binutils-gdb/gdb/gdbtypes.c
Joel Brobecker 3666a04883 Update copyright year range in all GDB files
This commits the result of running gdb/copyright.py as per our Start
of New Year procedure...

gdb/ChangeLog

        Update copyright year range in copyright header of all GDB files.
2021-01-01 12:12:21 +04:00

6331 lines
182 KiB
C
Raw Blame History

This file contains invisible Unicode characters

This file contains invisible Unicode characters that are indistinguishable to humans but may be processed differently by a computer. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

/* Support routines for manipulating internal types for GDB.
Copyright (C) 1992-2021 Free Software Foundation, Inc.
Contributed by Cygnus Support, using pieces from other GDB modules.
This file is part of GDB.
This program is free software; you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation; either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>. */
#include "defs.h"
#include "bfd.h"
#include "symtab.h"
#include "symfile.h"
#include "objfiles.h"
#include "gdbtypes.h"
#include "expression.h"
#include "language.h"
#include "target.h"
#include "value.h"
#include "demangle.h"
#include "complaints.h"
#include "gdbcmd.h"
#include "cp-abi.h"
#include "hashtab.h"
#include "cp-support.h"
#include "bcache.h"
#include "dwarf2/loc.h"
#include "gdbcore.h"
#include "floatformat.h"
#include "f-lang.h"
#include <algorithm>
#include "gmp-utils.h"
/* Initialize BADNESS constants. */
const struct rank LENGTH_MISMATCH_BADNESS = {100,0};
const struct rank TOO_FEW_PARAMS_BADNESS = {100,0};
const struct rank INCOMPATIBLE_TYPE_BADNESS = {100,0};
const struct rank EXACT_MATCH_BADNESS = {0,0};
const struct rank INTEGER_PROMOTION_BADNESS = {1,0};
const struct rank FLOAT_PROMOTION_BADNESS = {1,0};
const struct rank BASE_PTR_CONVERSION_BADNESS = {1,0};
const struct rank CV_CONVERSION_BADNESS = {1, 0};
const struct rank INTEGER_CONVERSION_BADNESS = {2,0};
const struct rank FLOAT_CONVERSION_BADNESS = {2,0};
const struct rank INT_FLOAT_CONVERSION_BADNESS = {2,0};
const struct rank VOID_PTR_CONVERSION_BADNESS = {2,0};
const struct rank BOOL_CONVERSION_BADNESS = {3,0};
const struct rank BASE_CONVERSION_BADNESS = {2,0};
const struct rank REFERENCE_CONVERSION_BADNESS = {2,0};
const struct rank REFERENCE_SEE_THROUGH_BADNESS = {0,1};
const struct rank NULL_POINTER_CONVERSION_BADNESS = {2,0};
const struct rank NS_POINTER_CONVERSION_BADNESS = {10,0};
const struct rank NS_INTEGER_POINTER_CONVERSION_BADNESS = {3,0};
/* Floatformat pairs. */
const struct floatformat *floatformats_ieee_half[BFD_ENDIAN_UNKNOWN] = {
&floatformat_ieee_half_big,
&floatformat_ieee_half_little
};
const struct floatformat *floatformats_ieee_single[BFD_ENDIAN_UNKNOWN] = {
&floatformat_ieee_single_big,
&floatformat_ieee_single_little
};
const struct floatformat *floatformats_ieee_double[BFD_ENDIAN_UNKNOWN] = {
&floatformat_ieee_double_big,
&floatformat_ieee_double_little
};
const struct floatformat *floatformats_ieee_double_littlebyte_bigword[BFD_ENDIAN_UNKNOWN] = {
&floatformat_ieee_double_big,
&floatformat_ieee_double_littlebyte_bigword
};
const struct floatformat *floatformats_i387_ext[BFD_ENDIAN_UNKNOWN] = {
&floatformat_i387_ext,
&floatformat_i387_ext
};
const struct floatformat *floatformats_m68881_ext[BFD_ENDIAN_UNKNOWN] = {
&floatformat_m68881_ext,
&floatformat_m68881_ext
};
const struct floatformat *floatformats_arm_ext[BFD_ENDIAN_UNKNOWN] = {
&floatformat_arm_ext_big,
&floatformat_arm_ext_littlebyte_bigword
};
const struct floatformat *floatformats_ia64_spill[BFD_ENDIAN_UNKNOWN] = {
&floatformat_ia64_spill_big,
&floatformat_ia64_spill_little
};
const struct floatformat *floatformats_ia64_quad[BFD_ENDIAN_UNKNOWN] = {
&floatformat_ia64_quad_big,
&floatformat_ia64_quad_little
};
const struct floatformat *floatformats_vax_f[BFD_ENDIAN_UNKNOWN] = {
&floatformat_vax_f,
&floatformat_vax_f
};
const struct floatformat *floatformats_vax_d[BFD_ENDIAN_UNKNOWN] = {
&floatformat_vax_d,
&floatformat_vax_d
};
const struct floatformat *floatformats_ibm_long_double[BFD_ENDIAN_UNKNOWN] = {
&floatformat_ibm_long_double_big,
&floatformat_ibm_long_double_little
};
const struct floatformat *floatformats_bfloat16[BFD_ENDIAN_UNKNOWN] = {
&floatformat_bfloat16_big,
&floatformat_bfloat16_little
};
/* Should opaque types be resolved? */
static bool opaque_type_resolution = true;
/* See gdbtypes.h. */
unsigned int overload_debug = 0;
/* A flag to enable strict type checking. */
static bool strict_type_checking = true;
/* A function to show whether opaque types are resolved. */
static void
show_opaque_type_resolution (struct ui_file *file, int from_tty,
struct cmd_list_element *c,
const char *value)
{
fprintf_filtered (file, _("Resolution of opaque struct/class/union types "
"(if set before loading symbols) is %s.\n"),
value);
}
/* A function to show whether C++ overload debugging is enabled. */
static void
show_overload_debug (struct ui_file *file, int from_tty,
struct cmd_list_element *c, const char *value)
{
fprintf_filtered (file, _("Debugging of C++ overloading is %s.\n"),
value);
}
/* A function to show the status of strict type checking. */
static void
show_strict_type_checking (struct ui_file *file, int from_tty,
struct cmd_list_element *c, const char *value)
{
fprintf_filtered (file, _("Strict type checking is %s.\n"), value);
}
/* Allocate a new OBJFILE-associated type structure and fill it
with some defaults. Space for the type structure is allocated
on the objfile's objfile_obstack. */
struct type *
alloc_type (struct objfile *objfile)
{
struct type *type;
gdb_assert (objfile != NULL);
/* Alloc the structure and start off with all fields zeroed. */
type = OBSTACK_ZALLOC (&objfile->objfile_obstack, struct type);
TYPE_MAIN_TYPE (type) = OBSTACK_ZALLOC (&objfile->objfile_obstack,
struct main_type);
OBJSTAT (objfile, n_types++);
TYPE_OBJFILE_OWNED (type) = 1;
TYPE_OWNER (type).objfile = objfile;
/* Initialize the fields that might not be zero. */
type->set_code (TYPE_CODE_UNDEF);
TYPE_CHAIN (type) = type; /* Chain back to itself. */
return type;
}
/* Allocate a new GDBARCH-associated type structure and fill it
with some defaults. Space for the type structure is allocated
on the obstack associated with GDBARCH. */
struct type *
alloc_type_arch (struct gdbarch *gdbarch)
{
struct type *type;
gdb_assert (gdbarch != NULL);
/* Alloc the structure and start off with all fields zeroed. */
type = GDBARCH_OBSTACK_ZALLOC (gdbarch, struct type);
TYPE_MAIN_TYPE (type) = GDBARCH_OBSTACK_ZALLOC (gdbarch, struct main_type);
TYPE_OBJFILE_OWNED (type) = 0;
TYPE_OWNER (type).gdbarch = gdbarch;
/* Initialize the fields that might not be zero. */
type->set_code (TYPE_CODE_UNDEF);
TYPE_CHAIN (type) = type; /* Chain back to itself. */
return type;
}
/* If TYPE is objfile-associated, allocate a new type structure
associated with the same objfile. If TYPE is gdbarch-associated,
allocate a new type structure associated with the same gdbarch. */
struct type *
alloc_type_copy (const struct type *type)
{
if (TYPE_OBJFILE_OWNED (type))
return alloc_type (TYPE_OWNER (type).objfile);
else
return alloc_type_arch (TYPE_OWNER (type).gdbarch);
}
/* If TYPE is gdbarch-associated, return that architecture.
If TYPE is objfile-associated, return that objfile's architecture. */
struct gdbarch *
get_type_arch (const struct type *type)
{
struct gdbarch *arch;
if (TYPE_OBJFILE_OWNED (type))
arch = TYPE_OWNER (type).objfile->arch ();
else
arch = TYPE_OWNER (type).gdbarch;
/* The ARCH can be NULL if TYPE is associated with neither an objfile nor
a gdbarch, however, this is very rare, and even then, in most cases
that get_type_arch is called, we assume that a non-NULL value is
returned. */
gdb_assert (arch != NULL);
return arch;
}
/* See gdbtypes.h. */
struct type *
get_target_type (struct type *type)
{
if (type != NULL)
{
type = TYPE_TARGET_TYPE (type);
if (type != NULL)
type = check_typedef (type);
}
return type;
}
/* See gdbtypes.h. */
unsigned int
type_length_units (struct type *type)
{
struct gdbarch *arch = get_type_arch (type);
int unit_size = gdbarch_addressable_memory_unit_size (arch);
return TYPE_LENGTH (type) / unit_size;
}
/* Alloc a new type instance structure, fill it with some defaults,
and point it at OLDTYPE. Allocate the new type instance from the
same place as OLDTYPE. */
static struct type *
alloc_type_instance (struct type *oldtype)
{
struct type *type;
/* Allocate the structure. */
if (! TYPE_OBJFILE_OWNED (oldtype))
type = GDBARCH_OBSTACK_ZALLOC (get_type_arch (oldtype), struct type);
else
type = OBSTACK_ZALLOC (&TYPE_OBJFILE (oldtype)->objfile_obstack,
struct type);
TYPE_MAIN_TYPE (type) = TYPE_MAIN_TYPE (oldtype);
TYPE_CHAIN (type) = type; /* Chain back to itself for now. */
return type;
}
/* Clear all remnants of the previous type at TYPE, in preparation for
replacing it with something else. Preserve owner information. */
static void
smash_type (struct type *type)
{
int objfile_owned = TYPE_OBJFILE_OWNED (type);
union type_owner owner = TYPE_OWNER (type);
memset (TYPE_MAIN_TYPE (type), 0, sizeof (struct main_type));
/* Restore owner information. */
TYPE_OBJFILE_OWNED (type) = objfile_owned;
TYPE_OWNER (type) = owner;
/* For now, delete the rings. */
TYPE_CHAIN (type) = type;
/* For now, leave the pointer/reference types alone. */
}
/* Lookup a pointer to a type TYPE. TYPEPTR, if nonzero, points
to a pointer to memory where the pointer type should be stored.
If *TYPEPTR is zero, update it to point to the pointer type we return.
We allocate new memory if needed. */
struct type *
make_pointer_type (struct type *type, struct type **typeptr)
{
struct type *ntype; /* New type */
struct type *chain;
ntype = TYPE_POINTER_TYPE (type);
if (ntype)
{
if (typeptr == 0)
return ntype; /* Don't care about alloc,
and have new type. */
else if (*typeptr == 0)
{
*typeptr = ntype; /* Tracking alloc, and have new type. */
return ntype;
}
}
if (typeptr == 0 || *typeptr == 0) /* We'll need to allocate one. */
{
ntype = alloc_type_copy (type);
if (typeptr)
*typeptr = ntype;
}
else /* We have storage, but need to reset it. */
{
ntype = *typeptr;
chain = TYPE_CHAIN (ntype);
smash_type (ntype);
TYPE_CHAIN (ntype) = chain;
}
TYPE_TARGET_TYPE (ntype) = type;
TYPE_POINTER_TYPE (type) = ntype;
/* FIXME! Assumes the machine has only one representation for pointers! */
TYPE_LENGTH (ntype)
= gdbarch_ptr_bit (get_type_arch (type)) / TARGET_CHAR_BIT;
ntype->set_code (TYPE_CODE_PTR);
/* Mark pointers as unsigned. The target converts between pointers
and addresses (CORE_ADDRs) using gdbarch_pointer_to_address and
gdbarch_address_to_pointer. */
ntype->set_is_unsigned (true);
/* Update the length of all the other variants of this type. */
chain = TYPE_CHAIN (ntype);
while (chain != ntype)
{
TYPE_LENGTH (chain) = TYPE_LENGTH (ntype);
chain = TYPE_CHAIN (chain);
}
return ntype;
}
/* Given a type TYPE, return a type of pointers to that type.
May need to construct such a type if this is the first use. */
struct type *
lookup_pointer_type (struct type *type)
{
return make_pointer_type (type, (struct type **) 0);
}
/* Lookup a C++ `reference' to a type TYPE. TYPEPTR, if nonzero,
points to a pointer to memory where the reference type should be
stored. If *TYPEPTR is zero, update it to point to the reference
type we return. We allocate new memory if needed. REFCODE denotes
the kind of reference type to lookup (lvalue or rvalue reference). */
struct type *
make_reference_type (struct type *type, struct type **typeptr,
enum type_code refcode)
{
struct type *ntype; /* New type */
struct type **reftype;
struct type *chain;
gdb_assert (refcode == TYPE_CODE_REF || refcode == TYPE_CODE_RVALUE_REF);
ntype = (refcode == TYPE_CODE_REF ? TYPE_REFERENCE_TYPE (type)
: TYPE_RVALUE_REFERENCE_TYPE (type));
if (ntype)
{
if (typeptr == 0)
return ntype; /* Don't care about alloc,
and have new type. */
else if (*typeptr == 0)
{
*typeptr = ntype; /* Tracking alloc, and have new type. */
return ntype;
}
}
if (typeptr == 0 || *typeptr == 0) /* We'll need to allocate one. */
{
ntype = alloc_type_copy (type);
if (typeptr)
*typeptr = ntype;
}
else /* We have storage, but need to reset it. */
{
ntype = *typeptr;
chain = TYPE_CHAIN (ntype);
smash_type (ntype);
TYPE_CHAIN (ntype) = chain;
}
TYPE_TARGET_TYPE (ntype) = type;
reftype = (refcode == TYPE_CODE_REF ? &TYPE_REFERENCE_TYPE (type)
: &TYPE_RVALUE_REFERENCE_TYPE (type));
*reftype = ntype;
/* FIXME! Assume the machine has only one representation for
references, and that it matches the (only) representation for
pointers! */
TYPE_LENGTH (ntype) =
gdbarch_ptr_bit (get_type_arch (type)) / TARGET_CHAR_BIT;
ntype->set_code (refcode);
*reftype = ntype;
/* Update the length of all the other variants of this type. */
chain = TYPE_CHAIN (ntype);
while (chain != ntype)
{
TYPE_LENGTH (chain) = TYPE_LENGTH (ntype);
chain = TYPE_CHAIN (chain);
}
return ntype;
}
/* Same as above, but caller doesn't care about memory allocation
details. */
struct type *
lookup_reference_type (struct type *type, enum type_code refcode)
{
return make_reference_type (type, (struct type **) 0, refcode);
}
/* Lookup the lvalue reference type for the type TYPE. */
struct type *
lookup_lvalue_reference_type (struct type *type)
{
return lookup_reference_type (type, TYPE_CODE_REF);
}
/* Lookup the rvalue reference type for the type TYPE. */
struct type *
lookup_rvalue_reference_type (struct type *type)
{
return lookup_reference_type (type, TYPE_CODE_RVALUE_REF);
}
/* Lookup a function type that returns type TYPE. TYPEPTR, if
nonzero, points to a pointer to memory where the function type
should be stored. If *TYPEPTR is zero, update it to point to the
function type we return. We allocate new memory if needed. */
struct type *
make_function_type (struct type *type, struct type **typeptr)
{
struct type *ntype; /* New type */
if (typeptr == 0 || *typeptr == 0) /* We'll need to allocate one. */
{
ntype = alloc_type_copy (type);
if (typeptr)
*typeptr = ntype;
}
else /* We have storage, but need to reset it. */
{
ntype = *typeptr;
smash_type (ntype);
}
TYPE_TARGET_TYPE (ntype) = type;
TYPE_LENGTH (ntype) = 1;
ntype->set_code (TYPE_CODE_FUNC);
INIT_FUNC_SPECIFIC (ntype);
return ntype;
}
/* Given a type TYPE, return a type of functions that return that type.
May need to construct such a type if this is the first use. */
struct type *
lookup_function_type (struct type *type)
{
return make_function_type (type, (struct type **) 0);
}
/* Given a type TYPE and argument types, return the appropriate
function type. If the final type in PARAM_TYPES is NULL, make a
varargs function. */
struct type *
lookup_function_type_with_arguments (struct type *type,
int nparams,
struct type **param_types)
{
struct type *fn = make_function_type (type, (struct type **) 0);
int i;
if (nparams > 0)
{
if (param_types[nparams - 1] == NULL)
{
--nparams;
fn->set_has_varargs (true);
}
else if (check_typedef (param_types[nparams - 1])->code ()
== TYPE_CODE_VOID)
{
--nparams;
/* Caller should have ensured this. */
gdb_assert (nparams == 0);
fn->set_is_prototyped (true);
}
else
fn->set_is_prototyped (true);
}
fn->set_num_fields (nparams);
fn->set_fields
((struct field *) TYPE_ZALLOC (fn, nparams * sizeof (struct field)));
for (i = 0; i < nparams; ++i)
fn->field (i).set_type (param_types[i]);
return fn;
}
/* Identify address space identifier by name -- return a
type_instance_flags. */
type_instance_flags
address_space_name_to_type_instance_flags (struct gdbarch *gdbarch,
const char *space_identifier)
{
type_instance_flags type_flags;
/* Check for known address space delimiters. */
if (!strcmp (space_identifier, "code"))
return TYPE_INSTANCE_FLAG_CODE_SPACE;
else if (!strcmp (space_identifier, "data"))
return TYPE_INSTANCE_FLAG_DATA_SPACE;
else if (gdbarch_address_class_name_to_type_flags_p (gdbarch)
&& gdbarch_address_class_name_to_type_flags (gdbarch,
space_identifier,
&type_flags))
return type_flags;
else
error (_("Unknown address space specifier: \"%s\""), space_identifier);
}
/* Identify address space identifier by type_instance_flags and return
the string version of the adress space name. */
const char *
address_space_type_instance_flags_to_name (struct gdbarch *gdbarch,
type_instance_flags space_flag)
{
if (space_flag & TYPE_INSTANCE_FLAG_CODE_SPACE)
return "code";
else if (space_flag & TYPE_INSTANCE_FLAG_DATA_SPACE)
return "data";
else if ((space_flag & TYPE_INSTANCE_FLAG_ADDRESS_CLASS_ALL)
&& gdbarch_address_class_type_flags_to_name_p (gdbarch))
return gdbarch_address_class_type_flags_to_name (gdbarch, space_flag);
else
return NULL;
}
/* Create a new type with instance flags NEW_FLAGS, based on TYPE.
If STORAGE is non-NULL, create the new type instance there.
STORAGE must be in the same obstack as TYPE. */
static struct type *
make_qualified_type (struct type *type, type_instance_flags new_flags,
struct type *storage)
{
struct type *ntype;
ntype = type;
do
{
if (ntype->instance_flags () == new_flags)
return ntype;
ntype = TYPE_CHAIN (ntype);
}
while (ntype != type);
/* Create a new type instance. */
if (storage == NULL)
ntype = alloc_type_instance (type);
else
{
/* If STORAGE was provided, it had better be in the same objfile
as TYPE. Otherwise, we can't link it into TYPE's cv chain:
if one objfile is freed and the other kept, we'd have
dangling pointers. */
gdb_assert (TYPE_OBJFILE (type) == TYPE_OBJFILE (storage));
ntype = storage;
TYPE_MAIN_TYPE (ntype) = TYPE_MAIN_TYPE (type);
TYPE_CHAIN (ntype) = ntype;
}
/* Pointers or references to the original type are not relevant to
the new type. */
TYPE_POINTER_TYPE (ntype) = (struct type *) 0;
TYPE_REFERENCE_TYPE (ntype) = (struct type *) 0;
/* Chain the new qualified type to the old type. */
TYPE_CHAIN (ntype) = TYPE_CHAIN (type);
TYPE_CHAIN (type) = ntype;
/* Now set the instance flags and return the new type. */
ntype->set_instance_flags (new_flags);
/* Set length of new type to that of the original type. */
TYPE_LENGTH (ntype) = TYPE_LENGTH (type);
return ntype;
}
/* Make an address-space-delimited variant of a type -- a type that
is identical to the one supplied except that it has an address
space attribute attached to it (such as "code" or "data").
The space attributes "code" and "data" are for Harvard
architectures. The address space attributes are for architectures
which have alternately sized pointers or pointers with alternate
representations. */
struct type *
make_type_with_address_space (struct type *type,
type_instance_flags space_flag)
{
type_instance_flags new_flags = ((type->instance_flags ()
& ~(TYPE_INSTANCE_FLAG_CODE_SPACE
| TYPE_INSTANCE_FLAG_DATA_SPACE
| TYPE_INSTANCE_FLAG_ADDRESS_CLASS_ALL))
| space_flag);
return make_qualified_type (type, new_flags, NULL);
}
/* Make a "c-v" variant of a type -- a type that is identical to the
one supplied except that it may have const or volatile attributes
CNST is a flag for setting the const attribute
VOLTL is a flag for setting the volatile attribute
TYPE is the base type whose variant we are creating.
If TYPEPTR and *TYPEPTR are non-zero, then *TYPEPTR points to
storage to hold the new qualified type; *TYPEPTR and TYPE must be
in the same objfile. Otherwise, allocate fresh memory for the new
type whereever TYPE lives. If TYPEPTR is non-zero, set it to the
new type we construct. */
struct type *
make_cv_type (int cnst, int voltl,
struct type *type,
struct type **typeptr)
{
struct type *ntype; /* New type */
type_instance_flags new_flags = (type->instance_flags ()
& ~(TYPE_INSTANCE_FLAG_CONST
| TYPE_INSTANCE_FLAG_VOLATILE));
if (cnst)
new_flags |= TYPE_INSTANCE_FLAG_CONST;
if (voltl)
new_flags |= TYPE_INSTANCE_FLAG_VOLATILE;
if (typeptr && *typeptr != NULL)
{
/* TYPE and *TYPEPTR must be in the same objfile. We can't have
a C-V variant chain that threads across objfiles: if one
objfile gets freed, then the other has a broken C-V chain.
This code used to try to copy over the main type from TYPE to
*TYPEPTR if they were in different objfiles, but that's
wrong, too: TYPE may have a field list or member function
lists, which refer to types of their own, etc. etc. The
whole shebang would need to be copied over recursively; you
can't have inter-objfile pointers. The only thing to do is
to leave stub types as stub types, and look them up afresh by
name each time you encounter them. */
gdb_assert (TYPE_OBJFILE (*typeptr) == TYPE_OBJFILE (type));
}
ntype = make_qualified_type (type, new_flags,
typeptr ? *typeptr : NULL);
if (typeptr != NULL)
*typeptr = ntype;
return ntype;
}
/* Make a 'restrict'-qualified version of TYPE. */
struct type *
make_restrict_type (struct type *type)
{
return make_qualified_type (type,
(type->instance_flags ()
| TYPE_INSTANCE_FLAG_RESTRICT),
NULL);
}
/* Make a type without const, volatile, or restrict. */
struct type *
make_unqualified_type (struct type *type)
{
return make_qualified_type (type,
(type->instance_flags ()
& ~(TYPE_INSTANCE_FLAG_CONST
| TYPE_INSTANCE_FLAG_VOLATILE
| TYPE_INSTANCE_FLAG_RESTRICT)),
NULL);
}
/* Make a '_Atomic'-qualified version of TYPE. */
struct type *
make_atomic_type (struct type *type)
{
return make_qualified_type (type,
(type->instance_flags ()
| TYPE_INSTANCE_FLAG_ATOMIC),
NULL);
}
/* Replace the contents of ntype with the type *type. This changes the
contents, rather than the pointer for TYPE_MAIN_TYPE (ntype); thus
the changes are propogated to all types in the TYPE_CHAIN.
In order to build recursive types, it's inevitable that we'll need
to update types in place --- but this sort of indiscriminate
smashing is ugly, and needs to be replaced with something more
controlled. TYPE_MAIN_TYPE is a step in this direction; it's not
clear if more steps are needed. */
void
replace_type (struct type *ntype, struct type *type)
{
struct type *chain;
/* These two types had better be in the same objfile. Otherwise,
the assignment of one type's main type structure to the other
will produce a type with references to objects (names; field
lists; etc.) allocated on an objfile other than its own. */
gdb_assert (TYPE_OBJFILE (ntype) == TYPE_OBJFILE (type));
*TYPE_MAIN_TYPE (ntype) = *TYPE_MAIN_TYPE (type);
/* The type length is not a part of the main type. Update it for
each type on the variant chain. */
chain = ntype;
do
{
/* Assert that this element of the chain has no address-class bits
set in its flags. Such type variants might have type lengths
which are supposed to be different from the non-address-class
variants. This assertion shouldn't ever be triggered because
symbol readers which do construct address-class variants don't
call replace_type(). */
gdb_assert (TYPE_ADDRESS_CLASS_ALL (chain) == 0);
TYPE_LENGTH (chain) = TYPE_LENGTH (type);
chain = TYPE_CHAIN (chain);
}
while (ntype != chain);
/* Assert that the two types have equivalent instance qualifiers.
This should be true for at least all of our debug readers. */
gdb_assert (ntype->instance_flags () == type->instance_flags ());
}
/* Implement direct support for MEMBER_TYPE in GNU C++.
May need to construct such a type if this is the first use.
The TYPE is the type of the member. The DOMAIN is the type
of the aggregate that the member belongs to. */
struct type *
lookup_memberptr_type (struct type *type, struct type *domain)
{
struct type *mtype;
mtype = alloc_type_copy (type);
smash_to_memberptr_type (mtype, domain, type);
return mtype;
}
/* Return a pointer-to-method type, for a method of type TO_TYPE. */
struct type *
lookup_methodptr_type (struct type *to_type)
{
struct type *mtype;
mtype = alloc_type_copy (to_type);
smash_to_methodptr_type (mtype, to_type);
return mtype;
}
/* Allocate a stub method whose return type is TYPE. This apparently
happens for speed of symbol reading, since parsing out the
arguments to the method is cpu-intensive, the way we are doing it.
So, we will fill in arguments later. This always returns a fresh
type. */
struct type *
allocate_stub_method (struct type *type)
{
struct type *mtype;
mtype = alloc_type_copy (type);
mtype->set_code (TYPE_CODE_METHOD);
TYPE_LENGTH (mtype) = 1;
mtype->set_is_stub (true);
TYPE_TARGET_TYPE (mtype) = type;
/* TYPE_SELF_TYPE (mtype) = unknown yet */
return mtype;
}
/* See gdbtypes.h. */
bool
operator== (const dynamic_prop &l, const dynamic_prop &r)
{
if (l.kind () != r.kind ())
return false;
switch (l.kind ())
{
case PROP_UNDEFINED:
return true;
case PROP_CONST:
return l.const_val () == r.const_val ();
case PROP_ADDR_OFFSET:
case PROP_LOCEXPR:
case PROP_LOCLIST:
return l.baton () == r.baton ();
case PROP_VARIANT_PARTS:
return l.variant_parts () == r.variant_parts ();
case PROP_TYPE:
return l.original_type () == r.original_type ();
}
gdb_assert_not_reached ("unhandled dynamic_prop kind");
}
/* See gdbtypes.h. */
bool
operator== (const range_bounds &l, const range_bounds &r)
{
#define FIELD_EQ(FIELD) (l.FIELD == r.FIELD)
return (FIELD_EQ (low)
&& FIELD_EQ (high)
&& FIELD_EQ (flag_upper_bound_is_count)
&& FIELD_EQ (flag_bound_evaluated)
&& FIELD_EQ (bias));
#undef FIELD_EQ
}
/* Create a range type with a dynamic range from LOW_BOUND to
HIGH_BOUND, inclusive. See create_range_type for further details. */
struct type *
create_range_type (struct type *result_type, struct type *index_type,
const struct dynamic_prop *low_bound,
const struct dynamic_prop *high_bound,
LONGEST bias)
{
/* The INDEX_TYPE should be a type capable of holding the upper and lower
bounds, as such a zero sized, or void type makes no sense. */
gdb_assert (index_type->code () != TYPE_CODE_VOID);
gdb_assert (TYPE_LENGTH (index_type) > 0);
if (result_type == NULL)
result_type = alloc_type_copy (index_type);
result_type->set_code (TYPE_CODE_RANGE);
TYPE_TARGET_TYPE (result_type) = index_type;
if (index_type->is_stub ())
result_type->set_target_is_stub (true);
else
TYPE_LENGTH (result_type) = TYPE_LENGTH (check_typedef (index_type));
range_bounds *bounds
= (struct range_bounds *) TYPE_ZALLOC (result_type, sizeof (range_bounds));
bounds->low = *low_bound;
bounds->high = *high_bound;
bounds->bias = bias;
bounds->stride.set_const_val (0);
result_type->set_bounds (bounds);
if (index_type->code () == TYPE_CODE_FIXED_POINT)
result_type->set_is_unsigned (index_type->is_unsigned ());
/* Note that the signed-ness of a range type can't simply be copied
from the underlying type. Consider a case where the underlying
type is 'int', but the range type can hold 0..65535, and where
the range is further specified to fit into 16 bits. In this
case, if we copy the underlying type's sign, then reading some
range values will cause an unwanted sign extension. So, we have
some heuristics here instead. */
else if (low_bound->kind () == PROP_CONST && low_bound->const_val () >= 0)
result_type->set_is_unsigned (true);
/* Ada allows the declaration of range types whose upper bound is
less than the lower bound, so checking the lower bound is not
enough. Make sure we do not mark a range type whose upper bound
is negative as unsigned. */
if (high_bound->kind () == PROP_CONST && high_bound->const_val () < 0)
result_type->set_is_unsigned (false);
result_type->set_endianity_is_not_default
(index_type->endianity_is_not_default ());
return result_type;
}
/* See gdbtypes.h. */
struct type *
create_range_type_with_stride (struct type *result_type,
struct type *index_type,
const struct dynamic_prop *low_bound,
const struct dynamic_prop *high_bound,
LONGEST bias,
const struct dynamic_prop *stride,
bool byte_stride_p)
{
result_type = create_range_type (result_type, index_type, low_bound,
high_bound, bias);
gdb_assert (stride != nullptr);
result_type->bounds ()->stride = *stride;
result_type->bounds ()->flag_is_byte_stride = byte_stride_p;
return result_type;
}
/* Create a range type using either a blank type supplied in
RESULT_TYPE, or creating a new type, inheriting the objfile from
INDEX_TYPE.
Indices will be of type INDEX_TYPE, and will range from LOW_BOUND
to HIGH_BOUND, inclusive.
FIXME: Maybe we should check the TYPE_CODE of RESULT_TYPE to make
sure it is TYPE_CODE_UNDEF before we bash it into a range type? */
struct type *
create_static_range_type (struct type *result_type, struct type *index_type,
LONGEST low_bound, LONGEST high_bound)
{
struct dynamic_prop low, high;
low.set_const_val (low_bound);
high.set_const_val (high_bound);
result_type = create_range_type (result_type, index_type, &low, &high, 0);
return result_type;
}
/* Predicate tests whether BOUNDS are static. Returns 1 if all bounds values
are static, otherwise returns 0. */
static bool
has_static_range (const struct range_bounds *bounds)
{
/* If the range doesn't have a defined stride then its stride field will
be initialized to the constant 0. */
return (bounds->low.kind () == PROP_CONST
&& bounds->high.kind () == PROP_CONST
&& bounds->stride.kind () == PROP_CONST);
}
/* See gdbtypes.h. */
gdb::optional<LONGEST>
get_discrete_low_bound (struct type *type)
{
type = check_typedef (type);
switch (type->code ())
{
case TYPE_CODE_RANGE:
{
/* This function only works for ranges with a constant low bound. */
if (type->bounds ()->low.kind () != PROP_CONST)
return {};
LONGEST low = type->bounds ()->low.const_val ();
if (TYPE_TARGET_TYPE (type)->code () == TYPE_CODE_ENUM)
{
gdb::optional<LONGEST> low_pos
= discrete_position (TYPE_TARGET_TYPE (type), low);
if (low_pos.has_value ())
low = *low_pos;
}
return low;
}
case TYPE_CODE_ENUM:
{
if (type->num_fields () > 0)
{
/* The enums may not be sorted by value, so search all
entries. */
LONGEST low = TYPE_FIELD_ENUMVAL (type, 0);
for (int i = 0; i < type->num_fields (); i++)
{
if (TYPE_FIELD_ENUMVAL (type, i) < low)
low = TYPE_FIELD_ENUMVAL (type, i);
}
/* Set unsigned indicator if warranted. */
if (low >= 0)
type->set_is_unsigned (true);
return low;
}
else
return 0;
}
case TYPE_CODE_BOOL:
return 0;
case TYPE_CODE_INT:
if (TYPE_LENGTH (type) > sizeof (LONGEST)) /* Too big */
return {};
if (!type->is_unsigned ())
return -(1 << (TYPE_LENGTH (type) * TARGET_CHAR_BIT - 1));
/* fall through */
case TYPE_CODE_CHAR:
return 0;
default:
return {};
}
}
/* See gdbtypes.h. */
gdb::optional<LONGEST>
get_discrete_high_bound (struct type *type)
{
type = check_typedef (type);
switch (type->code ())
{
case TYPE_CODE_RANGE:
{
/* This function only works for ranges with a constant high bound. */
if (type->bounds ()->high.kind () != PROP_CONST)
return {};
LONGEST high = type->bounds ()->high.const_val ();
if (TYPE_TARGET_TYPE (type)->code () == TYPE_CODE_ENUM)
{
gdb::optional<LONGEST> high_pos
= discrete_position (TYPE_TARGET_TYPE (type), high);
if (high_pos.has_value ())
high = *high_pos;
}
return high;
}
case TYPE_CODE_ENUM:
{
if (type->num_fields () > 0)
{
/* The enums may not be sorted by value, so search all
entries. */
LONGEST high = TYPE_FIELD_ENUMVAL (type, 0);
for (int i = 0; i < type->num_fields (); i++)
{
if (TYPE_FIELD_ENUMVAL (type, i) > high)
high = TYPE_FIELD_ENUMVAL (type, i);
}
return high;
}
else
return -1;
}
case TYPE_CODE_BOOL:
return 1;
case TYPE_CODE_INT:
if (TYPE_LENGTH (type) > sizeof (LONGEST)) /* Too big */
return {};
if (!type->is_unsigned ())
{
LONGEST low = -(1 << (TYPE_LENGTH (type) * TARGET_CHAR_BIT - 1));
return -low - 1;
}
/* fall through */
case TYPE_CODE_CHAR:
{
/* This round-about calculation is to avoid shifting by
TYPE_LENGTH (type) * TARGET_CHAR_BIT, which will not work
if TYPE_LENGTH (type) == sizeof (LONGEST). */
LONGEST high = 1 << (TYPE_LENGTH (type) * TARGET_CHAR_BIT - 1);
return (high - 1) | high;
}
default:
return {};
}
}
/* See gdbtypes.h. */
bool
get_discrete_bounds (struct type *type, LONGEST *lowp, LONGEST *highp)
{
gdb::optional<LONGEST> low = get_discrete_low_bound (type);
if (!low.has_value ())
return false;
gdb::optional<LONGEST> high = get_discrete_high_bound (type);
if (!high.has_value ())
return false;
*lowp = *low;
*highp = *high;
return true;
}
/* See gdbtypes.h */
bool
get_array_bounds (struct type *type, LONGEST *low_bound, LONGEST *high_bound)
{
struct type *index = type->index_type ();
LONGEST low = 0;
LONGEST high = 0;
if (index == NULL)
return false;
if (!get_discrete_bounds (index, &low, &high))
return false;
if (low_bound)
*low_bound = low;
if (high_bound)
*high_bound = high;
return true;
}
/* Assuming that TYPE is a discrete type and VAL is a valid integer
representation of a value of this type, save the corresponding
position number in POS.
Its differs from VAL only in the case of enumeration types. In
this case, the position number of the value of the first listed
enumeration literal is zero; the position number of the value of
each subsequent enumeration literal is one more than that of its
predecessor in the list.
Return 1 if the operation was successful. Return zero otherwise,
in which case the value of POS is unmodified.
*/
gdb::optional<LONGEST>
discrete_position (struct type *type, LONGEST val)
{
if (type->code () == TYPE_CODE_RANGE)
type = TYPE_TARGET_TYPE (type);
if (type->code () == TYPE_CODE_ENUM)
{
int i;
for (i = 0; i < type->num_fields (); i += 1)
{
if (val == TYPE_FIELD_ENUMVAL (type, i))
return i;
}
/* Invalid enumeration value. */
return {};
}
else
return val;
}
/* If the array TYPE has static bounds calculate and update its
size, then return true. Otherwise return false and leave TYPE
unchanged. */
static bool
update_static_array_size (struct type *type)
{
gdb_assert (type->code () == TYPE_CODE_ARRAY);
struct type *range_type = type->index_type ();
if (type->dyn_prop (DYN_PROP_BYTE_STRIDE) == nullptr
&& has_static_range (range_type->bounds ())
&& (!type_not_associated (type)
&& !type_not_allocated (type)))
{
LONGEST low_bound, high_bound;
int stride;
struct type *element_type;
/* If the array itself doesn't provide a stride value then take
whatever stride the range provides. Don't update BIT_STRIDE as
we don't want to place the stride value from the range into this
arrays bit size field. */
stride = TYPE_FIELD_BITSIZE (type, 0);
if (stride == 0)
stride = range_type->bit_stride ();
if (!get_discrete_bounds (range_type, &low_bound, &high_bound))
low_bound = high_bound = 0;
element_type = check_typedef (TYPE_TARGET_TYPE (type));
/* Be careful when setting the array length. Ada arrays can be
empty arrays with the high_bound being smaller than the low_bound.
In such cases, the array length should be zero. */
if (high_bound < low_bound)
TYPE_LENGTH (type) = 0;
else if (stride != 0)
{
/* Ensure that the type length is always positive, even in the
case where (for example in Fortran) we have a negative
stride. It is possible to have a single element array with a
negative stride in Fortran (this doesn't mean anything
special, it's still just a single element array) so do
consider that case when touching this code. */
LONGEST element_count = std::abs (high_bound - low_bound + 1);
TYPE_LENGTH (type)
= ((std::abs (stride) * element_count) + 7) / 8;
}
else
TYPE_LENGTH (type) =
TYPE_LENGTH (element_type) * (high_bound - low_bound + 1);
/* If this array's element is itself an array with a bit stride,
then we want to update this array's bit stride to reflect the
size of the sub-array. Otherwise, we'll end up using the
wrong size when trying to find elements of the outer
array. */
if (element_type->code () == TYPE_CODE_ARRAY
&& TYPE_LENGTH (element_type) != 0
&& TYPE_FIELD_BITSIZE (element_type, 0) != 0
&& get_array_bounds (element_type, &low_bound, &high_bound)
&& high_bound >= low_bound)
TYPE_FIELD_BITSIZE (type, 0)
= ((high_bound - low_bound + 1)
* TYPE_FIELD_BITSIZE (element_type, 0));
return true;
}
return false;
}
/* Create an array type using either a blank type supplied in
RESULT_TYPE, or creating a new type, inheriting the objfile from
RANGE_TYPE.
Elements will be of type ELEMENT_TYPE, the indices will be of type
RANGE_TYPE.
BYTE_STRIDE_PROP, when not NULL, provides the array's byte stride.
This byte stride property is added to the resulting array type
as a DYN_PROP_BYTE_STRIDE. As a consequence, the BYTE_STRIDE_PROP
argument can only be used to create types that are objfile-owned
(see add_dyn_prop), meaning that either this function must be called
with an objfile-owned RESULT_TYPE, or an objfile-owned RANGE_TYPE.
BIT_STRIDE is taken into account only when BYTE_STRIDE_PROP is NULL.
If BIT_STRIDE is not zero, build a packed array type whose element
size is BIT_STRIDE. Otherwise, ignore this parameter.
FIXME: Maybe we should check the TYPE_CODE of RESULT_TYPE to make
sure it is TYPE_CODE_UNDEF before we bash it into an array
type? */
struct type *
create_array_type_with_stride (struct type *result_type,
struct type *element_type,
struct type *range_type,
struct dynamic_prop *byte_stride_prop,
unsigned int bit_stride)
{
if (byte_stride_prop != NULL
&& byte_stride_prop->kind () == PROP_CONST)
{
/* The byte stride is actually not dynamic. Pretend we were
called with bit_stride set instead of byte_stride_prop.
This will give us the same result type, while avoiding
the need to handle this as a special case. */
bit_stride = byte_stride_prop->const_val () * 8;
byte_stride_prop = NULL;
}
if (result_type == NULL)
result_type = alloc_type_copy (range_type);
result_type->set_code (TYPE_CODE_ARRAY);
TYPE_TARGET_TYPE (result_type) = element_type;
result_type->set_num_fields (1);
result_type->set_fields
((struct field *) TYPE_ZALLOC (result_type, sizeof (struct field)));
result_type->set_index_type (range_type);
if (byte_stride_prop != NULL)
result_type->add_dyn_prop (DYN_PROP_BYTE_STRIDE, *byte_stride_prop);
else if (bit_stride > 0)
TYPE_FIELD_BITSIZE (result_type, 0) = bit_stride;
if (!update_static_array_size (result_type))
{
/* This type is dynamic and its length needs to be computed
on demand. In the meantime, avoid leaving the TYPE_LENGTH
undefined by setting it to zero. Although we are not expected
to trust TYPE_LENGTH in this case, setting the size to zero
allows us to avoid allocating objects of random sizes in case
we accidently do. */
TYPE_LENGTH (result_type) = 0;
}
/* TYPE_TARGET_STUB will take care of zero length arrays. */
if (TYPE_LENGTH (result_type) == 0)
result_type->set_target_is_stub (true);
return result_type;
}
/* Same as create_array_type_with_stride but with no bit_stride
(BIT_STRIDE = 0), thus building an unpacked array. */
struct type *
create_array_type (struct type *result_type,
struct type *element_type,
struct type *range_type)
{
return create_array_type_with_stride (result_type, element_type,
range_type, NULL, 0);
}
struct type *
lookup_array_range_type (struct type *element_type,
LONGEST low_bound, LONGEST high_bound)
{
struct type *index_type;
struct type *range_type;
if (TYPE_OBJFILE_OWNED (element_type))
index_type = objfile_type (TYPE_OWNER (element_type).objfile)->builtin_int;
else
index_type = builtin_type (get_type_arch (element_type))->builtin_int;
range_type = create_static_range_type (NULL, index_type,
low_bound, high_bound);
return create_array_type (NULL, element_type, range_type);
}
/* Create a string type using either a blank type supplied in
RESULT_TYPE, or creating a new type. String types are similar
enough to array of char types that we can use create_array_type to
build the basic type and then bash it into a string type.
For fixed length strings, the range type contains 0 as the lower
bound and the length of the string minus one as the upper bound.
FIXME: Maybe we should check the TYPE_CODE of RESULT_TYPE to make
sure it is TYPE_CODE_UNDEF before we bash it into a string
type? */
struct type *
create_string_type (struct type *result_type,
struct type *string_char_type,
struct type *range_type)
{
result_type = create_array_type (result_type,
string_char_type,
range_type);
result_type->set_code (TYPE_CODE_STRING);
return result_type;
}
struct type *
lookup_string_range_type (struct type *string_char_type,
LONGEST low_bound, LONGEST high_bound)
{
struct type *result_type;
result_type = lookup_array_range_type (string_char_type,
low_bound, high_bound);
result_type->set_code (TYPE_CODE_STRING);
return result_type;
}
struct type *
create_set_type (struct type *result_type, struct type *domain_type)
{
if (result_type == NULL)
result_type = alloc_type_copy (domain_type);
result_type->set_code (TYPE_CODE_SET);
result_type->set_num_fields (1);
result_type->set_fields
((struct field *) TYPE_ZALLOC (result_type, sizeof (struct field)));
if (!domain_type->is_stub ())
{
LONGEST low_bound, high_bound, bit_length;
if (!get_discrete_bounds (domain_type, &low_bound, &high_bound))
low_bound = high_bound = 0;
bit_length = high_bound - low_bound + 1;
TYPE_LENGTH (result_type)
= (bit_length + TARGET_CHAR_BIT - 1) / TARGET_CHAR_BIT;
if (low_bound >= 0)
result_type->set_is_unsigned (true);
}
result_type->field (0).set_type (domain_type);
return result_type;
}
/* Convert ARRAY_TYPE to a vector type. This may modify ARRAY_TYPE
and any array types nested inside it. */
void
make_vector_type (struct type *array_type)
{
struct type *inner_array, *elt_type;
/* Find the innermost array type, in case the array is
multi-dimensional. */
inner_array = array_type;
while (TYPE_TARGET_TYPE (inner_array)->code () == TYPE_CODE_ARRAY)
inner_array = TYPE_TARGET_TYPE (inner_array);
elt_type = TYPE_TARGET_TYPE (inner_array);
if (elt_type->code () == TYPE_CODE_INT)
{
type_instance_flags flags
= elt_type->instance_flags () | TYPE_INSTANCE_FLAG_NOTTEXT;
elt_type = make_qualified_type (elt_type, flags, NULL);
TYPE_TARGET_TYPE (inner_array) = elt_type;
}
array_type->set_is_vector (true);
}
struct type *
init_vector_type (struct type *elt_type, int n)
{
struct type *array_type;
array_type = lookup_array_range_type (elt_type, 0, n - 1);
make_vector_type (array_type);
return array_type;
}
/* Internal routine called by TYPE_SELF_TYPE to return the type that TYPE
belongs to. In c++ this is the class of "this", but TYPE_THIS_TYPE is too
confusing. "self" is a common enough replacement for "this".
TYPE must be one of TYPE_CODE_METHODPTR, TYPE_CODE_MEMBERPTR, or
TYPE_CODE_METHOD. */
struct type *
internal_type_self_type (struct type *type)
{
switch (type->code ())
{
case TYPE_CODE_METHODPTR:
case TYPE_CODE_MEMBERPTR:
if (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_NONE)
return NULL;
gdb_assert (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_SELF_TYPE);
return TYPE_MAIN_TYPE (type)->type_specific.self_type;
case TYPE_CODE_METHOD:
if (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_NONE)
return NULL;
gdb_assert (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_FUNC);
return TYPE_MAIN_TYPE (type)->type_specific.func_stuff->self_type;
default:
gdb_assert_not_reached ("bad type");
}
}
/* Set the type of the class that TYPE belongs to.
In c++ this is the class of "this".
TYPE must be one of TYPE_CODE_METHODPTR, TYPE_CODE_MEMBERPTR, or
TYPE_CODE_METHOD. */
void
set_type_self_type (struct type *type, struct type *self_type)
{
switch (type->code ())
{
case TYPE_CODE_METHODPTR:
case TYPE_CODE_MEMBERPTR:
if (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_NONE)
TYPE_SPECIFIC_FIELD (type) = TYPE_SPECIFIC_SELF_TYPE;
gdb_assert (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_SELF_TYPE);
TYPE_MAIN_TYPE (type)->type_specific.self_type = self_type;
break;
case TYPE_CODE_METHOD:
if (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_NONE)
INIT_FUNC_SPECIFIC (type);
gdb_assert (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_FUNC);
TYPE_MAIN_TYPE (type)->type_specific.func_stuff->self_type = self_type;
break;
default:
gdb_assert_not_reached ("bad type");
}
}
/* Smash TYPE to be a type of pointers to members of SELF_TYPE with type
TO_TYPE. A member pointer is a wierd thing -- it amounts to a
typed offset into a struct, e.g. "an int at offset 8". A MEMBER
TYPE doesn't include the offset (that's the value of the MEMBER
itself), but does include the structure type into which it points
(for some reason).
When "smashing" the type, we preserve the objfile that the old type
pointed to, since we aren't changing where the type is actually
allocated. */
void
smash_to_memberptr_type (struct type *type, struct type *self_type,
struct type *to_type)
{
smash_type (type);
type->set_code (TYPE_CODE_MEMBERPTR);
TYPE_TARGET_TYPE (type) = to_type;
set_type_self_type (type, self_type);
/* Assume that a data member pointer is the same size as a normal
pointer. */
TYPE_LENGTH (type)
= gdbarch_ptr_bit (get_type_arch (to_type)) / TARGET_CHAR_BIT;
}
/* Smash TYPE to be a type of pointer to methods type TO_TYPE.
When "smashing" the type, we preserve the objfile that the old type
pointed to, since we aren't changing where the type is actually
allocated. */
void
smash_to_methodptr_type (struct type *type, struct type *to_type)
{
smash_type (type);
type->set_code (TYPE_CODE_METHODPTR);
TYPE_TARGET_TYPE (type) = to_type;
set_type_self_type (type, TYPE_SELF_TYPE (to_type));
TYPE_LENGTH (type) = cplus_method_ptr_size (to_type);
}
/* Smash TYPE to be a type of method of SELF_TYPE with type TO_TYPE.
METHOD just means `function that gets an extra "this" argument'.
When "smashing" the type, we preserve the objfile that the old type
pointed to, since we aren't changing where the type is actually
allocated. */
void
smash_to_method_type (struct type *type, struct type *self_type,
struct type *to_type, struct field *args,
int nargs, int varargs)
{
smash_type (type);
type->set_code (TYPE_CODE_METHOD);
TYPE_TARGET_TYPE (type) = to_type;
set_type_self_type (type, self_type);
type->set_fields (args);
type->set_num_fields (nargs);
if (varargs)
type->set_has_varargs (true);
TYPE_LENGTH (type) = 1; /* In practice, this is never needed. */
}
/* A wrapper of TYPE_NAME which calls error if the type is anonymous.
Since GCC PR debug/47510 DWARF provides associated information to detect the
anonymous class linkage name from its typedef.
Parameter TYPE should not yet have CHECK_TYPEDEF applied, this function will
apply it itself. */
const char *
type_name_or_error (struct type *type)
{
struct type *saved_type = type;
const char *name;
struct objfile *objfile;
type = check_typedef (type);
name = type->name ();
if (name != NULL)
return name;
name = saved_type->name ();
objfile = TYPE_OBJFILE (saved_type);
error (_("Invalid anonymous type %s [in module %s], GCC PR debug/47510 bug?"),
name ? name : "<anonymous>",
objfile ? objfile_name (objfile) : "<arch>");
}
/* Lookup a typedef or primitive type named NAME, visible in lexical
block BLOCK. If NOERR is nonzero, return zero if NAME is not
suitably defined. */
struct type *
lookup_typename (const struct language_defn *language,
const char *name,
const struct block *block, int noerr)
{
struct symbol *sym;
sym = lookup_symbol_in_language (name, block, VAR_DOMAIN,
language->la_language, NULL).symbol;
if (sym != NULL && SYMBOL_CLASS (sym) == LOC_TYPEDEF)
return SYMBOL_TYPE (sym);
if (noerr)
return NULL;
error (_("No type named %s."), name);
}
struct type *
lookup_unsigned_typename (const struct language_defn *language,
const char *name)
{
char *uns = (char *) alloca (strlen (name) + 10);
strcpy (uns, "unsigned ");
strcpy (uns + 9, name);
return lookup_typename (language, uns, NULL, 0);
}
struct type *
lookup_signed_typename (const struct language_defn *language, const char *name)
{
struct type *t;
char *uns = (char *) alloca (strlen (name) + 8);
strcpy (uns, "signed ");
strcpy (uns + 7, name);
t = lookup_typename (language, uns, NULL, 1);
/* If we don't find "signed FOO" just try again with plain "FOO". */
if (t != NULL)
return t;
return lookup_typename (language, name, NULL, 0);
}
/* Lookup a structure type named "struct NAME",
visible in lexical block BLOCK. */
struct type *
lookup_struct (const char *name, const struct block *block)
{
struct symbol *sym;
sym = lookup_symbol (name, block, STRUCT_DOMAIN, 0).symbol;
if (sym == NULL)
{
error (_("No struct type named %s."), name);
}
if (SYMBOL_TYPE (sym)->code () != TYPE_CODE_STRUCT)
{
error (_("This context has class, union or enum %s, not a struct."),
name);
}
return (SYMBOL_TYPE (sym));
}
/* Lookup a union type named "union NAME",
visible in lexical block BLOCK. */
struct type *
lookup_union (const char *name, const struct block *block)
{
struct symbol *sym;
struct type *t;
sym = lookup_symbol (name, block, STRUCT_DOMAIN, 0).symbol;
if (sym == NULL)
error (_("No union type named %s."), name);
t = SYMBOL_TYPE (sym);
if (t->code () == TYPE_CODE_UNION)
return t;
/* If we get here, it's not a union. */
error (_("This context has class, struct or enum %s, not a union."),
name);
}
/* Lookup an enum type named "enum NAME",
visible in lexical block BLOCK. */
struct type *
lookup_enum (const char *name, const struct block *block)
{
struct symbol *sym;
sym = lookup_symbol (name, block, STRUCT_DOMAIN, 0).symbol;
if (sym == NULL)
{
error (_("No enum type named %s."), name);
}
if (SYMBOL_TYPE (sym)->code () != TYPE_CODE_ENUM)
{
error (_("This context has class, struct or union %s, not an enum."),
name);
}
return (SYMBOL_TYPE (sym));
}
/* Lookup a template type named "template NAME<TYPE>",
visible in lexical block BLOCK. */
struct type *
lookup_template_type (const char *name, struct type *type,
const struct block *block)
{
struct symbol *sym;
char *nam = (char *)
alloca (strlen (name) + strlen (type->name ()) + 4);
strcpy (nam, name);
strcat (nam, "<");
strcat (nam, type->name ());
strcat (nam, " >"); /* FIXME, extra space still introduced in gcc? */
sym = lookup_symbol (nam, block, VAR_DOMAIN, 0).symbol;
if (sym == NULL)
{
error (_("No template type named %s."), name);
}
if (SYMBOL_TYPE (sym)->code () != TYPE_CODE_STRUCT)
{
error (_("This context has class, union or enum %s, not a struct."),
name);
}
return (SYMBOL_TYPE (sym));
}
/* See gdbtypes.h. */
struct_elt
lookup_struct_elt (struct type *type, const char *name, int noerr)
{
int i;
for (;;)
{
type = check_typedef (type);
if (type->code () != TYPE_CODE_PTR
&& type->code () != TYPE_CODE_REF)
break;
type = TYPE_TARGET_TYPE (type);
}
if (type->code () != TYPE_CODE_STRUCT
&& type->code () != TYPE_CODE_UNION)
{
std::string type_name = type_to_string (type);
error (_("Type %s is not a structure or union type."),
type_name.c_str ());
}
for (i = type->num_fields () - 1; i >= TYPE_N_BASECLASSES (type); i--)
{
const char *t_field_name = TYPE_FIELD_NAME (type, i);
if (t_field_name && (strcmp_iw (t_field_name, name) == 0))
{
return {&type->field (i), TYPE_FIELD_BITPOS (type, i)};
}
else if (!t_field_name || *t_field_name == '\0')
{
struct_elt elt
= lookup_struct_elt (type->field (i).type (), name, 1);
if (elt.field != NULL)
{
elt.offset += TYPE_FIELD_BITPOS (type, i);
return elt;
}
}
}
/* OK, it's not in this class. Recursively check the baseclasses. */
for (i = TYPE_N_BASECLASSES (type) - 1; i >= 0; i--)
{
struct_elt elt = lookup_struct_elt (TYPE_BASECLASS (type, i), name, 1);
if (elt.field != NULL)
return elt;
}
if (noerr)
return {nullptr, 0};
std::string type_name = type_to_string (type);
error (_("Type %s has no component named %s."), type_name.c_str (), name);
}
/* See gdbtypes.h. */
struct type *
lookup_struct_elt_type (struct type *type, const char *name, int noerr)
{
struct_elt elt = lookup_struct_elt (type, name, noerr);
if (elt.field != NULL)
return elt.field->type ();
else
return NULL;
}
/* Store in *MAX the largest number representable by unsigned integer type
TYPE. */
void
get_unsigned_type_max (struct type *type, ULONGEST *max)
{
unsigned int n;
type = check_typedef (type);
gdb_assert (type->code () == TYPE_CODE_INT && type->is_unsigned ());
gdb_assert (TYPE_LENGTH (type) <= sizeof (ULONGEST));
/* Written this way to avoid overflow. */
n = TYPE_LENGTH (type) * TARGET_CHAR_BIT;
*max = ((((ULONGEST) 1 << (n - 1)) - 1) << 1) | 1;
}
/* Store in *MIN, *MAX the smallest and largest numbers representable by
signed integer type TYPE. */
void
get_signed_type_minmax (struct type *type, LONGEST *min, LONGEST *max)
{
unsigned int n;
type = check_typedef (type);
gdb_assert (type->code () == TYPE_CODE_INT && !type->is_unsigned ());
gdb_assert (TYPE_LENGTH (type) <= sizeof (LONGEST));
n = TYPE_LENGTH (type) * TARGET_CHAR_BIT;
*min = -((ULONGEST) 1 << (n - 1));
*max = ((ULONGEST) 1 << (n - 1)) - 1;
}
/* Internal routine called by TYPE_VPTR_FIELDNO to return the value of
cplus_stuff.vptr_fieldno.
cplus_stuff is initialized to cplus_struct_default which does not
set vptr_fieldno to -1 for portability reasons (IWBN to use C99
designated initializers). We cope with that here. */
int
internal_type_vptr_fieldno (struct type *type)
{
type = check_typedef (type);
gdb_assert (type->code () == TYPE_CODE_STRUCT
|| type->code () == TYPE_CODE_UNION);
if (!HAVE_CPLUS_STRUCT (type))
return -1;
return TYPE_RAW_CPLUS_SPECIFIC (type)->vptr_fieldno;
}
/* Set the value of cplus_stuff.vptr_fieldno. */
void
set_type_vptr_fieldno (struct type *type, int fieldno)
{
type = check_typedef (type);
gdb_assert (type->code () == TYPE_CODE_STRUCT
|| type->code () == TYPE_CODE_UNION);
if (!HAVE_CPLUS_STRUCT (type))
ALLOCATE_CPLUS_STRUCT_TYPE (type);
TYPE_RAW_CPLUS_SPECIFIC (type)->vptr_fieldno = fieldno;
}
/* Internal routine called by TYPE_VPTR_BASETYPE to return the value of
cplus_stuff.vptr_basetype. */
struct type *
internal_type_vptr_basetype (struct type *type)
{
type = check_typedef (type);
gdb_assert (type->code () == TYPE_CODE_STRUCT
|| type->code () == TYPE_CODE_UNION);
gdb_assert (TYPE_SPECIFIC_FIELD (type) == TYPE_SPECIFIC_CPLUS_STUFF);
return TYPE_RAW_CPLUS_SPECIFIC (type)->vptr_basetype;
}
/* Set the value of cplus_stuff.vptr_basetype. */
void
set_type_vptr_basetype (struct type *type, struct type *basetype)
{
type = check_typedef (type);
gdb_assert (type->code () == TYPE_CODE_STRUCT
|| type->code () == TYPE_CODE_UNION);
if (!HAVE_CPLUS_STRUCT (type))
ALLOCATE_CPLUS_STRUCT_TYPE (type);
TYPE_RAW_CPLUS_SPECIFIC (type)->vptr_basetype = basetype;
}
/* Lookup the vptr basetype/fieldno values for TYPE.
If found store vptr_basetype in *BASETYPEP if non-NULL, and return
vptr_fieldno. Also, if found and basetype is from the same objfile,
cache the results.
If not found, return -1 and ignore BASETYPEP.
Callers should be aware that in some cases (for example,
the type or one of its baseclasses is a stub type and we are
debugging a .o file, or the compiler uses DWARF-2 and is not GCC),
this function will not be able to find the
virtual function table pointer, and vptr_fieldno will remain -1 and
vptr_basetype will remain NULL or incomplete. */
int
get_vptr_fieldno (struct type *type, struct type **basetypep)
{
type = check_typedef (type);
if (TYPE_VPTR_FIELDNO (type) < 0)
{
int i;
/* We must start at zero in case the first (and only) baseclass
is virtual (and hence we cannot share the table pointer). */
for (i = 0; i < TYPE_N_BASECLASSES (type); i++)
{
struct type *baseclass = check_typedef (TYPE_BASECLASS (type, i));
int fieldno;
struct type *basetype;
fieldno = get_vptr_fieldno (baseclass, &basetype);
if (fieldno >= 0)
{
/* If the type comes from a different objfile we can't cache
it, it may have a different lifetime. PR 2384 */
if (TYPE_OBJFILE (type) == TYPE_OBJFILE (basetype))
{
set_type_vptr_fieldno (type, fieldno);
set_type_vptr_basetype (type, basetype);
}
if (basetypep)
*basetypep = basetype;
return fieldno;
}
}
/* Not found. */
return -1;
}
else
{
if (basetypep)
*basetypep = TYPE_VPTR_BASETYPE (type);
return TYPE_VPTR_FIELDNO (type);
}
}
static void
stub_noname_complaint (void)
{
complaint (_("stub type has NULL name"));
}
/* Return nonzero if TYPE has a DYN_PROP_BYTE_STRIDE dynamic property
attached to it, and that property has a non-constant value. */
static int
array_type_has_dynamic_stride (struct type *type)
{
struct dynamic_prop *prop = type->dyn_prop (DYN_PROP_BYTE_STRIDE);
return (prop != NULL && prop->kind () != PROP_CONST);
}
/* Worker for is_dynamic_type. */
static int
is_dynamic_type_internal (struct type *type, int top_level)
{
type = check_typedef (type);
/* We only want to recognize references at the outermost level. */
if (top_level && type->code () == TYPE_CODE_REF)
type = check_typedef (TYPE_TARGET_TYPE (type));
/* Types that have a dynamic TYPE_DATA_LOCATION are considered
dynamic, even if the type itself is statically defined.
From a user's point of view, this may appear counter-intuitive;
but it makes sense in this context, because the point is to determine
whether any part of the type needs to be resolved before it can
be exploited. */
if (TYPE_DATA_LOCATION (type) != NULL
&& (TYPE_DATA_LOCATION_KIND (type) == PROP_LOCEXPR
|| TYPE_DATA_LOCATION_KIND (type) == PROP_LOCLIST))
return 1;
if (TYPE_ASSOCIATED_PROP (type))
return 1;
if (TYPE_ALLOCATED_PROP (type))
return 1;
struct dynamic_prop *prop = type->dyn_prop (DYN_PROP_VARIANT_PARTS);
if (prop != nullptr && prop->kind () != PROP_TYPE)
return 1;
if (TYPE_HAS_DYNAMIC_LENGTH (type))
return 1;
switch (type->code ())
{
case TYPE_CODE_RANGE:
{
/* A range type is obviously dynamic if it has at least one
dynamic bound. But also consider the range type to be
dynamic when its subtype is dynamic, even if the bounds
of the range type are static. It allows us to assume that
the subtype of a static range type is also static. */
return (!has_static_range (type->bounds ())
|| is_dynamic_type_internal (TYPE_TARGET_TYPE (type), 0));
}
case TYPE_CODE_STRING:
/* Strings are very much like an array of characters, and can be
treated as one here. */
case TYPE_CODE_ARRAY:
{
gdb_assert (type->num_fields () == 1);
/* The array is dynamic if either the bounds are dynamic... */
if (is_dynamic_type_internal (type->index_type (), 0))
return 1;
/* ... or the elements it contains have a dynamic contents... */
if (is_dynamic_type_internal (TYPE_TARGET_TYPE (type), 0))
return 1;
/* ... or if it has a dynamic stride... */
if (array_type_has_dynamic_stride (type))
return 1;
return 0;
}
case TYPE_CODE_STRUCT:
case TYPE_CODE_UNION:
{
int i;
bool is_cplus = HAVE_CPLUS_STRUCT (type);
for (i = 0; i < type->num_fields (); ++i)
{
/* Static fields can be ignored here. */
if (field_is_static (&type->field (i)))
continue;
/* If the field has dynamic type, then so does TYPE. */
if (is_dynamic_type_internal (type->field (i).type (), 0))
return 1;
/* If the field is at a fixed offset, then it is not
dynamic. */
if (TYPE_FIELD_LOC_KIND (type, i) != FIELD_LOC_KIND_DWARF_BLOCK)
continue;
/* Do not consider C++ virtual base types to be dynamic
due to the field's offset being dynamic; these are
handled via other means. */
if (is_cplus && BASETYPE_VIA_VIRTUAL (type, i))
continue;
return 1;
}
}
break;
}
return 0;
}
/* See gdbtypes.h. */
int
is_dynamic_type (struct type *type)
{
return is_dynamic_type_internal (type, 1);
}
static struct type *resolve_dynamic_type_internal
(struct type *type, struct property_addr_info *addr_stack, int top_level);
/* Given a dynamic range type (dyn_range_type) and a stack of
struct property_addr_info elements, return a static version
of that type.
When RESOLVE_P is true then the returned static range is created by
actually evaluating any dynamic properties within the range type, while
when RESOLVE_P is false the returned static range has all of the bounds
and stride information set to undefined. The RESOLVE_P set to false
case will be used when evaluating a dynamic array that is not
allocated, or not associated, i.e. the bounds information might not be
initialized yet. */
static struct type *
resolve_dynamic_range (struct type *dyn_range_type,
struct property_addr_info *addr_stack,
bool resolve_p = true)
{
CORE_ADDR value;
struct type *static_range_type, *static_target_type;
struct dynamic_prop low_bound, high_bound, stride;
gdb_assert (dyn_range_type->code () == TYPE_CODE_RANGE);
const struct dynamic_prop *prop = &dyn_range_type->bounds ()->low;
if (resolve_p && dwarf2_evaluate_property (prop, NULL, addr_stack, &value))
low_bound.set_const_val (value);
else
low_bound.set_undefined ();
prop = &dyn_range_type->bounds ()->high;
if (resolve_p && dwarf2_evaluate_property (prop, NULL, addr_stack, &value))
{
high_bound.set_const_val (value);
if (dyn_range_type->bounds ()->flag_upper_bound_is_count)
high_bound.set_const_val
(low_bound.const_val () + high_bound.const_val () - 1);
}
else
high_bound.set_undefined ();
bool byte_stride_p = dyn_range_type->bounds ()->flag_is_byte_stride;
prop = &dyn_range_type->bounds ()->stride;
if (resolve_p && dwarf2_evaluate_property (prop, NULL, addr_stack, &value))
{
stride.set_const_val (value);
/* If we have a bit stride that is not an exact number of bytes then
I really don't think this is going to work with current GDB, the
array indexing code in GDB seems to be pretty heavily tied to byte
offsets right now. Assuming 8 bits in a byte. */
struct gdbarch *gdbarch = get_type_arch (dyn_range_type);
int unit_size = gdbarch_addressable_memory_unit_size (gdbarch);
if (!byte_stride_p && (value % (unit_size * 8)) != 0)
error (_("bit strides that are not a multiple of the byte size "
"are currently not supported"));
}
else
{
stride.set_undefined ();
byte_stride_p = true;
}
static_target_type
= resolve_dynamic_type_internal (TYPE_TARGET_TYPE (dyn_range_type),
addr_stack, 0);
LONGEST bias = dyn_range_type->bounds ()->bias;
static_range_type = create_range_type_with_stride
(copy_type (dyn_range_type), static_target_type,
&low_bound, &high_bound, bias, &stride, byte_stride_p);
static_range_type->bounds ()->flag_bound_evaluated = 1;
return static_range_type;
}
/* Resolves dynamic bound values of an array or string type TYPE to static
ones. ADDR_STACK is a stack of struct property_addr_info to be used if
needed during the dynamic resolution.
When RESOLVE_P is true then the dynamic properties of TYPE are
evaluated, otherwise the dynamic properties of TYPE are not evaluated,
instead we assume the array is not allocated/associated yet. */
static struct type *
resolve_dynamic_array_or_string (struct type *type,
struct property_addr_info *addr_stack,
bool resolve_p = true)
{
CORE_ADDR value;
struct type *elt_type;
struct type *range_type;
struct type *ary_dim;
struct dynamic_prop *prop;
unsigned int bit_stride = 0;
/* For dynamic type resolution strings can be treated like arrays of
characters. */
gdb_assert (type->code () == TYPE_CODE_ARRAY
|| type->code () == TYPE_CODE_STRING);
type = copy_type (type);
/* Resolve the allocated and associated properties before doing anything
else. If an array is not allocated or not associated then (at least
for Fortran) there is no guarantee that the data to define the upper
bound, lower bound, or stride will be correct. If RESOLVE_P is
already false at this point then this is not the first dimension of
the array and a more outer dimension has already marked this array as
not allocated/associated, as such we just ignore this property. This
is fine as GDB only checks the allocated/associated on the outer most
dimension of the array. */
prop = TYPE_ALLOCATED_PROP (type);
if (prop != NULL && resolve_p
&& dwarf2_evaluate_property (prop, NULL, addr_stack, &value))
{
prop->set_const_val (value);
if (value == 0)
resolve_p = false;
}
prop = TYPE_ASSOCIATED_PROP (type);
if (prop != NULL && resolve_p
&& dwarf2_evaluate_property (prop, NULL, addr_stack, &value))
{
prop->set_const_val (value);
if (value == 0)
resolve_p = false;
}
range_type = check_typedef (type->index_type ());
range_type = resolve_dynamic_range (range_type, addr_stack, resolve_p);
ary_dim = check_typedef (TYPE_TARGET_TYPE (type));
if (ary_dim != NULL && ary_dim->code () == TYPE_CODE_ARRAY)
elt_type = resolve_dynamic_array_or_string (ary_dim, addr_stack, resolve_p);
else
elt_type = TYPE_TARGET_TYPE (type);
prop = type->dyn_prop (DYN_PROP_BYTE_STRIDE);
if (prop != NULL && resolve_p)
{
if (dwarf2_evaluate_property (prop, NULL, addr_stack, &value))
{
type->remove_dyn_prop (DYN_PROP_BYTE_STRIDE);
bit_stride = (unsigned int) (value * 8);
}
else
{
/* Could be a bug in our code, but it could also happen
if the DWARF info is not correct. Issue a warning,
and assume no byte/bit stride (leave bit_stride = 0). */
warning (_("cannot determine array stride for type %s"),
type->name () ? type->name () : "<no name>");
}
}
else
bit_stride = TYPE_FIELD_BITSIZE (type, 0);
return create_array_type_with_stride (type, elt_type, range_type, NULL,
bit_stride);
}
/* Resolve dynamic bounds of members of the union TYPE to static
bounds. ADDR_STACK is a stack of struct property_addr_info
to be used if needed during the dynamic resolution. */
static struct type *
resolve_dynamic_union (struct type *type,
struct property_addr_info *addr_stack)
{
struct type *resolved_type;
int i;
unsigned int max_len = 0;
gdb_assert (type->code () == TYPE_CODE_UNION);
resolved_type = copy_type (type);
resolved_type->set_fields
((struct field *)
TYPE_ALLOC (resolved_type,
resolved_type->num_fields () * sizeof (struct field)));
memcpy (resolved_type->fields (),
type->fields (),
resolved_type->num_fields () * sizeof (struct field));
for (i = 0; i < resolved_type->num_fields (); ++i)
{
struct type *t;
if (field_is_static (&type->field (i)))
continue;
t = resolve_dynamic_type_internal (resolved_type->field (i).type (),
addr_stack, 0);
resolved_type->field (i).set_type (t);
struct type *real_type = check_typedef (t);
if (TYPE_LENGTH (real_type) > max_len)
max_len = TYPE_LENGTH (real_type);
}
TYPE_LENGTH (resolved_type) = max_len;
return resolved_type;
}
/* See gdbtypes.h. */
bool
variant::matches (ULONGEST value, bool is_unsigned) const
{
for (const discriminant_range &range : discriminants)
if (range.contains (value, is_unsigned))
return true;
return false;
}
static void
compute_variant_fields_inner (struct type *type,
struct property_addr_info *addr_stack,
const variant_part &part,
std::vector<bool> &flags);
/* A helper function to determine which variant fields will be active.
This handles both the variant's direct fields, and any variant
parts embedded in this variant. TYPE is the type we're examining.
ADDR_STACK holds information about the concrete object. VARIANT is
the current variant to be handled. FLAGS is where the results are
stored -- this function sets the Nth element in FLAGS if the
corresponding field is enabled. ENABLED is whether this variant is
enabled or not. */
static void
compute_variant_fields_recurse (struct type *type,
struct property_addr_info *addr_stack,
const variant &variant,
std::vector<bool> &flags,
bool enabled)
{
for (int field = variant.first_field; field < variant.last_field; ++field)
flags[field] = enabled;
for (const variant_part &new_part : variant.parts)
{
if (enabled)
compute_variant_fields_inner (type, addr_stack, new_part, flags);
else
{
for (const auto &sub_variant : new_part.variants)
compute_variant_fields_recurse (type, addr_stack, sub_variant,
flags, enabled);
}
}
}
/* A helper function to determine which variant fields will be active.
This evaluates the discriminant, decides which variant (if any) is
active, and then updates FLAGS to reflect which fields should be
available. TYPE is the type we're examining. ADDR_STACK holds
information about the concrete object. VARIANT is the current
variant to be handled. FLAGS is where the results are stored --
this function sets the Nth element in FLAGS if the corresponding
field is enabled. */
static void
compute_variant_fields_inner (struct type *type,
struct property_addr_info *addr_stack,
const variant_part &part,
std::vector<bool> &flags)
{
/* Evaluate the discriminant. */
gdb::optional<ULONGEST> discr_value;
if (part.discriminant_index != -1)
{
int idx = part.discriminant_index;
if (TYPE_FIELD_LOC_KIND (type, idx) != FIELD_LOC_KIND_BITPOS)
error (_("Cannot determine struct field location"
" (invalid location kind)"));
if (addr_stack->valaddr.data () != NULL)
discr_value = unpack_field_as_long (type, addr_stack->valaddr.data (),
idx);
else
{
CORE_ADDR addr = (addr_stack->addr
+ (TYPE_FIELD_BITPOS (type, idx)
/ TARGET_CHAR_BIT));
LONGEST bitsize = TYPE_FIELD_BITSIZE (type, idx);
LONGEST size = bitsize / 8;
if (size == 0)
size = TYPE_LENGTH (type->field (idx).type ());
gdb_byte bits[sizeof (ULONGEST)];
read_memory (addr, bits, size);
LONGEST bitpos = (TYPE_FIELD_BITPOS (type, idx)
% TARGET_CHAR_BIT);
discr_value = unpack_bits_as_long (type->field (idx).type (),
bits, bitpos, bitsize);
}
}
/* Go through each variant and see which applies. */
const variant *default_variant = nullptr;
const variant *applied_variant = nullptr;
for (const auto &variant : part.variants)
{
if (variant.is_default ())
default_variant = &variant;
else if (discr_value.has_value ()
&& variant.matches (*discr_value, part.is_unsigned))
{
applied_variant = &variant;
break;
}
}
if (applied_variant == nullptr)
applied_variant = default_variant;
for (const auto &variant : part.variants)
compute_variant_fields_recurse (type, addr_stack, variant,
flags, applied_variant == &variant);
}
/* Determine which variant fields are available in TYPE. The enabled
fields are stored in RESOLVED_TYPE. ADDR_STACK holds information
about the concrete object. PARTS describes the top-level variant
parts for this type. */
static void
compute_variant_fields (struct type *type,
struct type *resolved_type,
struct property_addr_info *addr_stack,
const gdb::array_view<variant_part> &parts)
{
/* Assume all fields are included by default. */
std::vector<bool> flags (resolved_type->num_fields (), true);
/* Now disable fields based on the variants that control them. */
for (const auto &part : parts)
compute_variant_fields_inner (type, addr_stack, part, flags);
resolved_type->set_num_fields
(std::count (flags.begin (), flags.end (), true));
resolved_type->set_fields
((struct field *)
TYPE_ALLOC (resolved_type,
resolved_type->num_fields () * sizeof (struct field)));
int out = 0;
for (int i = 0; i < type->num_fields (); ++i)
{
if (!flags[i])
continue;
resolved_type->field (out) = type->field (i);
++out;
}
}
/* Resolve dynamic bounds of members of the struct TYPE to static
bounds. ADDR_STACK is a stack of struct property_addr_info to
be used if needed during the dynamic resolution. */
static struct type *
resolve_dynamic_struct (struct type *type,
struct property_addr_info *addr_stack)
{
struct type *resolved_type;
int i;
unsigned resolved_type_bit_length = 0;
gdb_assert (type->code () == TYPE_CODE_STRUCT);
gdb_assert (type->num_fields () > 0);
resolved_type = copy_type (type);
dynamic_prop *variant_prop = resolved_type->dyn_prop (DYN_PROP_VARIANT_PARTS);
if (variant_prop != nullptr && variant_prop->kind () == PROP_VARIANT_PARTS)
{
compute_variant_fields (type, resolved_type, addr_stack,
*variant_prop->variant_parts ());
/* We want to leave the property attached, so that the Rust code
can tell whether the type was originally an enum. */
variant_prop->set_original_type (type);
}
else
{
resolved_type->set_fields
((struct field *)
TYPE_ALLOC (resolved_type,
resolved_type->num_fields () * sizeof (struct field)));
memcpy (resolved_type->fields (),
type->fields (),
resolved_type->num_fields () * sizeof (struct field));
}
for (i = 0; i < resolved_type->num_fields (); ++i)
{
unsigned new_bit_length;
struct property_addr_info pinfo;
if (field_is_static (&resolved_type->field (i)))
continue;
if (TYPE_FIELD_LOC_KIND (resolved_type, i) == FIELD_LOC_KIND_DWARF_BLOCK)
{
struct dwarf2_property_baton baton;
baton.property_type
= lookup_pointer_type (resolved_type->field (i).type ());
baton.locexpr = *TYPE_FIELD_DWARF_BLOCK (resolved_type, i);
struct dynamic_prop prop;
prop.set_locexpr (&baton);
CORE_ADDR addr;
if (dwarf2_evaluate_property (&prop, nullptr, addr_stack, &addr,
true))
SET_FIELD_BITPOS (resolved_type->field (i),
TARGET_CHAR_BIT * (addr - addr_stack->addr));
}
/* As we know this field is not a static field, the field's
field_loc_kind should be FIELD_LOC_KIND_BITPOS. Verify
this is the case, but only trigger a simple error rather
than an internal error if that fails. While failing
that verification indicates a bug in our code, the error
is not severe enough to suggest to the user he stops
his debugging session because of it. */
if (TYPE_FIELD_LOC_KIND (resolved_type, i) != FIELD_LOC_KIND_BITPOS)
error (_("Cannot determine struct field location"
" (invalid location kind)"));
pinfo.type = check_typedef (resolved_type->field (i).type ());
pinfo.valaddr = addr_stack->valaddr;
pinfo.addr
= (addr_stack->addr
+ (TYPE_FIELD_BITPOS (resolved_type, i) / TARGET_CHAR_BIT));
pinfo.next = addr_stack;
resolved_type->field (i).set_type
(resolve_dynamic_type_internal (resolved_type->field (i).type (),
&pinfo, 0));
gdb_assert (TYPE_FIELD_LOC_KIND (resolved_type, i)
== FIELD_LOC_KIND_BITPOS);
new_bit_length = TYPE_FIELD_BITPOS (resolved_type, i);
if (TYPE_FIELD_BITSIZE (resolved_type, i) != 0)
new_bit_length += TYPE_FIELD_BITSIZE (resolved_type, i);
else
{
struct type *real_type
= check_typedef (resolved_type->field (i).type ());
new_bit_length += (TYPE_LENGTH (real_type) * TARGET_CHAR_BIT);
}
/* Normally, we would use the position and size of the last field
to determine the size of the enclosing structure. But GCC seems
to be encoding the position of some fields incorrectly when
the struct contains a dynamic field that is not placed last.
So we compute the struct size based on the field that has
the highest position + size - probably the best we can do. */
if (new_bit_length > resolved_type_bit_length)
resolved_type_bit_length = new_bit_length;
}
/* The length of a type won't change for fortran, but it does for C and Ada.
For fortran the size of dynamic fields might change over time but not the
type length of the structure. If we adapt it, we run into problems
when calculating the element offset for arrays of structs. */
if (current_language->la_language != language_fortran)
TYPE_LENGTH (resolved_type)
= (resolved_type_bit_length + TARGET_CHAR_BIT - 1) / TARGET_CHAR_BIT;
/* The Ada language uses this field as a cache for static fixed types: reset
it as RESOLVED_TYPE must have its own static fixed type. */
TYPE_TARGET_TYPE (resolved_type) = NULL;
return resolved_type;
}
/* Worker for resolved_dynamic_type. */
static struct type *
resolve_dynamic_type_internal (struct type *type,
struct property_addr_info *addr_stack,
int top_level)
{
struct type *real_type = check_typedef (type);
struct type *resolved_type = nullptr;
struct dynamic_prop *prop;
CORE_ADDR value;
if (!is_dynamic_type_internal (real_type, top_level))
return type;
gdb::optional<CORE_ADDR> type_length;
prop = TYPE_DYNAMIC_LENGTH (type);
if (prop != NULL
&& dwarf2_evaluate_property (prop, NULL, addr_stack, &value))
type_length = value;
if (type->code () == TYPE_CODE_TYPEDEF)
{
resolved_type = copy_type (type);
TYPE_TARGET_TYPE (resolved_type)
= resolve_dynamic_type_internal (TYPE_TARGET_TYPE (type), addr_stack,
top_level);
}
else
{
/* Before trying to resolve TYPE, make sure it is not a stub. */
type = real_type;
switch (type->code ())
{
case TYPE_CODE_REF:
{
struct property_addr_info pinfo;
pinfo.type = check_typedef (TYPE_TARGET_TYPE (type));
pinfo.valaddr = {};
if (addr_stack->valaddr.data () != NULL)
pinfo.addr = extract_typed_address (addr_stack->valaddr.data (),
type);
else
pinfo.addr = read_memory_typed_address (addr_stack->addr, type);
pinfo.next = addr_stack;
resolved_type = copy_type (type);
TYPE_TARGET_TYPE (resolved_type)
= resolve_dynamic_type_internal (TYPE_TARGET_TYPE (type),
&pinfo, top_level);
break;
}
case TYPE_CODE_STRING:
/* Strings are very much like an array of characters, and can be
treated as one here. */
case TYPE_CODE_ARRAY:
resolved_type = resolve_dynamic_array_or_string (type, addr_stack);
break;
case TYPE_CODE_RANGE:
resolved_type = resolve_dynamic_range (type, addr_stack);
break;
case TYPE_CODE_UNION:
resolved_type = resolve_dynamic_union (type, addr_stack);
break;
case TYPE_CODE_STRUCT:
resolved_type = resolve_dynamic_struct (type, addr_stack);
break;
}
}
if (resolved_type == nullptr)
return type;
if (type_length.has_value ())
{
TYPE_LENGTH (resolved_type) = *type_length;
resolved_type->remove_dyn_prop (DYN_PROP_BYTE_SIZE);
}
/* Resolve data_location attribute. */
prop = TYPE_DATA_LOCATION (resolved_type);
if (prop != NULL
&& dwarf2_evaluate_property (prop, NULL, addr_stack, &value))
{
/* Start of Fortran hack. See comment in f-lang.h for what is going
on here.*/
if (current_language->la_language == language_fortran
&& resolved_type->code () == TYPE_CODE_ARRAY)
value = fortran_adjust_dynamic_array_base_address_hack (resolved_type,
value);
/* End of Fortran hack. */
prop->set_const_val (value);
}
return resolved_type;
}
/* See gdbtypes.h */
struct type *
resolve_dynamic_type (struct type *type,
gdb::array_view<const gdb_byte> valaddr,
CORE_ADDR addr)
{
struct property_addr_info pinfo
= {check_typedef (type), valaddr, addr, NULL};
return resolve_dynamic_type_internal (type, &pinfo, 1);
}
/* See gdbtypes.h */
dynamic_prop *
type::dyn_prop (dynamic_prop_node_kind prop_kind) const
{
dynamic_prop_list *node = this->main_type->dyn_prop_list;
while (node != NULL)
{
if (node->prop_kind == prop_kind)
return &node->prop;
node = node->next;
}
return NULL;
}
/* See gdbtypes.h */
void
type::add_dyn_prop (dynamic_prop_node_kind prop_kind, dynamic_prop prop)
{
struct dynamic_prop_list *temp;
gdb_assert (TYPE_OBJFILE_OWNED (this));
temp = XOBNEW (&TYPE_OBJFILE (this)->objfile_obstack,
struct dynamic_prop_list);
temp->prop_kind = prop_kind;
temp->prop = prop;
temp->next = this->main_type->dyn_prop_list;
this->main_type->dyn_prop_list = temp;
}
/* See gdbtypes.h. */
void
type::remove_dyn_prop (dynamic_prop_node_kind kind)
{
struct dynamic_prop_list *prev_node, *curr_node;
curr_node = this->main_type->dyn_prop_list;
prev_node = NULL;
while (NULL != curr_node)
{
if (curr_node->prop_kind == kind)
{
/* Update the linked list but don't free anything.
The property was allocated on objstack and it is not known
if we are on top of it. Nevertheless, everything is released
when the complete objstack is freed. */
if (NULL == prev_node)
this->main_type->dyn_prop_list = curr_node->next;
else
prev_node->next = curr_node->next;
return;
}
prev_node = curr_node;
curr_node = curr_node->next;
}
}
/* Find the real type of TYPE. This function returns the real type,
after removing all layers of typedefs, and completing opaque or stub
types. Completion changes the TYPE argument, but stripping of
typedefs does not.
Instance flags (e.g. const/volatile) are preserved as typedefs are
stripped. If necessary a new qualified form of the underlying type
is created.
NOTE: This will return a typedef if TYPE_TARGET_TYPE for the typedef has
not been computed and we're either in the middle of reading symbols, or
there was no name for the typedef in the debug info.
NOTE: Lookup of opaque types can throw errors for invalid symbol files.
QUITs in the symbol reading code can also throw.
Thus this function can throw an exception.
If TYPE is a TYPE_CODE_TYPEDEF, its length is updated to the length of
the target type.
If this is a stubbed struct (i.e. declared as struct foo *), see if
we can find a full definition in some other file. If so, copy this
definition, so we can use it in future. There used to be a comment
(but not any code) that if we don't find a full definition, we'd
set a flag so we don't spend time in the future checking the same
type. That would be a mistake, though--we might load in more
symbols which contain a full definition for the type. */
struct type *
check_typedef (struct type *type)
{
struct type *orig_type = type;
gdb_assert (type);
/* While we're removing typedefs, we don't want to lose qualifiers.
E.g., const/volatile. */
type_instance_flags instance_flags = type->instance_flags ();
while (type->code () == TYPE_CODE_TYPEDEF)
{
if (!TYPE_TARGET_TYPE (type))
{
const char *name;
struct symbol *sym;
/* It is dangerous to call lookup_symbol if we are currently
reading a symtab. Infinite recursion is one danger. */
if (currently_reading_symtab)
return make_qualified_type (type, instance_flags, NULL);
name = type->name ();
/* FIXME: shouldn't we look in STRUCT_DOMAIN and/or
VAR_DOMAIN as appropriate? */
if (name == NULL)
{
stub_noname_complaint ();
return make_qualified_type (type, instance_flags, NULL);
}
sym = lookup_symbol (name, 0, STRUCT_DOMAIN, 0).symbol;
if (sym)
TYPE_TARGET_TYPE (type) = SYMBOL_TYPE (sym);
else /* TYPE_CODE_UNDEF */
TYPE_TARGET_TYPE (type) = alloc_type_arch (get_type_arch (type));
}
type = TYPE_TARGET_TYPE (type);
/* Preserve the instance flags as we traverse down the typedef chain.
Handling address spaces/classes is nasty, what do we do if there's a
conflict?
E.g., what if an outer typedef marks the type as class_1 and an inner
typedef marks the type as class_2?
This is the wrong place to do such error checking. We leave it to
the code that created the typedef in the first place to flag the
error. We just pick the outer address space (akin to letting the
outer cast in a chain of casting win), instead of assuming
"it can't happen". */
{
const type_instance_flags ALL_SPACES
= (TYPE_INSTANCE_FLAG_CODE_SPACE
| TYPE_INSTANCE_FLAG_DATA_SPACE);
const type_instance_flags ALL_CLASSES
= TYPE_INSTANCE_FLAG_ADDRESS_CLASS_ALL;
type_instance_flags new_instance_flags = type->instance_flags ();
/* Treat code vs data spaces and address classes separately. */
if ((instance_flags & ALL_SPACES) != 0)
new_instance_flags &= ~ALL_SPACES;
if ((instance_flags & ALL_CLASSES) != 0)
new_instance_flags &= ~ALL_CLASSES;
instance_flags |= new_instance_flags;
}
}
/* If this is a struct/class/union with no fields, then check
whether a full definition exists somewhere else. This is for
systems where a type definition with no fields is issued for such
types, instead of identifying them as stub types in the first
place. */
if (TYPE_IS_OPAQUE (type)
&& opaque_type_resolution
&& !currently_reading_symtab)
{
const char *name = type->name ();
struct type *newtype;
if (name == NULL)
{
stub_noname_complaint ();
return make_qualified_type (type, instance_flags, NULL);
}
newtype = lookup_transparent_type (name);
if (newtype)
{
/* If the resolved type and the stub are in the same
objfile, then replace the stub type with the real deal.
But if they're in separate objfiles, leave the stub
alone; we'll just look up the transparent type every time
we call check_typedef. We can't create pointers between
types allocated to different objfiles, since they may
have different lifetimes. Trying to copy NEWTYPE over to
TYPE's objfile is pointless, too, since you'll have to
move over any other types NEWTYPE refers to, which could
be an unbounded amount of stuff. */
if (TYPE_OBJFILE (newtype) == TYPE_OBJFILE (type))
type = make_qualified_type (newtype, type->instance_flags (), type);
else
type = newtype;
}
}
/* Otherwise, rely on the stub flag being set for opaque/stubbed
types. */
else if (type->is_stub () && !currently_reading_symtab)
{
const char *name = type->name ();
/* FIXME: shouldn't we look in STRUCT_DOMAIN and/or VAR_DOMAIN
as appropriate? */
struct symbol *sym;
if (name == NULL)
{
stub_noname_complaint ();
return make_qualified_type (type, instance_flags, NULL);
}
sym = lookup_symbol (name, 0, STRUCT_DOMAIN, 0).symbol;
if (sym)
{
/* Same as above for opaque types, we can replace the stub
with the complete type only if they are in the same
objfile. */
if (TYPE_OBJFILE (SYMBOL_TYPE (sym)) == TYPE_OBJFILE (type))
type = make_qualified_type (SYMBOL_TYPE (sym),
type->instance_flags (), type);
else
type = SYMBOL_TYPE (sym);
}
}
if (type->target_is_stub ())
{
struct type *target_type = check_typedef (TYPE_TARGET_TYPE (type));
if (target_type->is_stub () || target_type->target_is_stub ())
{
/* Nothing we can do. */
}
else if (type->code () == TYPE_CODE_RANGE)
{
TYPE_LENGTH (type) = TYPE_LENGTH (target_type);
type->set_target_is_stub (false);
}
else if (type->code () == TYPE_CODE_ARRAY
&& update_static_array_size (type))
type->set_target_is_stub (false);
}
type = make_qualified_type (type, instance_flags, NULL);
/* Cache TYPE_LENGTH for future use. */
TYPE_LENGTH (orig_type) = TYPE_LENGTH (type);
return type;
}
/* Parse a type expression in the string [P..P+LENGTH). If an error
occurs, silently return a void type. */
static struct type *
safe_parse_type (struct gdbarch *gdbarch, const char *p, int length)
{
struct ui_file *saved_gdb_stderr;
struct type *type = NULL; /* Initialize to keep gcc happy. */
/* Suppress error messages. */
saved_gdb_stderr = gdb_stderr;
gdb_stderr = &null_stream;
/* Call parse_and_eval_type() without fear of longjmp()s. */
try
{
type = parse_and_eval_type (p, length);
}
catch (const gdb_exception_error &except)
{
type = builtin_type (gdbarch)->builtin_void;
}
/* Stop suppressing error messages. */
gdb_stderr = saved_gdb_stderr;
return type;
}
/* Ugly hack to convert method stubs into method types.
He ain't kiddin'. This demangles the name of the method into a
string including argument types, parses out each argument type,
generates a string casting a zero to that type, evaluates the
string, and stuffs the resulting type into an argtype vector!!!
Then it knows the type of the whole function (including argument
types for overloading), which info used to be in the stab's but was
removed to hack back the space required for them. */
static void
check_stub_method (struct type *type, int method_id, int signature_id)
{
struct gdbarch *gdbarch = get_type_arch (type);
struct fn_field *f;
char *mangled_name = gdb_mangle_name (type, method_id, signature_id);
char *demangled_name = gdb_demangle (mangled_name,
DMGL_PARAMS | DMGL_ANSI);
char *argtypetext, *p;
int depth = 0, argcount = 1;
struct field *argtypes;
struct type *mtype;
/* Make sure we got back a function string that we can use. */
if (demangled_name)
p = strchr (demangled_name, '(');
else
p = NULL;
if (demangled_name == NULL || p == NULL)
error (_("Internal: Cannot demangle mangled name `%s'."),
mangled_name);
/* Now, read in the parameters that define this type. */
p += 1;
argtypetext = p;
while (*p)
{
if (*p == '(' || *p == '<')
{
depth += 1;
}
else if (*p == ')' || *p == '>')
{
depth -= 1;
}
else if (*p == ',' && depth == 0)
{
argcount += 1;
}
p += 1;
}
/* If we read one argument and it was ``void'', don't count it. */
if (startswith (argtypetext, "(void)"))
argcount -= 1;
/* We need one extra slot, for the THIS pointer. */
argtypes = (struct field *)
TYPE_ALLOC (type, (argcount + 1) * sizeof (struct field));
p = argtypetext;
/* Add THIS pointer for non-static methods. */
f = TYPE_FN_FIELDLIST1 (type, method_id);
if (TYPE_FN_FIELD_STATIC_P (f, signature_id))
argcount = 0;
else
{
argtypes[0].set_type (lookup_pointer_type (type));
argcount = 1;
}
if (*p != ')') /* () means no args, skip while. */
{
depth = 0;
while (*p)
{
if (depth <= 0 && (*p == ',' || *p == ')'))
{
/* Avoid parsing of ellipsis, they will be handled below.
Also avoid ``void'' as above. */
if (strncmp (argtypetext, "...", p - argtypetext) != 0
&& strncmp (argtypetext, "void", p - argtypetext) != 0)
{
argtypes[argcount].set_type
(safe_parse_type (gdbarch, argtypetext, p - argtypetext));
argcount += 1;
}
argtypetext = p + 1;
}
if (*p == '(' || *p == '<')
{
depth += 1;
}
else if (*p == ')' || *p == '>')
{
depth -= 1;
}
p += 1;
}
}
TYPE_FN_FIELD_PHYSNAME (f, signature_id) = mangled_name;
/* Now update the old "stub" type into a real type. */
mtype = TYPE_FN_FIELD_TYPE (f, signature_id);
/* MTYPE may currently be a function (TYPE_CODE_FUNC).
We want a method (TYPE_CODE_METHOD). */
smash_to_method_type (mtype, type, TYPE_TARGET_TYPE (mtype),
argtypes, argcount, p[-2] == '.');
mtype->set_is_stub (false);
TYPE_FN_FIELD_STUB (f, signature_id) = 0;
xfree (demangled_name);
}
/* This is the external interface to check_stub_method, above. This
function unstubs all of the signatures for TYPE's METHOD_ID method
name. After calling this function TYPE_FN_FIELD_STUB will be
cleared for each signature and TYPE_FN_FIELDLIST_NAME will be
correct.
This function unfortunately can not die until stabs do. */
void
check_stub_method_group (struct type *type, int method_id)
{
int len = TYPE_FN_FIELDLIST_LENGTH (type, method_id);
struct fn_field *f = TYPE_FN_FIELDLIST1 (type, method_id);
for (int j = 0; j < len; j++)
{
if (TYPE_FN_FIELD_STUB (f, j))
check_stub_method (type, method_id, j);
}
}
/* Ensure it is in .rodata (if available) by working around GCC PR 44690. */
const struct cplus_struct_type cplus_struct_default = { };
void
allocate_cplus_struct_type (struct type *type)
{
if (HAVE_CPLUS_STRUCT (type))
/* Structure was already allocated. Nothing more to do. */
return;
TYPE_SPECIFIC_FIELD (type) = TYPE_SPECIFIC_CPLUS_STUFF;
TYPE_RAW_CPLUS_SPECIFIC (type) = (struct cplus_struct_type *)
TYPE_ALLOC (type, sizeof (struct cplus_struct_type));
*(TYPE_RAW_CPLUS_SPECIFIC (type)) = cplus_struct_default;
set_type_vptr_fieldno (type, -1);
}
const struct gnat_aux_type gnat_aux_default =
{ NULL };
/* Set the TYPE's type-specific kind to TYPE_SPECIFIC_GNAT_STUFF,
and allocate the associated gnat-specific data. The gnat-specific
data is also initialized to gnat_aux_default. */
void
allocate_gnat_aux_type (struct type *type)
{
TYPE_SPECIFIC_FIELD (type) = TYPE_SPECIFIC_GNAT_STUFF;
TYPE_GNAT_SPECIFIC (type) = (struct gnat_aux_type *)
TYPE_ALLOC (type, sizeof (struct gnat_aux_type));
*(TYPE_GNAT_SPECIFIC (type)) = gnat_aux_default;
}
/* Helper function to initialize a newly allocated type. Set type code
to CODE and initialize the type-specific fields accordingly. */
static void
set_type_code (struct type *type, enum type_code code)
{
type->set_code (code);
switch (code)
{
case TYPE_CODE_STRUCT:
case TYPE_CODE_UNION:
case TYPE_CODE_NAMESPACE:
INIT_CPLUS_SPECIFIC (type);
break;
case TYPE_CODE_FLT:
TYPE_SPECIFIC_FIELD (type) = TYPE_SPECIFIC_FLOATFORMAT;
break;
case TYPE_CODE_FUNC:
INIT_FUNC_SPECIFIC (type);
break;
case TYPE_CODE_FIXED_POINT:
INIT_FIXED_POINT_SPECIFIC (type);
break;
}
}
/* Helper function to verify floating-point format and size.
BIT is the type size in bits; if BIT equals -1, the size is
determined by the floatformat. Returns size to be used. */
static int
verify_floatformat (int bit, const struct floatformat *floatformat)
{
gdb_assert (floatformat != NULL);
if (bit == -1)
bit = floatformat->totalsize;
gdb_assert (bit >= 0);
gdb_assert (bit >= floatformat->totalsize);
return bit;
}
/* Return the floating-point format for a floating-point variable of
type TYPE. */
const struct floatformat *
floatformat_from_type (const struct type *type)
{
gdb_assert (type->code () == TYPE_CODE_FLT);
gdb_assert (TYPE_FLOATFORMAT (type));
return TYPE_FLOATFORMAT (type);
}
/* Helper function to initialize the standard scalar types.
If NAME is non-NULL, then it is used to initialize the type name.
Note that NAME is not copied; it is required to have a lifetime at
least as long as OBJFILE. */
struct type *
init_type (struct objfile *objfile, enum type_code code, int bit,
const char *name)
{
struct type *type;
type = alloc_type (objfile);
set_type_code (type, code);
gdb_assert ((bit % TARGET_CHAR_BIT) == 0);
TYPE_LENGTH (type) = bit / TARGET_CHAR_BIT;
type->set_name (name);
return type;
}
/* Allocate a TYPE_CODE_ERROR type structure associated with OBJFILE,
to use with variables that have no debug info. NAME is the type
name. */
static struct type *
init_nodebug_var_type (struct objfile *objfile, const char *name)
{
return init_type (objfile, TYPE_CODE_ERROR, 0, name);
}
/* Allocate a TYPE_CODE_INT type structure associated with OBJFILE.
BIT is the type size in bits. If UNSIGNED_P is non-zero, set
the type's TYPE_UNSIGNED flag. NAME is the type name. */
struct type *
init_integer_type (struct objfile *objfile,
int bit, int unsigned_p, const char *name)
{
struct type *t;
t = init_type (objfile, TYPE_CODE_INT, bit, name);
if (unsigned_p)
t->set_is_unsigned (true);
TYPE_SPECIFIC_FIELD (t) = TYPE_SPECIFIC_INT;
TYPE_MAIN_TYPE (t)->type_specific.int_stuff.bit_size = bit;
TYPE_MAIN_TYPE (t)->type_specific.int_stuff.bit_offset = 0;
return t;
}
/* Allocate a TYPE_CODE_CHAR type structure associated with OBJFILE.
BIT is the type size in bits. If UNSIGNED_P is non-zero, set
the type's TYPE_UNSIGNED flag. NAME is the type name. */
struct type *
init_character_type (struct objfile *objfile,
int bit, int unsigned_p, const char *name)
{
struct type *t;
t = init_type (objfile, TYPE_CODE_CHAR, bit, name);
if (unsigned_p)
t->set_is_unsigned (true);
return t;
}
/* Allocate a TYPE_CODE_BOOL type structure associated with OBJFILE.
BIT is the type size in bits. If UNSIGNED_P is non-zero, set
the type's TYPE_UNSIGNED flag. NAME is the type name. */
struct type *
init_boolean_type (struct objfile *objfile,
int bit, int unsigned_p, const char *name)
{
struct type *t;
t = init_type (objfile, TYPE_CODE_BOOL, bit, name);
if (unsigned_p)
t->set_is_unsigned (true);
TYPE_SPECIFIC_FIELD (t) = TYPE_SPECIFIC_INT;
TYPE_MAIN_TYPE (t)->type_specific.int_stuff.bit_size = bit;
TYPE_MAIN_TYPE (t)->type_specific.int_stuff.bit_offset = 0;
return t;
}
/* Allocate a TYPE_CODE_FLT type structure associated with OBJFILE.
BIT is the type size in bits; if BIT equals -1, the size is
determined by the floatformat. NAME is the type name. Set the
TYPE_FLOATFORMAT from FLOATFORMATS. BYTE_ORDER is the byte order
to use. If it is BFD_ENDIAN_UNKNOWN (the default), then the byte
order of the objfile's architecture is used. */
struct type *
init_float_type (struct objfile *objfile,
int bit, const char *name,
const struct floatformat **floatformats,
enum bfd_endian byte_order)
{
if (byte_order == BFD_ENDIAN_UNKNOWN)
{
struct gdbarch *gdbarch = objfile->arch ();
byte_order = gdbarch_byte_order (gdbarch);
}
const struct floatformat *fmt = floatformats[byte_order];
struct type *t;
bit = verify_floatformat (bit, fmt);
t = init_type (objfile, TYPE_CODE_FLT, bit, name);
TYPE_FLOATFORMAT (t) = fmt;
return t;
}
/* Allocate a TYPE_CODE_DECFLOAT type structure associated with OBJFILE.
BIT is the type size in bits. NAME is the type name. */
struct type *
init_decfloat_type (struct objfile *objfile, int bit, const char *name)
{
struct type *t;
t = init_type (objfile, TYPE_CODE_DECFLOAT, bit, name);
return t;
}
/* Allocate a TYPE_CODE_COMPLEX type structure. NAME is the type
name. TARGET_TYPE is the component type. */
struct type *
init_complex_type (const char *name, struct type *target_type)
{
struct type *t;
gdb_assert (target_type->code () == TYPE_CODE_INT
|| target_type->code () == TYPE_CODE_FLT);
if (TYPE_MAIN_TYPE (target_type)->flds_bnds.complex_type == nullptr)
{
if (name == nullptr && target_type->name () != nullptr)
{
char *new_name
= (char *) TYPE_ALLOC (target_type,
strlen (target_type->name ())
+ strlen ("_Complex ") + 1);
strcpy (new_name, "_Complex ");
strcat (new_name, target_type->name ());
name = new_name;
}
t = alloc_type_copy (target_type);
set_type_code (t, TYPE_CODE_COMPLEX);
TYPE_LENGTH (t) = 2 * TYPE_LENGTH (target_type);
t->set_name (name);
TYPE_TARGET_TYPE (t) = target_type;
TYPE_MAIN_TYPE (target_type)->flds_bnds.complex_type = t;
}
return TYPE_MAIN_TYPE (target_type)->flds_bnds.complex_type;
}
/* Allocate a TYPE_CODE_PTR type structure associated with OBJFILE.
BIT is the pointer type size in bits. NAME is the type name.
TARGET_TYPE is the pointer target type. Always sets the pointer type's
TYPE_UNSIGNED flag. */
struct type *
init_pointer_type (struct objfile *objfile,
int bit, const char *name, struct type *target_type)
{
struct type *t;
t = init_type (objfile, TYPE_CODE_PTR, bit, name);
TYPE_TARGET_TYPE (t) = target_type;
t->set_is_unsigned (true);
return t;
}
/* Allocate a TYPE_CODE_FIXED_POINT type structure associated with OBJFILE.
BIT is the pointer type size in bits.
UNSIGNED_P should be nonzero if the type is unsigned.
NAME is the type name. */
struct type *
init_fixed_point_type (struct objfile *objfile,
int bit, int unsigned_p, const char *name)
{
struct type *t;
t = init_type (objfile, TYPE_CODE_FIXED_POINT, bit, name);
if (unsigned_p)
t->set_is_unsigned (true);
return t;
}
/* See gdbtypes.h. */
unsigned
type_raw_align (struct type *type)
{
if (type->align_log2 != 0)
return 1 << (type->align_log2 - 1);
return 0;
}
/* See gdbtypes.h. */
unsigned
type_align (struct type *type)
{
/* Check alignment provided in the debug information. */
unsigned raw_align = type_raw_align (type);
if (raw_align != 0)
return raw_align;
/* Allow the architecture to provide an alignment. */
struct gdbarch *arch = get_type_arch (type);
ULONGEST align = gdbarch_type_align (arch, type);
if (align != 0)
return align;
switch (type->code ())
{
case TYPE_CODE_PTR:
case TYPE_CODE_FUNC:
case TYPE_CODE_FLAGS:
case TYPE_CODE_INT:
case TYPE_CODE_RANGE:
case TYPE_CODE_FLT:
case TYPE_CODE_ENUM:
case TYPE_CODE_REF:
case TYPE_CODE_RVALUE_REF:
case TYPE_CODE_CHAR:
case TYPE_CODE_BOOL:
case TYPE_CODE_DECFLOAT:
case TYPE_CODE_METHODPTR:
case TYPE_CODE_MEMBERPTR:
align = type_length_units (check_typedef (type));
break;
case TYPE_CODE_ARRAY:
case TYPE_CODE_COMPLEX:
case TYPE_CODE_TYPEDEF:
align = type_align (TYPE_TARGET_TYPE (type));
break;
case TYPE_CODE_STRUCT:
case TYPE_CODE_UNION:
{
int number_of_non_static_fields = 0;
for (unsigned i = 0; i < type->num_fields (); ++i)
{
if (!field_is_static (&type->field (i)))
{
number_of_non_static_fields++;
ULONGEST f_align = type_align (type->field (i).type ());
if (f_align == 0)
{
/* Don't pretend we know something we don't. */
align = 0;
break;
}
if (f_align > align)
align = f_align;
}
}
/* A struct with no fields, or with only static fields has an
alignment of 1. */
if (number_of_non_static_fields == 0)
align = 1;
}
break;
case TYPE_CODE_SET:
case TYPE_CODE_STRING:
/* Not sure what to do here, and these can't appear in C or C++
anyway. */
break;
case TYPE_CODE_VOID:
align = 1;
break;
case TYPE_CODE_ERROR:
case TYPE_CODE_METHOD:
default:
break;
}
if ((align & (align - 1)) != 0)
{
/* Not a power of 2, so pass. */
align = 0;
}
return align;
}
/* See gdbtypes.h. */
bool
set_type_align (struct type *type, ULONGEST align)
{
/* Must be a power of 2. Zero is ok. */
gdb_assert ((align & (align - 1)) == 0);
unsigned result = 0;
while (align != 0)
{
++result;
align >>= 1;
}
if (result >= (1 << TYPE_ALIGN_BITS))
return false;
type->align_log2 = result;
return true;
}
/* Queries on types. */
int
can_dereference (struct type *t)
{
/* FIXME: Should we return true for references as well as
pointers? */
t = check_typedef (t);
return
(t != NULL
&& t->code () == TYPE_CODE_PTR
&& TYPE_TARGET_TYPE (t)->code () != TYPE_CODE_VOID);
}
int
is_integral_type (struct type *t)
{
t = check_typedef (t);
return
((t != NULL)
&& !is_fixed_point_type (t)
&& ((t->code () == TYPE_CODE_INT)
|| (t->code () == TYPE_CODE_ENUM)
|| (t->code () == TYPE_CODE_FLAGS)
|| (t->code () == TYPE_CODE_CHAR)
|| (t->code () == TYPE_CODE_RANGE)
|| (t->code () == TYPE_CODE_BOOL)));
}
int
is_floating_type (struct type *t)
{
t = check_typedef (t);
return
((t != NULL)
&& ((t->code () == TYPE_CODE_FLT)
|| (t->code () == TYPE_CODE_DECFLOAT)));
}
/* Return true if TYPE is scalar. */
int
is_scalar_type (struct type *type)
{
type = check_typedef (type);
if (is_fixed_point_type (type))
return 0; /* Implemented as a scalar, but more like a floating point. */
switch (type->code ())
{
case TYPE_CODE_ARRAY:
case TYPE_CODE_STRUCT:
case TYPE_CODE_UNION:
case TYPE_CODE_SET:
case TYPE_CODE_STRING:
return 0;
default:
return 1;
}
}
/* Return true if T is scalar, or a composite type which in practice has
the memory layout of a scalar type. E.g., an array or struct with only
one scalar element inside it, or a union with only scalar elements. */
int
is_scalar_type_recursive (struct type *t)
{
t = check_typedef (t);
if (is_scalar_type (t))
return 1;
/* Are we dealing with an array or string of known dimensions? */
else if ((t->code () == TYPE_CODE_ARRAY
|| t->code () == TYPE_CODE_STRING) && t->num_fields () == 1
&& t->index_type ()->code () == TYPE_CODE_RANGE)
{
LONGEST low_bound, high_bound;
struct type *elt_type = check_typedef (TYPE_TARGET_TYPE (t));
if (get_discrete_bounds (t->index_type (), &low_bound, &high_bound))
return (high_bound == low_bound
&& is_scalar_type_recursive (elt_type));
else
return 0;
}
/* Are we dealing with a struct with one element? */
else if (t->code () == TYPE_CODE_STRUCT && t->num_fields () == 1)
return is_scalar_type_recursive (t->field (0).type ());
else if (t->code () == TYPE_CODE_UNION)
{
int i, n = t->num_fields ();
/* If all elements of the union are scalar, then the union is scalar. */
for (i = 0; i < n; i++)
if (!is_scalar_type_recursive (t->field (i).type ()))
return 0;
return 1;
}
return 0;
}
/* Return true is T is a class or a union. False otherwise. */
int
class_or_union_p (const struct type *t)
{
return (t->code () == TYPE_CODE_STRUCT
|| t->code () == TYPE_CODE_UNION);
}
/* A helper function which returns true if types A and B represent the
"same" class type. This is true if the types have the same main
type, or the same name. */
int
class_types_same_p (const struct type *a, const struct type *b)
{
return (TYPE_MAIN_TYPE (a) == TYPE_MAIN_TYPE (b)
|| (a->name () && b->name ()
&& !strcmp (a->name (), b->name ())));
}
/* If BASE is an ancestor of DCLASS return the distance between them.
otherwise return -1;
eg:
class A {};
class B: public A {};
class C: public B {};
class D: C {};
distance_to_ancestor (A, A, 0) = 0
distance_to_ancestor (A, B, 0) = 1
distance_to_ancestor (A, C, 0) = 2
distance_to_ancestor (A, D, 0) = 3
If PUBLIC is 1 then only public ancestors are considered,
and the function returns the distance only if BASE is a public ancestor
of DCLASS.
Eg:
distance_to_ancestor (A, D, 1) = -1. */
static int
distance_to_ancestor (struct type *base, struct type *dclass, int is_public)
{
int i;
int d;
base = check_typedef (base);
dclass = check_typedef (dclass);
if (class_types_same_p (base, dclass))
return 0;
for (i = 0; i < TYPE_N_BASECLASSES (dclass); i++)
{
if (is_public && ! BASETYPE_VIA_PUBLIC (dclass, i))
continue;
d = distance_to_ancestor (base, TYPE_BASECLASS (dclass, i), is_public);
if (d >= 0)
return 1 + d;
}
return -1;
}
/* Check whether BASE is an ancestor or base class or DCLASS
Return 1 if so, and 0 if not.
Note: If BASE and DCLASS are of the same type, this function
will return 1. So for some class A, is_ancestor (A, A) will
return 1. */
int
is_ancestor (struct type *base, struct type *dclass)
{
return distance_to_ancestor (base, dclass, 0) >= 0;
}
/* Like is_ancestor, but only returns true when BASE is a public
ancestor of DCLASS. */
int
is_public_ancestor (struct type *base, struct type *dclass)
{
return distance_to_ancestor (base, dclass, 1) >= 0;
}
/* A helper function for is_unique_ancestor. */
static int
is_unique_ancestor_worker (struct type *base, struct type *dclass,
int *offset,
const gdb_byte *valaddr, int embedded_offset,
CORE_ADDR address, struct value *val)
{
int i, count = 0;
base = check_typedef (base);
dclass = check_typedef (dclass);
for (i = 0; i < TYPE_N_BASECLASSES (dclass) && count < 2; ++i)
{
struct type *iter;
int this_offset;
iter = check_typedef (TYPE_BASECLASS (dclass, i));
this_offset = baseclass_offset (dclass, i, valaddr, embedded_offset,
address, val);
if (class_types_same_p (base, iter))
{
/* If this is the first subclass, set *OFFSET and set count
to 1. Otherwise, if this is at the same offset as
previous instances, do nothing. Otherwise, increment
count. */
if (*offset == -1)
{
*offset = this_offset;
count = 1;
}
else if (this_offset == *offset)
{
/* Nothing. */
}
else
++count;
}
else
count += is_unique_ancestor_worker (base, iter, offset,
valaddr,
embedded_offset + this_offset,
address, val);
}
return count;
}
/* Like is_ancestor, but only returns true if BASE is a unique base
class of the type of VAL. */
int
is_unique_ancestor (struct type *base, struct value *val)
{
int offset = -1;
return is_unique_ancestor_worker (base, value_type (val), &offset,
value_contents_for_printing (val),
value_embedded_offset (val),
value_address (val), val) == 1;
}
/* See gdbtypes.h. */
enum bfd_endian
type_byte_order (const struct type *type)
{
bfd_endian byteorder = gdbarch_byte_order (get_type_arch (type));
if (type->endianity_is_not_default ())
{
if (byteorder == BFD_ENDIAN_BIG)
return BFD_ENDIAN_LITTLE;
else
{
gdb_assert (byteorder == BFD_ENDIAN_LITTLE);
return BFD_ENDIAN_BIG;
}
}
return byteorder;
}
/* Overload resolution. */
/* Return the sum of the rank of A with the rank of B. */
struct rank
sum_ranks (struct rank a, struct rank b)
{
struct rank c;
c.rank = a.rank + b.rank;
c.subrank = a.subrank + b.subrank;
return c;
}
/* Compare rank A and B and return:
0 if a = b
1 if a is better than b
-1 if b is better than a. */
int
compare_ranks (struct rank a, struct rank b)
{
if (a.rank == b.rank)
{
if (a.subrank == b.subrank)
return 0;
if (a.subrank < b.subrank)
return 1;
if (a.subrank > b.subrank)
return -1;
}
if (a.rank < b.rank)
return 1;
/* a.rank > b.rank */
return -1;
}
/* Functions for overload resolution begin here. */
/* Compare two badness vectors A and B and return the result.
0 => A and B are identical
1 => A and B are incomparable
2 => A is better than B
3 => A is worse than B */
int
compare_badness (const badness_vector &a, const badness_vector &b)
{
int i;
int tmp;
short found_pos = 0; /* any positives in c? */
short found_neg = 0; /* any negatives in c? */
/* differing sizes => incomparable */
if (a.size () != b.size ())
return 1;
/* Subtract b from a */
for (i = 0; i < a.size (); i++)
{
tmp = compare_ranks (b[i], a[i]);
if (tmp > 0)
found_pos = 1;
else if (tmp < 0)
found_neg = 1;
}
if (found_pos)
{
if (found_neg)
return 1; /* incomparable */
else
return 3; /* A > B */
}
else
/* no positives */
{
if (found_neg)
return 2; /* A < B */
else
return 0; /* A == B */
}
}
/* Rank a function by comparing its parameter types (PARMS), to the
types of an argument list (ARGS). Return the badness vector. This
has ARGS.size() + 1 entries. */
badness_vector
rank_function (gdb::array_view<type *> parms,
gdb::array_view<value *> args)
{
/* add 1 for the length-match rank. */
badness_vector bv;
bv.reserve (1 + args.size ());
/* First compare the lengths of the supplied lists.
If there is a mismatch, set it to a high value. */
/* pai/1997-06-03 FIXME: when we have debug info about default
arguments and ellipsis parameter lists, we should consider those
and rank the length-match more finely. */
bv.push_back ((args.size () != parms.size ())
? LENGTH_MISMATCH_BADNESS
: EXACT_MATCH_BADNESS);
/* Now rank all the parameters of the candidate function. */
size_t min_len = std::min (parms.size (), args.size ());
for (size_t i = 0; i < min_len; i++)
bv.push_back (rank_one_type (parms[i], value_type (args[i]),
args[i]));
/* If more arguments than parameters, add dummy entries. */
for (size_t i = min_len; i < args.size (); i++)
bv.push_back (TOO_FEW_PARAMS_BADNESS);
return bv;
}
/* Compare the names of two integer types, assuming that any sign
qualifiers have been checked already. We do it this way because
there may be an "int" in the name of one of the types. */
static int
integer_types_same_name_p (const char *first, const char *second)
{
int first_p, second_p;
/* If both are shorts, return 1; if neither is a short, keep
checking. */
first_p = (strstr (first, "short") != NULL);
second_p = (strstr (second, "short") != NULL);
if (first_p && second_p)
return 1;
if (first_p || second_p)
return 0;
/* Likewise for long. */
first_p = (strstr (first, "long") != NULL);
second_p = (strstr (second, "long") != NULL);
if (first_p && second_p)
return 1;
if (first_p || second_p)
return 0;
/* Likewise for char. */
first_p = (strstr (first, "char") != NULL);
second_p = (strstr (second, "char") != NULL);
if (first_p && second_p)
return 1;
if (first_p || second_p)
return 0;
/* They must both be ints. */
return 1;
}
/* Compares type A to type B. Returns true if they represent the same
type, false otherwise. */
bool
types_equal (struct type *a, struct type *b)
{
/* Identical type pointers. */
/* However, this still doesn't catch all cases of same type for b
and a. The reason is that builtin types are different from
the same ones constructed from the object. */
if (a == b)
return true;
/* Resolve typedefs */
if (a->code () == TYPE_CODE_TYPEDEF)
a = check_typedef (a);
if (b->code () == TYPE_CODE_TYPEDEF)
b = check_typedef (b);
/* If after resolving typedefs a and b are not of the same type
code then they are not equal. */
if (a->code () != b->code ())
return false;
/* If a and b are both pointers types or both reference types then
they are equal of the same type iff the objects they refer to are
of the same type. */
if (a->code () == TYPE_CODE_PTR
|| a->code () == TYPE_CODE_REF)
return types_equal (TYPE_TARGET_TYPE (a),
TYPE_TARGET_TYPE (b));
/* Well, damnit, if the names are exactly the same, I'll say they
are exactly the same. This happens when we generate method
stubs. The types won't point to the same address, but they
really are the same. */
if (a->name () && b->name ()
&& strcmp (a->name (), b->name ()) == 0)
return true;
/* Check if identical after resolving typedefs. */
if (a == b)
return true;
/* Two function types are equal if their argument and return types
are equal. */
if (a->code () == TYPE_CODE_FUNC)
{
int i;
if (a->num_fields () != b->num_fields ())
return false;
if (!types_equal (TYPE_TARGET_TYPE (a), TYPE_TARGET_TYPE (b)))
return false;
for (i = 0; i < a->num_fields (); ++i)
if (!types_equal (a->field (i).type (), b->field (i).type ()))
return false;
return true;
}
return false;
}
/* Deep comparison of types. */
/* An entry in the type-equality bcache. */
struct type_equality_entry
{
type_equality_entry (struct type *t1, struct type *t2)
: type1 (t1),
type2 (t2)
{
}
struct type *type1, *type2;
};
/* A helper function to compare two strings. Returns true if they are
the same, false otherwise. Handles NULLs properly. */
static bool
compare_maybe_null_strings (const char *s, const char *t)
{
if (s == NULL || t == NULL)
return s == t;
return strcmp (s, t) == 0;
}
/* A helper function for check_types_worklist that checks two types for
"deep" equality. Returns true if the types are considered the
same, false otherwise. */
static bool
check_types_equal (struct type *type1, struct type *type2,
std::vector<type_equality_entry> *worklist)
{
type1 = check_typedef (type1);
type2 = check_typedef (type2);
if (type1 == type2)
return true;
if (type1->code () != type2->code ()
|| TYPE_LENGTH (type1) != TYPE_LENGTH (type2)
|| type1->is_unsigned () != type2->is_unsigned ()
|| type1->has_no_signedness () != type2->has_no_signedness ()
|| type1->endianity_is_not_default () != type2->endianity_is_not_default ()
|| type1->has_varargs () != type2->has_varargs ()
|| type1->is_vector () != type2->is_vector ()
|| TYPE_NOTTEXT (type1) != TYPE_NOTTEXT (type2)
|| type1->instance_flags () != type2->instance_flags ()
|| type1->num_fields () != type2->num_fields ())
return false;
if (!compare_maybe_null_strings (type1->name (), type2->name ()))
return false;
if (!compare_maybe_null_strings (type1->name (), type2->name ()))
return false;
if (type1->code () == TYPE_CODE_RANGE)
{
if (*type1->bounds () != *type2->bounds ())
return false;
}
else
{
int i;
for (i = 0; i < type1->num_fields (); ++i)
{
const struct field *field1 = &type1->field (i);
const struct field *field2 = &type2->field (i);
if (FIELD_ARTIFICIAL (*field1) != FIELD_ARTIFICIAL (*field2)
|| FIELD_BITSIZE (*field1) != FIELD_BITSIZE (*field2)
|| FIELD_LOC_KIND (*field1) != FIELD_LOC_KIND (*field2))
return false;
if (!compare_maybe_null_strings (FIELD_NAME (*field1),
FIELD_NAME (*field2)))
return false;
switch (FIELD_LOC_KIND (*field1))
{
case FIELD_LOC_KIND_BITPOS:
if (FIELD_BITPOS (*field1) != FIELD_BITPOS (*field2))
return false;
break;
case FIELD_LOC_KIND_ENUMVAL:
if (FIELD_ENUMVAL (*field1) != FIELD_ENUMVAL (*field2))
return false;
/* Don't compare types of enum fields, because they don't
have a type. */
continue;
case FIELD_LOC_KIND_PHYSADDR:
if (FIELD_STATIC_PHYSADDR (*field1)
!= FIELD_STATIC_PHYSADDR (*field2))
return false;
break;
case FIELD_LOC_KIND_PHYSNAME:
if (!compare_maybe_null_strings (FIELD_STATIC_PHYSNAME (*field1),
FIELD_STATIC_PHYSNAME (*field2)))
return false;
break;
case FIELD_LOC_KIND_DWARF_BLOCK:
{
struct dwarf2_locexpr_baton *block1, *block2;
block1 = FIELD_DWARF_BLOCK (*field1);
block2 = FIELD_DWARF_BLOCK (*field2);
if (block1->per_cu != block2->per_cu
|| block1->size != block2->size
|| memcmp (block1->data, block2->data, block1->size) != 0)
return false;
}
break;
default:
internal_error (__FILE__, __LINE__, _("Unsupported field kind "
"%d by check_types_equal"),
FIELD_LOC_KIND (*field1));
}
worklist->emplace_back (field1->type (), field2->type ());
}
}
if (TYPE_TARGET_TYPE (type1) != NULL)
{
if (TYPE_TARGET_TYPE (type2) == NULL)
return false;
worklist->emplace_back (TYPE_TARGET_TYPE (type1),
TYPE_TARGET_TYPE (type2));
}
else if (TYPE_TARGET_TYPE (type2) != NULL)
return false;
return true;
}
/* Check types on a worklist for equality. Returns false if any pair
is not equal, true if they are all considered equal. */
static bool
check_types_worklist (std::vector<type_equality_entry> *worklist,
gdb::bcache *cache)
{
while (!worklist->empty ())
{
bool added;
struct type_equality_entry entry = std::move (worklist->back ());
worklist->pop_back ();
/* If the type pair has already been visited, we know it is
ok. */
cache->insert (&entry, sizeof (entry), &added);
if (!added)
continue;
if (!check_types_equal (entry.type1, entry.type2, worklist))
return false;
}
return true;
}
/* Return true if types TYPE1 and TYPE2 are equal, as determined by a
"deep comparison". Otherwise return false. */
bool
types_deeply_equal (struct type *type1, struct type *type2)
{
std::vector<type_equality_entry> worklist;
gdb_assert (type1 != NULL && type2 != NULL);
/* Early exit for the simple case. */
if (type1 == type2)
return true;
gdb::bcache cache;
worklist.emplace_back (type1, type2);
return check_types_worklist (&worklist, &cache);
}
/* Allocated status of type TYPE. Return zero if type TYPE is allocated.
Otherwise return one. */
int
type_not_allocated (const struct type *type)
{
struct dynamic_prop *prop = TYPE_ALLOCATED_PROP (type);
return (prop != nullptr && prop->kind () == PROP_CONST
&& prop->const_val () == 0);
}
/* Associated status of type TYPE. Return zero if type TYPE is associated.
Otherwise return one. */
int
type_not_associated (const struct type *type)
{
struct dynamic_prop *prop = TYPE_ASSOCIATED_PROP (type);
return (prop != nullptr && prop->kind () == PROP_CONST
&& prop->const_val () == 0);
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_PTR. */
static struct rank
rank_one_type_parm_ptr (struct type *parm, struct type *arg, struct value *value)
{
struct rank rank = {0,0};
switch (arg->code ())
{
case TYPE_CODE_PTR:
/* Allowed pointer conversions are:
(a) pointer to void-pointer conversion. */
if (TYPE_TARGET_TYPE (parm)->code () == TYPE_CODE_VOID)
return VOID_PTR_CONVERSION_BADNESS;
/* (b) pointer to ancestor-pointer conversion. */
rank.subrank = distance_to_ancestor (TYPE_TARGET_TYPE (parm),
TYPE_TARGET_TYPE (arg),
0);
if (rank.subrank >= 0)
return sum_ranks (BASE_PTR_CONVERSION_BADNESS, rank);
return INCOMPATIBLE_TYPE_BADNESS;
case TYPE_CODE_ARRAY:
{
struct type *t1 = TYPE_TARGET_TYPE (parm);
struct type *t2 = TYPE_TARGET_TYPE (arg);
if (types_equal (t1, t2))
{
/* Make sure they are CV equal. */
if (TYPE_CONST (t1) != TYPE_CONST (t2))
rank.subrank |= CV_CONVERSION_CONST;
if (TYPE_VOLATILE (t1) != TYPE_VOLATILE (t2))
rank.subrank |= CV_CONVERSION_VOLATILE;
if (rank.subrank != 0)
return sum_ranks (CV_CONVERSION_BADNESS, rank);
return EXACT_MATCH_BADNESS;
}
return INCOMPATIBLE_TYPE_BADNESS;
}
case TYPE_CODE_FUNC:
return rank_one_type (TYPE_TARGET_TYPE (parm), arg, NULL);
case TYPE_CODE_INT:
if (value != NULL && value_type (value)->code () == TYPE_CODE_INT)
{
if (value_as_long (value) == 0)
{
/* Null pointer conversion: allow it to be cast to a pointer.
[4.10.1 of C++ standard draft n3290] */
return NULL_POINTER_CONVERSION_BADNESS;
}
else
{
/* If type checking is disabled, allow the conversion. */
if (!strict_type_checking)
return NS_INTEGER_POINTER_CONVERSION_BADNESS;
}
}
/* fall through */
case TYPE_CODE_ENUM:
case TYPE_CODE_FLAGS:
case TYPE_CODE_CHAR:
case TYPE_CODE_RANGE:
case TYPE_CODE_BOOL:
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_ARRAY. */
static struct rank
rank_one_type_parm_array (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
case TYPE_CODE_PTR:
case TYPE_CODE_ARRAY:
return rank_one_type (TYPE_TARGET_TYPE (parm),
TYPE_TARGET_TYPE (arg), NULL);
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_FUNC. */
static struct rank
rank_one_type_parm_func (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
case TYPE_CODE_PTR: /* funcptr -> func */
return rank_one_type (parm, TYPE_TARGET_TYPE (arg), NULL);
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_INT. */
static struct rank
rank_one_type_parm_int (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
case TYPE_CODE_INT:
if (TYPE_LENGTH (arg) == TYPE_LENGTH (parm))
{
/* Deal with signed, unsigned, and plain chars and
signed and unsigned ints. */
if (parm->has_no_signedness ())
{
/* This case only for character types. */
if (arg->has_no_signedness ())
return EXACT_MATCH_BADNESS; /* plain char -> plain char */
else /* signed/unsigned char -> plain char */
return INTEGER_CONVERSION_BADNESS;
}
else if (parm->is_unsigned ())
{
if (arg->is_unsigned ())
{
/* unsigned int -> unsigned int, or
unsigned long -> unsigned long */
if (integer_types_same_name_p (parm->name (),
arg->name ()))
return EXACT_MATCH_BADNESS;
else if (integer_types_same_name_p (arg->name (),
"int")
&& integer_types_same_name_p (parm->name (),
"long"))
/* unsigned int -> unsigned long */
return INTEGER_PROMOTION_BADNESS;
else
/* unsigned long -> unsigned int */
return INTEGER_CONVERSION_BADNESS;
}
else
{
if (integer_types_same_name_p (arg->name (),
"long")
&& integer_types_same_name_p (parm->name (),
"int"))
/* signed long -> unsigned int */
return INTEGER_CONVERSION_BADNESS;
else
/* signed int/long -> unsigned int/long */
return INTEGER_CONVERSION_BADNESS;
}
}
else if (!arg->has_no_signedness () && !arg->is_unsigned ())
{
if (integer_types_same_name_p (parm->name (),
arg->name ()))
return EXACT_MATCH_BADNESS;
else if (integer_types_same_name_p (arg->name (),
"int")
&& integer_types_same_name_p (parm->name (),
"long"))
return INTEGER_PROMOTION_BADNESS;
else
return INTEGER_CONVERSION_BADNESS;
}
else
return INTEGER_CONVERSION_BADNESS;
}
else if (TYPE_LENGTH (arg) < TYPE_LENGTH (parm))
return INTEGER_PROMOTION_BADNESS;
else
return INTEGER_CONVERSION_BADNESS;
case TYPE_CODE_ENUM:
case TYPE_CODE_FLAGS:
case TYPE_CODE_CHAR:
case TYPE_CODE_RANGE:
case TYPE_CODE_BOOL:
if (TYPE_DECLARED_CLASS (arg))
return INCOMPATIBLE_TYPE_BADNESS;
return INTEGER_PROMOTION_BADNESS;
case TYPE_CODE_FLT:
return INT_FLOAT_CONVERSION_BADNESS;
case TYPE_CODE_PTR:
return NS_POINTER_CONVERSION_BADNESS;
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_ENUM. */
static struct rank
rank_one_type_parm_enum (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
case TYPE_CODE_INT:
case TYPE_CODE_CHAR:
case TYPE_CODE_RANGE:
case TYPE_CODE_BOOL:
case TYPE_CODE_ENUM:
if (TYPE_DECLARED_CLASS (parm) || TYPE_DECLARED_CLASS (arg))
return INCOMPATIBLE_TYPE_BADNESS;
return INTEGER_CONVERSION_BADNESS;
case TYPE_CODE_FLT:
return INT_FLOAT_CONVERSION_BADNESS;
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_CHAR. */
static struct rank
rank_one_type_parm_char (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
case TYPE_CODE_RANGE:
case TYPE_CODE_BOOL:
case TYPE_CODE_ENUM:
if (TYPE_DECLARED_CLASS (arg))
return INCOMPATIBLE_TYPE_BADNESS;
return INTEGER_CONVERSION_BADNESS;
case TYPE_CODE_FLT:
return INT_FLOAT_CONVERSION_BADNESS;
case TYPE_CODE_INT:
if (TYPE_LENGTH (arg) > TYPE_LENGTH (parm))
return INTEGER_CONVERSION_BADNESS;
else if (TYPE_LENGTH (arg) < TYPE_LENGTH (parm))
return INTEGER_PROMOTION_BADNESS;
/* fall through */
case TYPE_CODE_CHAR:
/* Deal with signed, unsigned, and plain chars for C++ and
with int cases falling through from previous case. */
if (parm->has_no_signedness ())
{
if (arg->has_no_signedness ())
return EXACT_MATCH_BADNESS;
else
return INTEGER_CONVERSION_BADNESS;
}
else if (parm->is_unsigned ())
{
if (arg->is_unsigned ())
return EXACT_MATCH_BADNESS;
else
return INTEGER_PROMOTION_BADNESS;
}
else if (!arg->has_no_signedness () && !arg->is_unsigned ())
return EXACT_MATCH_BADNESS;
else
return INTEGER_CONVERSION_BADNESS;
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_RANGE. */
static struct rank
rank_one_type_parm_range (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
case TYPE_CODE_INT:
case TYPE_CODE_CHAR:
case TYPE_CODE_RANGE:
case TYPE_CODE_BOOL:
case TYPE_CODE_ENUM:
return INTEGER_CONVERSION_BADNESS;
case TYPE_CODE_FLT:
return INT_FLOAT_CONVERSION_BADNESS;
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_BOOL. */
static struct rank
rank_one_type_parm_bool (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
/* n3290 draft, section 4.12.1 (conv.bool):
"A prvalue of arithmetic, unscoped enumeration, pointer, or
pointer to member type can be converted to a prvalue of type
bool. A zero value, null pointer value, or null member pointer
value is converted to false; any other value is converted to
true. A prvalue of type std::nullptr_t can be converted to a
prvalue of type bool; the resulting value is false." */
case TYPE_CODE_INT:
case TYPE_CODE_CHAR:
case TYPE_CODE_ENUM:
case TYPE_CODE_FLT:
case TYPE_CODE_MEMBERPTR:
case TYPE_CODE_PTR:
return BOOL_CONVERSION_BADNESS;
case TYPE_CODE_RANGE:
return INCOMPATIBLE_TYPE_BADNESS;
case TYPE_CODE_BOOL:
return EXACT_MATCH_BADNESS;
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_FLOAT. */
static struct rank
rank_one_type_parm_float (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
case TYPE_CODE_FLT:
if (TYPE_LENGTH (arg) < TYPE_LENGTH (parm))
return FLOAT_PROMOTION_BADNESS;
else if (TYPE_LENGTH (arg) == TYPE_LENGTH (parm))
return EXACT_MATCH_BADNESS;
else
return FLOAT_CONVERSION_BADNESS;
case TYPE_CODE_INT:
case TYPE_CODE_BOOL:
case TYPE_CODE_ENUM:
case TYPE_CODE_RANGE:
case TYPE_CODE_CHAR:
return INT_FLOAT_CONVERSION_BADNESS;
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_COMPLEX. */
static struct rank
rank_one_type_parm_complex (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{ /* Strictly not needed for C++, but... */
case TYPE_CODE_FLT:
return FLOAT_PROMOTION_BADNESS;
case TYPE_CODE_COMPLEX:
return EXACT_MATCH_BADNESS;
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_STRUCT. */
static struct rank
rank_one_type_parm_struct (struct type *parm, struct type *arg, struct value *value)
{
struct rank rank = {0, 0};
switch (arg->code ())
{
case TYPE_CODE_STRUCT:
/* Check for derivation */
rank.subrank = distance_to_ancestor (parm, arg, 0);
if (rank.subrank >= 0)
return sum_ranks (BASE_CONVERSION_BADNESS, rank);
/* fall through */
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* rank_one_type helper for when PARM's type code is TYPE_CODE_SET. */
static struct rank
rank_one_type_parm_set (struct type *parm, struct type *arg, struct value *value)
{
switch (arg->code ())
{
/* Not in C++ */
case TYPE_CODE_SET:
return rank_one_type (parm->field (0).type (),
arg->field (0).type (), NULL);
default:
return INCOMPATIBLE_TYPE_BADNESS;
}
}
/* Compare one type (PARM) for compatibility with another (ARG).
* PARM is intended to be the parameter type of a function; and
* ARG is the supplied argument's type. This function tests if
* the latter can be converted to the former.
* VALUE is the argument's value or NULL if none (or called recursively)
*
* Return 0 if they are identical types;
* Otherwise, return an integer which corresponds to how compatible
* PARM is to ARG. The higher the return value, the worse the match.
* Generally the "bad" conversions are all uniformly assigned a 100. */
struct rank
rank_one_type (struct type *parm, struct type *arg, struct value *value)
{
struct rank rank = {0,0};
/* Resolve typedefs */
if (parm->code () == TYPE_CODE_TYPEDEF)
parm = check_typedef (parm);
if (arg->code () == TYPE_CODE_TYPEDEF)
arg = check_typedef (arg);
if (TYPE_IS_REFERENCE (parm) && value != NULL)
{
if (VALUE_LVAL (value) == not_lval)
{
/* Rvalues should preferably bind to rvalue references or const
lvalue references. */
if (parm->code () == TYPE_CODE_RVALUE_REF)
rank.subrank = REFERENCE_CONVERSION_RVALUE;
else if (TYPE_CONST (TYPE_TARGET_TYPE (parm)))
rank.subrank = REFERENCE_CONVERSION_CONST_LVALUE;
else
return INCOMPATIBLE_TYPE_BADNESS;
return sum_ranks (rank, REFERENCE_CONVERSION_BADNESS);
}
else
{
/* It's illegal to pass an lvalue as an rvalue. */
if (parm->code () == TYPE_CODE_RVALUE_REF)
return INCOMPATIBLE_TYPE_BADNESS;
}
}
if (types_equal (parm, arg))
{
struct type *t1 = parm;
struct type *t2 = arg;
/* For pointers and references, compare target type. */
if (parm->code () == TYPE_CODE_PTR || TYPE_IS_REFERENCE (parm))
{
t1 = TYPE_TARGET_TYPE (parm);
t2 = TYPE_TARGET_TYPE (arg);
}
/* Make sure they are CV equal, too. */
if (TYPE_CONST (t1) != TYPE_CONST (t2))
rank.subrank |= CV_CONVERSION_CONST;
if (TYPE_VOLATILE (t1) != TYPE_VOLATILE (t2))
rank.subrank |= CV_CONVERSION_VOLATILE;
if (rank.subrank != 0)
return sum_ranks (CV_CONVERSION_BADNESS, rank);
return EXACT_MATCH_BADNESS;
}
/* See through references, since we can almost make non-references
references. */
if (TYPE_IS_REFERENCE (arg))
return (sum_ranks (rank_one_type (parm, TYPE_TARGET_TYPE (arg), NULL),
REFERENCE_SEE_THROUGH_BADNESS));
if (TYPE_IS_REFERENCE (parm))
return (sum_ranks (rank_one_type (TYPE_TARGET_TYPE (parm), arg, NULL),
REFERENCE_SEE_THROUGH_BADNESS));
if (overload_debug)
/* Debugging only. */
fprintf_filtered (gdb_stderr,
"------ Arg is %s [%d], parm is %s [%d]\n",
arg->name (), arg->code (),
parm->name (), parm->code ());
/* x -> y means arg of type x being supplied for parameter of type y. */
switch (parm->code ())
{
case TYPE_CODE_PTR:
return rank_one_type_parm_ptr (parm, arg, value);
case TYPE_CODE_ARRAY:
return rank_one_type_parm_array (parm, arg, value);
case TYPE_CODE_FUNC:
return rank_one_type_parm_func (parm, arg, value);
case TYPE_CODE_INT:
return rank_one_type_parm_int (parm, arg, value);
case TYPE_CODE_ENUM:
return rank_one_type_parm_enum (parm, arg, value);
case TYPE_CODE_CHAR:
return rank_one_type_parm_char (parm, arg, value);
case TYPE_CODE_RANGE:
return rank_one_type_parm_range (parm, arg, value);
case TYPE_CODE_BOOL:
return rank_one_type_parm_bool (parm, arg, value);
case TYPE_CODE_FLT:
return rank_one_type_parm_float (parm, arg, value);
case TYPE_CODE_COMPLEX:
return rank_one_type_parm_complex (parm, arg, value);
case TYPE_CODE_STRUCT:
return rank_one_type_parm_struct (parm, arg, value);
case TYPE_CODE_SET:
return rank_one_type_parm_set (parm, arg, value);
default:
return INCOMPATIBLE_TYPE_BADNESS;
} /* switch (arg->code ()) */
}
/* End of functions for overload resolution. */
/* Routines to pretty-print types. */
static void
print_bit_vector (B_TYPE *bits, int nbits)
{
int bitno;
for (bitno = 0; bitno < nbits; bitno++)
{
if ((bitno % 8) == 0)
{
puts_filtered (" ");
}
if (B_TST (bits, bitno))
printf_filtered (("1"));
else
printf_filtered (("0"));
}
}
/* Note the first arg should be the "this" pointer, we may not want to
include it since we may get into a infinitely recursive
situation. */
static void
print_args (struct field *args, int nargs, int spaces)
{
if (args != NULL)
{
int i;
for (i = 0; i < nargs; i++)
{
printf_filtered ("%*s[%d] name '%s'\n", spaces, "", i,
args[i].name != NULL ? args[i].name : "<NULL>");
recursive_dump_type (args[i].type (), spaces + 2);
}
}
}
int
field_is_static (struct field *f)
{
/* "static" fields are the fields whose location is not relative
to the address of the enclosing struct. It would be nice to
have a dedicated flag that would be set for static fields when
the type is being created. But in practice, checking the field
loc_kind should give us an accurate answer. */
return (FIELD_LOC_KIND (*f) == FIELD_LOC_KIND_PHYSNAME
|| FIELD_LOC_KIND (*f) == FIELD_LOC_KIND_PHYSADDR);
}
static void
dump_fn_fieldlists (struct type *type, int spaces)
{
int method_idx;
int overload_idx;
struct fn_field *f;
printf_filtered ("%*sfn_fieldlists ", spaces, "");
gdb_print_host_address (TYPE_FN_FIELDLISTS (type), gdb_stdout);
printf_filtered ("\n");
for (method_idx = 0; method_idx < TYPE_NFN_FIELDS (type); method_idx++)
{
f = TYPE_FN_FIELDLIST1 (type, method_idx);
printf_filtered ("%*s[%d] name '%s' (", spaces + 2, "",
method_idx,
TYPE_FN_FIELDLIST_NAME (type, method_idx));
gdb_print_host_address (TYPE_FN_FIELDLIST_NAME (type, method_idx),
gdb_stdout);
printf_filtered (_(") length %d\n"),
TYPE_FN_FIELDLIST_LENGTH (type, method_idx));
for (overload_idx = 0;
overload_idx < TYPE_FN_FIELDLIST_LENGTH (type, method_idx);
overload_idx++)
{
printf_filtered ("%*s[%d] physname '%s' (",
spaces + 4, "", overload_idx,
TYPE_FN_FIELD_PHYSNAME (f, overload_idx));
gdb_print_host_address (TYPE_FN_FIELD_PHYSNAME (f, overload_idx),
gdb_stdout);
printf_filtered (")\n");
printf_filtered ("%*stype ", spaces + 8, "");
gdb_print_host_address (TYPE_FN_FIELD_TYPE (f, overload_idx),
gdb_stdout);
printf_filtered ("\n");
recursive_dump_type (TYPE_FN_FIELD_TYPE (f, overload_idx),
spaces + 8 + 2);
printf_filtered ("%*sargs ", spaces + 8, "");
gdb_print_host_address (TYPE_FN_FIELD_ARGS (f, overload_idx),
gdb_stdout);
printf_filtered ("\n");
print_args (TYPE_FN_FIELD_ARGS (f, overload_idx),
TYPE_FN_FIELD_TYPE (f, overload_idx)->num_fields (),
spaces + 8 + 2);
printf_filtered ("%*sfcontext ", spaces + 8, "");
gdb_print_host_address (TYPE_FN_FIELD_FCONTEXT (f, overload_idx),
gdb_stdout);
printf_filtered ("\n");
printf_filtered ("%*sis_const %d\n", spaces + 8, "",
TYPE_FN_FIELD_CONST (f, overload_idx));
printf_filtered ("%*sis_volatile %d\n", spaces + 8, "",
TYPE_FN_FIELD_VOLATILE (f, overload_idx));
printf_filtered ("%*sis_private %d\n", spaces + 8, "",
TYPE_FN_FIELD_PRIVATE (f, overload_idx));
printf_filtered ("%*sis_protected %d\n", spaces + 8, "",
TYPE_FN_FIELD_PROTECTED (f, overload_idx));
printf_filtered ("%*sis_stub %d\n", spaces + 8, "",
TYPE_FN_FIELD_STUB (f, overload_idx));
printf_filtered ("%*sdefaulted %d\n", spaces + 8, "",
TYPE_FN_FIELD_DEFAULTED (f, overload_idx));
printf_filtered ("%*sis_deleted %d\n", spaces + 8, "",
TYPE_FN_FIELD_DELETED (f, overload_idx));
printf_filtered ("%*svoffset %u\n", spaces + 8, "",
TYPE_FN_FIELD_VOFFSET (f, overload_idx));
}
}
}
static void
print_cplus_stuff (struct type *type, int spaces)
{
printf_filtered ("%*svptr_fieldno %d\n", spaces, "",
TYPE_VPTR_FIELDNO (type));
printf_filtered ("%*svptr_basetype ", spaces, "");
gdb_print_host_address (TYPE_VPTR_BASETYPE (type), gdb_stdout);
puts_filtered ("\n");
if (TYPE_VPTR_BASETYPE (type) != NULL)
recursive_dump_type (TYPE_VPTR_BASETYPE (type), spaces + 2);
printf_filtered ("%*sn_baseclasses %d\n", spaces, "",
TYPE_N_BASECLASSES (type));
printf_filtered ("%*snfn_fields %d\n", spaces, "",
TYPE_NFN_FIELDS (type));
if (TYPE_N_BASECLASSES (type) > 0)
{
printf_filtered ("%*svirtual_field_bits (%d bits at *",
spaces, "", TYPE_N_BASECLASSES (type));
gdb_print_host_address (TYPE_FIELD_VIRTUAL_BITS (type),
gdb_stdout);
printf_filtered (")");
print_bit_vector (TYPE_FIELD_VIRTUAL_BITS (type),
TYPE_N_BASECLASSES (type));
puts_filtered ("\n");
}
if (type->num_fields () > 0)
{
if (TYPE_FIELD_PRIVATE_BITS (type) != NULL)
{
printf_filtered ("%*sprivate_field_bits (%d bits at *",
spaces, "", type->num_fields ());
gdb_print_host_address (TYPE_FIELD_PRIVATE_BITS (type),
gdb_stdout);
printf_filtered (")");
print_bit_vector (TYPE_FIELD_PRIVATE_BITS (type),
type->num_fields ());
puts_filtered ("\n");
}
if (TYPE_FIELD_PROTECTED_BITS (type) != NULL)
{
printf_filtered ("%*sprotected_field_bits (%d bits at *",
spaces, "", type->num_fields ());
gdb_print_host_address (TYPE_FIELD_PROTECTED_BITS (type),
gdb_stdout);
printf_filtered (")");
print_bit_vector (TYPE_FIELD_PROTECTED_BITS (type),
type->num_fields ());
puts_filtered ("\n");
}
}
if (TYPE_NFN_FIELDS (type) > 0)
{
dump_fn_fieldlists (type, spaces);
}
printf_filtered ("%*scalling_convention %d\n", spaces, "",
TYPE_CPLUS_CALLING_CONVENTION (type));
}
/* Print the contents of the TYPE's type_specific union, assuming that
its type-specific kind is TYPE_SPECIFIC_GNAT_STUFF. */
static void
print_gnat_stuff (struct type *type, int spaces)
{
struct type *descriptive_type = TYPE_DESCRIPTIVE_TYPE (type);
if (descriptive_type == NULL)
printf_filtered ("%*sno descriptive type\n", spaces + 2, "");
else
{
printf_filtered ("%*sdescriptive type\n", spaces + 2, "");
recursive_dump_type (descriptive_type, spaces + 4);
}
}
/* Print the contents of the TYPE's type_specific union, assuming that
its type-specific kind is TYPE_SPECIFIC_FIXED_POINT. */
static void
print_fixed_point_type_info (struct type *type, int spaces)
{
printf_filtered ("%*sscaling factor: %s\n", spaces + 2, "",
type->fixed_point_scaling_factor ().str ().c_str ());
}
static struct obstack dont_print_type_obstack;
/* Print the dynamic_prop PROP. */
static void
dump_dynamic_prop (dynamic_prop const& prop)
{
switch (prop.kind ())
{
case PROP_CONST:
printf_filtered ("%s", plongest (prop.const_val ()));
break;
case PROP_UNDEFINED:
printf_filtered ("(undefined)");
break;
case PROP_LOCEXPR:
case PROP_LOCLIST:
printf_filtered ("(dynamic)");
break;
default:
gdb_assert_not_reached ("unhandled prop kind");
break;
}
}
void
recursive_dump_type (struct type *type, int spaces)
{
int idx;
if (spaces == 0)
obstack_begin (&dont_print_type_obstack, 0);
if (type->num_fields () > 0
|| (HAVE_CPLUS_STRUCT (type) && TYPE_NFN_FIELDS (type) > 0))
{
struct type **first_dont_print
= (struct type **) obstack_base (&dont_print_type_obstack);
int i = (struct type **)
obstack_next_free (&dont_print_type_obstack) - first_dont_print;
while (--i >= 0)
{
if (type == first_dont_print[i])
{
printf_filtered ("%*stype node ", spaces, "");
gdb_print_host_address (type, gdb_stdout);
printf_filtered (_(" <same as already seen type>\n"));
return;
}
}
obstack_ptr_grow (&dont_print_type_obstack, type);
}
printf_filtered ("%*stype node ", spaces, "");
gdb_print_host_address (type, gdb_stdout);
printf_filtered ("\n");
printf_filtered ("%*sname '%s' (", spaces, "",
type->name () ? type->name () : "<NULL>");
gdb_print_host_address (type->name (), gdb_stdout);
printf_filtered (")\n");
printf_filtered ("%*scode 0x%x ", spaces, "", type->code ());
switch (type->code ())
{
case TYPE_CODE_UNDEF:
printf_filtered ("(TYPE_CODE_UNDEF)");
break;
case TYPE_CODE_PTR:
printf_filtered ("(TYPE_CODE_PTR)");
break;
case TYPE_CODE_ARRAY:
printf_filtered ("(TYPE_CODE_ARRAY)");
break;
case TYPE_CODE_STRUCT:
printf_filtered ("(TYPE_CODE_STRUCT)");
break;
case TYPE_CODE_UNION:
printf_filtered ("(TYPE_CODE_UNION)");
break;
case TYPE_CODE_ENUM:
printf_filtered ("(TYPE_CODE_ENUM)");
break;
case TYPE_CODE_FLAGS:
printf_filtered ("(TYPE_CODE_FLAGS)");
break;
case TYPE_CODE_FUNC:
printf_filtered ("(TYPE_CODE_FUNC)");
break;
case TYPE_CODE_INT:
printf_filtered ("(TYPE_CODE_INT)");
break;
case TYPE_CODE_FLT:
printf_filtered ("(TYPE_CODE_FLT)");
break;
case TYPE_CODE_VOID:
printf_filtered ("(TYPE_CODE_VOID)");
break;
case TYPE_CODE_SET:
printf_filtered ("(TYPE_CODE_SET)");
break;
case TYPE_CODE_RANGE:
printf_filtered ("(TYPE_CODE_RANGE)");
break;
case TYPE_CODE_STRING:
printf_filtered ("(TYPE_CODE_STRING)");
break;
case TYPE_CODE_ERROR:
printf_filtered ("(TYPE_CODE_ERROR)");
break;
case TYPE_CODE_MEMBERPTR:
printf_filtered ("(TYPE_CODE_MEMBERPTR)");
break;
case TYPE_CODE_METHODPTR:
printf_filtered ("(TYPE_CODE_METHODPTR)");
break;
case TYPE_CODE_METHOD:
printf_filtered ("(TYPE_CODE_METHOD)");
break;
case TYPE_CODE_REF:
printf_filtered ("(TYPE_CODE_REF)");
break;
case TYPE_CODE_CHAR:
printf_filtered ("(TYPE_CODE_CHAR)");
break;
case TYPE_CODE_BOOL:
printf_filtered ("(TYPE_CODE_BOOL)");
break;
case TYPE_CODE_COMPLEX:
printf_filtered ("(TYPE_CODE_COMPLEX)");
break;
case TYPE_CODE_TYPEDEF:
printf_filtered ("(TYPE_CODE_TYPEDEF)");
break;
case TYPE_CODE_NAMESPACE:
printf_filtered ("(TYPE_CODE_NAMESPACE)");
break;
case TYPE_CODE_FIXED_POINT:
printf_filtered ("(TYPE_CODE_FIXED_POINT)");
break;
default:
printf_filtered ("(UNKNOWN TYPE CODE)");
break;
}
puts_filtered ("\n");
printf_filtered ("%*slength %s\n", spaces, "",
pulongest (TYPE_LENGTH (type)));
if (TYPE_OBJFILE_OWNED (type))
{
printf_filtered ("%*sobjfile ", spaces, "");
gdb_print_host_address (TYPE_OWNER (type).objfile, gdb_stdout);
}
else
{
printf_filtered ("%*sgdbarch ", spaces, "");
gdb_print_host_address (TYPE_OWNER (type).gdbarch, gdb_stdout);
}
printf_filtered ("\n");
printf_filtered ("%*starget_type ", spaces, "");
gdb_print_host_address (TYPE_TARGET_TYPE (type), gdb_stdout);
printf_filtered ("\n");
if (TYPE_TARGET_TYPE (type) != NULL)
{
recursive_dump_type (TYPE_TARGET_TYPE (type), spaces + 2);
}
printf_filtered ("%*spointer_type ", spaces, "");
gdb_print_host_address (TYPE_POINTER_TYPE (type), gdb_stdout);
printf_filtered ("\n");
printf_filtered ("%*sreference_type ", spaces, "");
gdb_print_host_address (TYPE_REFERENCE_TYPE (type), gdb_stdout);
printf_filtered ("\n");
printf_filtered ("%*stype_chain ", spaces, "");
gdb_print_host_address (TYPE_CHAIN (type), gdb_stdout);
printf_filtered ("\n");
printf_filtered ("%*sinstance_flags 0x%x", spaces, "",
(unsigned) type->instance_flags ());
if (TYPE_CONST (type))
{
puts_filtered (" TYPE_CONST");
}
if (TYPE_VOLATILE (type))
{
puts_filtered (" TYPE_VOLATILE");
}
if (TYPE_CODE_SPACE (type))
{
puts_filtered (" TYPE_CODE_SPACE");
}
if (TYPE_DATA_SPACE (type))
{
puts_filtered (" TYPE_DATA_SPACE");
}
if (TYPE_ADDRESS_CLASS_1 (type))
{
puts_filtered (" TYPE_ADDRESS_CLASS_1");
}
if (TYPE_ADDRESS_CLASS_2 (type))
{
puts_filtered (" TYPE_ADDRESS_CLASS_2");
}
if (TYPE_RESTRICT (type))
{
puts_filtered (" TYPE_RESTRICT");
}
if (TYPE_ATOMIC (type))
{
puts_filtered (" TYPE_ATOMIC");
}
puts_filtered ("\n");
printf_filtered ("%*sflags", spaces, "");
if (type->is_unsigned ())
{
puts_filtered (" TYPE_UNSIGNED");
}
if (type->has_no_signedness ())
{
puts_filtered (" TYPE_NOSIGN");
}
if (type->endianity_is_not_default ())
{
puts_filtered (" TYPE_ENDIANITY_NOT_DEFAULT");
}
if (type->is_stub ())
{
puts_filtered (" TYPE_STUB");
}
if (type->target_is_stub ())
{
puts_filtered (" TYPE_TARGET_STUB");
}
if (type->is_prototyped ())
{
puts_filtered (" TYPE_PROTOTYPED");
}
if (type->has_varargs ())
{
puts_filtered (" TYPE_VARARGS");
}
/* This is used for things like AltiVec registers on ppc. Gcc emits
an attribute for the array type, which tells whether or not we
have a vector, instead of a regular array. */
if (type->is_vector ())
{
puts_filtered (" TYPE_VECTOR");
}
if (type->is_fixed_instance ())
{
puts_filtered (" TYPE_FIXED_INSTANCE");
}
if (type->stub_is_supported ())
{
puts_filtered (" TYPE_STUB_SUPPORTED");
}
if (TYPE_NOTTEXT (type))
{
puts_filtered (" TYPE_NOTTEXT");
}
puts_filtered ("\n");
printf_filtered ("%*snfields %d ", spaces, "", type->num_fields ());
if (TYPE_ASSOCIATED_PROP (type) != nullptr
|| TYPE_ALLOCATED_PROP (type) != nullptr)
{
printf_filtered ("%*s", spaces, "");
if (TYPE_ASSOCIATED_PROP (type) != nullptr)
{
printf_filtered ("associated ");
dump_dynamic_prop (*TYPE_ASSOCIATED_PROP (type));
}
if (TYPE_ALLOCATED_PROP (type) != nullptr)
{
if (TYPE_ASSOCIATED_PROP (type) != nullptr)
printf_filtered (" ");
printf_filtered ("allocated ");
dump_dynamic_prop (*TYPE_ALLOCATED_PROP (type));
}
printf_filtered ("\n");
}
gdb_print_host_address (type->fields (), gdb_stdout);
puts_filtered ("\n");
for (idx = 0; idx < type->num_fields (); idx++)
{
if (type->code () == TYPE_CODE_ENUM)
printf_filtered ("%*s[%d] enumval %s type ", spaces + 2, "",
idx, plongest (TYPE_FIELD_ENUMVAL (type, idx)));
else
printf_filtered ("%*s[%d] bitpos %s bitsize %d type ", spaces + 2, "",
idx, plongest (TYPE_FIELD_BITPOS (type, idx)),
TYPE_FIELD_BITSIZE (type, idx));
gdb_print_host_address (type->field (idx).type (), gdb_stdout);
printf_filtered (" name '%s' (",
TYPE_FIELD_NAME (type, idx) != NULL
? TYPE_FIELD_NAME (type, idx)
: "<NULL>");
gdb_print_host_address (TYPE_FIELD_NAME (type, idx), gdb_stdout);
printf_filtered (")\n");
if (type->field (idx).type () != NULL)
{
recursive_dump_type (type->field (idx).type (), spaces + 4);
}
}
if (type->code () == TYPE_CODE_RANGE)
{
printf_filtered ("%*slow ", spaces, "");
dump_dynamic_prop (type->bounds ()->low);
printf_filtered (" high ");
dump_dynamic_prop (type->bounds ()->high);
printf_filtered ("\n");
}
switch (TYPE_SPECIFIC_FIELD (type))
{
case TYPE_SPECIFIC_CPLUS_STUFF:
printf_filtered ("%*scplus_stuff ", spaces, "");
gdb_print_host_address (TYPE_CPLUS_SPECIFIC (type),
gdb_stdout);
puts_filtered ("\n");
print_cplus_stuff (type, spaces);
break;
case TYPE_SPECIFIC_GNAT_STUFF:
printf_filtered ("%*sgnat_stuff ", spaces, "");
gdb_print_host_address (TYPE_GNAT_SPECIFIC (type), gdb_stdout);
puts_filtered ("\n");
print_gnat_stuff (type, spaces);
break;
case TYPE_SPECIFIC_FLOATFORMAT:
printf_filtered ("%*sfloatformat ", spaces, "");
if (TYPE_FLOATFORMAT (type) == NULL
|| TYPE_FLOATFORMAT (type)->name == NULL)
puts_filtered ("(null)");
else
puts_filtered (TYPE_FLOATFORMAT (type)->name);
puts_filtered ("\n");
break;
case TYPE_SPECIFIC_FUNC:
printf_filtered ("%*scalling_convention %d\n", spaces, "",
TYPE_CALLING_CONVENTION (type));
/* tail_call_list is not printed. */
break;
case TYPE_SPECIFIC_SELF_TYPE:
printf_filtered ("%*sself_type ", spaces, "");
gdb_print_host_address (TYPE_SELF_TYPE (type), gdb_stdout);
puts_filtered ("\n");
break;
case TYPE_SPECIFIC_FIXED_POINT:
printf_filtered ("%*sfixed_point_info ", spaces, "");
print_fixed_point_type_info (type, spaces);
puts_filtered ("\n");
break;
case TYPE_SPECIFIC_INT:
if (type->bit_size_differs_p ())
{
unsigned bit_size = type->bit_size ();
unsigned bit_off = type->bit_offset ();
printf_filtered ("%*s bit size = %u, bit offset = %u\n", spaces, "",
bit_size, bit_off);
}
break;
}
if (spaces == 0)
obstack_free (&dont_print_type_obstack, NULL);
}
/* Trivial helpers for the libiberty hash table, for mapping one
type to another. */
struct type_pair : public allocate_on_obstack
{
type_pair (struct type *old_, struct type *newobj_)
: old (old_), newobj (newobj_)
{}
struct type * const old, * const newobj;
};
static hashval_t
type_pair_hash (const void *item)
{
const struct type_pair *pair = (const struct type_pair *) item;
return htab_hash_pointer (pair->old);
}
static int
type_pair_eq (const void *item_lhs, const void *item_rhs)
{
const struct type_pair *lhs = (const struct type_pair *) item_lhs;
const struct type_pair *rhs = (const struct type_pair *) item_rhs;
return lhs->old == rhs->old;
}
/* Allocate the hash table used by copy_type_recursive to walk
types without duplicates. We use OBJFILE's obstack, because
OBJFILE is about to be deleted. */
htab_up
create_copied_types_hash (struct objfile *objfile)
{
return htab_up (htab_create_alloc_ex (1, type_pair_hash, type_pair_eq,
NULL, &objfile->objfile_obstack,
hashtab_obstack_allocate,
dummy_obstack_deallocate));
}
/* Recursively copy (deep copy) a dynamic attribute list of a type. */
static struct dynamic_prop_list *
copy_dynamic_prop_list (struct obstack *objfile_obstack,
struct dynamic_prop_list *list)
{
struct dynamic_prop_list *copy = list;
struct dynamic_prop_list **node_ptr = &copy;
while (*node_ptr != NULL)
{
struct dynamic_prop_list *node_copy;
node_copy = ((struct dynamic_prop_list *)
obstack_copy (objfile_obstack, *node_ptr,
sizeof (struct dynamic_prop_list)));
node_copy->prop = (*node_ptr)->prop;
*node_ptr = node_copy;
node_ptr = &node_copy->next;
}
return copy;
}
/* Recursively copy (deep copy) TYPE, if it is associated with
OBJFILE. Return a new type owned by the gdbarch associated with the type, a
saved type if we have already visited TYPE (using COPIED_TYPES), or TYPE if
it is not associated with OBJFILE. */
struct type *
copy_type_recursive (struct objfile *objfile,
struct type *type,
htab_t copied_types)
{
void **slot;
struct type *new_type;
if (! TYPE_OBJFILE_OWNED (type))
return type;
/* This type shouldn't be pointing to any types in other objfiles;
if it did, the type might disappear unexpectedly. */
gdb_assert (TYPE_OBJFILE (type) == objfile);
struct type_pair pair (type, nullptr);
slot = htab_find_slot (copied_types, &pair, INSERT);
if (*slot != NULL)
return ((struct type_pair *) *slot)->newobj;
new_type = alloc_type_arch (get_type_arch (type));
/* We must add the new type to the hash table immediately, in case
we encounter this type again during a recursive call below. */
struct type_pair *stored
= new (&objfile->objfile_obstack) struct type_pair (type, new_type);
*slot = stored;
/* Copy the common fields of types. For the main type, we simply
copy the entire thing and then update specific fields as needed. */
*TYPE_MAIN_TYPE (new_type) = *TYPE_MAIN_TYPE (type);
TYPE_OBJFILE_OWNED (new_type) = 0;
TYPE_OWNER (new_type).gdbarch = get_type_arch (type);
if (type->name ())
new_type->set_name (xstrdup (type->name ()));
new_type->set_instance_flags (type->instance_flags ());
TYPE_LENGTH (new_type) = TYPE_LENGTH (type);
/* Copy the fields. */
if (type->num_fields ())
{
int i, nfields;
nfields = type->num_fields ();
new_type->set_fields
((struct field *)
TYPE_ZALLOC (new_type, nfields * sizeof (struct field)));
for (i = 0; i < nfields; i++)
{
TYPE_FIELD_ARTIFICIAL (new_type, i) =
TYPE_FIELD_ARTIFICIAL (type, i);
TYPE_FIELD_BITSIZE (new_type, i) = TYPE_FIELD_BITSIZE (type, i);
if (type->field (i).type ())
new_type->field (i).set_type
(copy_type_recursive (objfile, type->field (i).type (),
copied_types));
if (TYPE_FIELD_NAME (type, i))
TYPE_FIELD_NAME (new_type, i) =
xstrdup (TYPE_FIELD_NAME (type, i));
switch (TYPE_FIELD_LOC_KIND (type, i))
{
case FIELD_LOC_KIND_BITPOS:
SET_FIELD_BITPOS (new_type->field (i),
TYPE_FIELD_BITPOS (type, i));
break;
case FIELD_LOC_KIND_ENUMVAL:
SET_FIELD_ENUMVAL (new_type->field (i),
TYPE_FIELD_ENUMVAL (type, i));
break;
case FIELD_LOC_KIND_PHYSADDR:
SET_FIELD_PHYSADDR (new_type->field (i),
TYPE_FIELD_STATIC_PHYSADDR (type, i));
break;
case FIELD_LOC_KIND_PHYSNAME:
SET_FIELD_PHYSNAME (new_type->field (i),
xstrdup (TYPE_FIELD_STATIC_PHYSNAME (type,
i)));
break;
default:
internal_error (__FILE__, __LINE__,
_("Unexpected type field location kind: %d"),
TYPE_FIELD_LOC_KIND (type, i));
}
}
}
/* For range types, copy the bounds information. */
if (type->code () == TYPE_CODE_RANGE)
{
range_bounds *bounds
= ((struct range_bounds *) TYPE_ALLOC
(new_type, sizeof (struct range_bounds)));
*bounds = *type->bounds ();
new_type->set_bounds (bounds);
}
if (type->main_type->dyn_prop_list != NULL)
new_type->main_type->dyn_prop_list
= copy_dynamic_prop_list (&objfile->objfile_obstack,
type->main_type->dyn_prop_list);
/* Copy pointers to other types. */
if (TYPE_TARGET_TYPE (type))
TYPE_TARGET_TYPE (new_type) =
copy_type_recursive (objfile,
TYPE_TARGET_TYPE (type),
copied_types);
/* Maybe copy the type_specific bits.
NOTE drow/2005-12-09: We do not copy the C++-specific bits like
base classes and methods. There's no fundamental reason why we
can't, but at the moment it is not needed. */
switch (TYPE_SPECIFIC_FIELD (type))
{
case TYPE_SPECIFIC_NONE:
break;
case TYPE_SPECIFIC_FUNC:
INIT_FUNC_SPECIFIC (new_type);
TYPE_CALLING_CONVENTION (new_type) = TYPE_CALLING_CONVENTION (type);
TYPE_NO_RETURN (new_type) = TYPE_NO_RETURN (type);
TYPE_TAIL_CALL_LIST (new_type) = NULL;
break;
case TYPE_SPECIFIC_FLOATFORMAT:
TYPE_FLOATFORMAT (new_type) = TYPE_FLOATFORMAT (type);
break;
case TYPE_SPECIFIC_CPLUS_STUFF:
INIT_CPLUS_SPECIFIC (new_type);
break;
case TYPE_SPECIFIC_GNAT_STUFF:
INIT_GNAT_SPECIFIC (new_type);
break;
case TYPE_SPECIFIC_SELF_TYPE:
set_type_self_type (new_type,
copy_type_recursive (objfile, TYPE_SELF_TYPE (type),
copied_types));
break;
case TYPE_SPECIFIC_FIXED_POINT:
INIT_FIXED_POINT_SPECIFIC (new_type);
new_type->fixed_point_info ().scaling_factor
= type->fixed_point_info ().scaling_factor;
break;
case TYPE_SPECIFIC_INT:
TYPE_SPECIFIC_FIELD (new_type) = TYPE_SPECIFIC_INT;
TYPE_MAIN_TYPE (new_type)->type_specific.int_stuff
= TYPE_MAIN_TYPE (type)->type_specific.int_stuff;
break;
default:
gdb_assert_not_reached ("bad type_specific_kind");
}
return new_type;
}
/* Make a copy of the given TYPE, except that the pointer & reference
types are not preserved.
This function assumes that the given type has an associated objfile.
This objfile is used to allocate the new type. */
struct type *
copy_type (const struct type *type)
{
struct type *new_type;
gdb_assert (TYPE_OBJFILE_OWNED (type));
new_type = alloc_type_copy (type);
new_type->set_instance_flags (type->instance_flags ());
TYPE_LENGTH (new_type) = TYPE_LENGTH (type);
memcpy (TYPE_MAIN_TYPE (new_type), TYPE_MAIN_TYPE (type),
sizeof (struct main_type));
if (type->main_type->dyn_prop_list != NULL)
new_type->main_type->dyn_prop_list
= copy_dynamic_prop_list (&TYPE_OBJFILE (type) -> objfile_obstack,
type->main_type->dyn_prop_list);
return new_type;
}
/* Helper functions to initialize architecture-specific types. */
/* Allocate a type structure associated with GDBARCH and set its
CODE, LENGTH, and NAME fields. */
struct type *
arch_type (struct gdbarch *gdbarch,
enum type_code code, int bit, const char *name)
{
struct type *type;
type = alloc_type_arch (gdbarch);
set_type_code (type, code);
gdb_assert ((bit % TARGET_CHAR_BIT) == 0);
TYPE_LENGTH (type) = bit / TARGET_CHAR_BIT;
if (name)
type->set_name (gdbarch_obstack_strdup (gdbarch, name));
return type;
}
/* Allocate a TYPE_CODE_INT type structure associated with GDBARCH.
BIT is the type size in bits. If UNSIGNED_P is non-zero, set
the type's TYPE_UNSIGNED flag. NAME is the type name. */
struct type *
arch_integer_type (struct gdbarch *gdbarch,
int bit, int unsigned_p, const char *name)
{
struct type *t;
t = arch_type (gdbarch, TYPE_CODE_INT, bit, name);
if (unsigned_p)
t->set_is_unsigned (true);
return t;
}
/* Allocate a TYPE_CODE_CHAR type structure associated with GDBARCH.
BIT is the type size in bits. If UNSIGNED_P is non-zero, set
the type's TYPE_UNSIGNED flag. NAME is the type name. */
struct type *
arch_character_type (struct gdbarch *gdbarch,
int bit, int unsigned_p, const char *name)
{
struct type *t;
t = arch_type (gdbarch, TYPE_CODE_CHAR, bit, name);
if (unsigned_p)
t->set_is_unsigned (true);
return t;
}
/* Allocate a TYPE_CODE_BOOL type structure associated with GDBARCH.
BIT is the type size in bits. If UNSIGNED_P is non-zero, set
the type's TYPE_UNSIGNED flag. NAME is the type name. */
struct type *
arch_boolean_type (struct gdbarch *gdbarch,
int bit, int unsigned_p, const char *name)
{
struct type *t;
t = arch_type (gdbarch, TYPE_CODE_BOOL, bit, name);
if (unsigned_p)
t->set_is_unsigned (true);
return t;
}
/* Allocate a TYPE_CODE_FLT type structure associated with GDBARCH.
BIT is the type size in bits; if BIT equals -1, the size is
determined by the floatformat. NAME is the type name. Set the
TYPE_FLOATFORMAT from FLOATFORMATS. */
struct type *
arch_float_type (struct gdbarch *gdbarch,
int bit, const char *name,
const struct floatformat **floatformats)
{
const struct floatformat *fmt = floatformats[gdbarch_byte_order (gdbarch)];
struct type *t;
bit = verify_floatformat (bit, fmt);
t = arch_type (gdbarch, TYPE_CODE_FLT, bit, name);
TYPE_FLOATFORMAT (t) = fmt;
return t;
}
/* Allocate a TYPE_CODE_DECFLOAT type structure associated with GDBARCH.
BIT is the type size in bits. NAME is the type name. */
struct type *
arch_decfloat_type (struct gdbarch *gdbarch, int bit, const char *name)
{
struct type *t;
t = arch_type (gdbarch, TYPE_CODE_DECFLOAT, bit, name);
return t;
}
/* Allocate a TYPE_CODE_PTR type structure associated with GDBARCH.
BIT is the pointer type size in bits. NAME is the type name.
TARGET_TYPE is the pointer target type. Always sets the pointer type's
TYPE_UNSIGNED flag. */
struct type *
arch_pointer_type (struct gdbarch *gdbarch,
int bit, const char *name, struct type *target_type)
{
struct type *t;
t = arch_type (gdbarch, TYPE_CODE_PTR, bit, name);
TYPE_TARGET_TYPE (t) = target_type;
t->set_is_unsigned (true);
return t;
}
/* Allocate a TYPE_CODE_FLAGS type structure associated with GDBARCH.
NAME is the type name. BIT is the size of the flag word in bits. */
struct type *
arch_flags_type (struct gdbarch *gdbarch, const char *name, int bit)
{
struct type *type;
type = arch_type (gdbarch, TYPE_CODE_FLAGS, bit, name);
type->set_is_unsigned (true);
type->set_num_fields (0);
/* Pre-allocate enough space assuming every field is one bit. */
type->set_fields
((struct field *) TYPE_ZALLOC (type, bit * sizeof (struct field)));
return type;
}
/* Add field to TYPE_CODE_FLAGS type TYPE to indicate the bit at
position BITPOS is called NAME. Pass NAME as "" for fields that
should not be printed. */
void
append_flags_type_field (struct type *type, int start_bitpos, int nr_bits,
struct type *field_type, const char *name)
{
int type_bitsize = TYPE_LENGTH (type) * TARGET_CHAR_BIT;
int field_nr = type->num_fields ();
gdb_assert (type->code () == TYPE_CODE_FLAGS);
gdb_assert (type->num_fields () + 1 <= type_bitsize);
gdb_assert (start_bitpos >= 0 && start_bitpos < type_bitsize);
gdb_assert (nr_bits >= 1 && nr_bits <= type_bitsize);
gdb_assert (name != NULL);
TYPE_FIELD_NAME (type, field_nr) = xstrdup (name);
type->field (field_nr).set_type (field_type);
SET_FIELD_BITPOS (type->field (field_nr), start_bitpos);
TYPE_FIELD_BITSIZE (type, field_nr) = nr_bits;
type->set_num_fields (type->num_fields () + 1);
}
/* Special version of append_flags_type_field to add a flag field.
Add field to TYPE_CODE_FLAGS type TYPE to indicate the bit at
position BITPOS is called NAME. */
void
append_flags_type_flag (struct type *type, int bitpos, const char *name)
{
struct gdbarch *gdbarch = get_type_arch (type);
append_flags_type_field (type, bitpos, 1,
builtin_type (gdbarch)->builtin_bool,
name);
}
/* Allocate a TYPE_CODE_STRUCT or TYPE_CODE_UNION type structure (as
specified by CODE) associated with GDBARCH. NAME is the type name. */
struct type *
arch_composite_type (struct gdbarch *gdbarch, const char *name,
enum type_code code)
{
struct type *t;
gdb_assert (code == TYPE_CODE_STRUCT || code == TYPE_CODE_UNION);
t = arch_type (gdbarch, code, 0, NULL);
t->set_name (name);
INIT_CPLUS_SPECIFIC (t);
return t;
}
/* Add new field with name NAME and type FIELD to composite type T.
Do not set the field's position or adjust the type's length;
the caller should do so. Return the new field. */
struct field *
append_composite_type_field_raw (struct type *t, const char *name,
struct type *field)
{
struct field *f;
t->set_num_fields (t->num_fields () + 1);
t->set_fields (XRESIZEVEC (struct field, t->fields (),
t->num_fields ()));
f = &t->field (t->num_fields () - 1);
memset (f, 0, sizeof f[0]);
f[0].set_type (field);
FIELD_NAME (f[0]) = name;
return f;
}
/* Add new field with name NAME and type FIELD to composite type T.
ALIGNMENT (if non-zero) specifies the minimum field alignment. */
void
append_composite_type_field_aligned (struct type *t, const char *name,
struct type *field, int alignment)
{
struct field *f = append_composite_type_field_raw (t, name, field);
if (t->code () == TYPE_CODE_UNION)
{
if (TYPE_LENGTH (t) < TYPE_LENGTH (field))
TYPE_LENGTH (t) = TYPE_LENGTH (field);
}
else if (t->code () == TYPE_CODE_STRUCT)
{
TYPE_LENGTH (t) = TYPE_LENGTH (t) + TYPE_LENGTH (field);
if (t->num_fields () > 1)
{
SET_FIELD_BITPOS (f[0],
(FIELD_BITPOS (f[-1])
+ (TYPE_LENGTH (f[-1].type ())
* TARGET_CHAR_BIT)));
if (alignment)
{
int left;
alignment *= TARGET_CHAR_BIT;
left = FIELD_BITPOS (f[0]) % alignment;
if (left)
{
SET_FIELD_BITPOS (f[0], FIELD_BITPOS (f[0]) + (alignment - left));
TYPE_LENGTH (t) += (alignment - left) / TARGET_CHAR_BIT;
}
}
}
}
}
/* Add new field with name NAME and type FIELD to composite type T. */
void
append_composite_type_field (struct type *t, const char *name,
struct type *field)
{
append_composite_type_field_aligned (t, name, field, 0);
}
/* We manage the lifetimes of fixed_point_type_info objects by
attaching them to the objfile. Currently, these objects are
modified during construction, and GMP does not provide a way to
hash the contents of an mpq_t; so it's a bit of a pain to hash-cons
them. If we did do this, they could be moved to the per-BFD and
shared across objfiles. */
typedef std::vector<std::unique_ptr<fixed_point_type_info>>
fixed_point_type_storage;
/* Key used for managing the storage of fixed-point type info. */
static const struct objfile_key<fixed_point_type_storage>
fixed_point_objfile_key;
/* See gdbtypes.h. */
void
allocate_fixed_point_type_info (struct type *type)
{
std::unique_ptr<fixed_point_type_info> up (new fixed_point_type_info);
fixed_point_type_info *info;
if (TYPE_OBJFILE_OWNED (type))
{
fixed_point_type_storage *storage
= fixed_point_objfile_key.get (TYPE_OBJFILE (type));
if (storage == nullptr)
storage = fixed_point_objfile_key.emplace (TYPE_OBJFILE (type));
info = up.get ();
storage->push_back (std::move (up));
}
else
{
/* We just leak the memory, because that's what we do generally
for non-objfile-attached types. */
info = up.release ();
}
type->set_fixed_point_info (info);
}
/* See gdbtypes.h. */
bool
is_fixed_point_type (struct type *type)
{
while (check_typedef (type)->code () == TYPE_CODE_RANGE)
type = TYPE_TARGET_TYPE (check_typedef (type));
type = check_typedef (type);
return type->code () == TYPE_CODE_FIXED_POINT;
}
/* See gdbtypes.h. */
struct type *
type::fixed_point_type_base_type ()
{
struct type *type = this;
while (check_typedef (type)->code () == TYPE_CODE_RANGE)
type = TYPE_TARGET_TYPE (check_typedef (type));
type = check_typedef (type);
gdb_assert (type->code () == TYPE_CODE_FIXED_POINT);
return type;
}
/* See gdbtypes.h. */
const gdb_mpq &
type::fixed_point_scaling_factor ()
{
struct type *type = this->fixed_point_type_base_type ();
return type->fixed_point_info ().scaling_factor;
}
static struct gdbarch_data *gdbtypes_data;
const struct builtin_type *
builtin_type (struct gdbarch *gdbarch)
{
return (const struct builtin_type *) gdbarch_data (gdbarch, gdbtypes_data);
}
static void *
gdbtypes_post_init (struct gdbarch *gdbarch)
{
struct builtin_type *builtin_type
= GDBARCH_OBSTACK_ZALLOC (gdbarch, struct builtin_type);
/* Basic types. */
builtin_type->builtin_void
= arch_type (gdbarch, TYPE_CODE_VOID, TARGET_CHAR_BIT, "void");
builtin_type->builtin_char
= arch_integer_type (gdbarch, TARGET_CHAR_BIT,
!gdbarch_char_signed (gdbarch), "char");
builtin_type->builtin_char->set_has_no_signedness (true);
builtin_type->builtin_signed_char
= arch_integer_type (gdbarch, TARGET_CHAR_BIT,
0, "signed char");
builtin_type->builtin_unsigned_char
= arch_integer_type (gdbarch, TARGET_CHAR_BIT,
1, "unsigned char");
builtin_type->builtin_short
= arch_integer_type (gdbarch, gdbarch_short_bit (gdbarch),
0, "short");
builtin_type->builtin_unsigned_short
= arch_integer_type (gdbarch, gdbarch_short_bit (gdbarch),
1, "unsigned short");
builtin_type->builtin_int
= arch_integer_type (gdbarch, gdbarch_int_bit (gdbarch),
0, "int");
builtin_type->builtin_unsigned_int
= arch_integer_type (gdbarch, gdbarch_int_bit (gdbarch),
1, "unsigned int");
builtin_type->builtin_long
= arch_integer_type (gdbarch, gdbarch_long_bit (gdbarch),
0, "long");
builtin_type->builtin_unsigned_long
= arch_integer_type (gdbarch, gdbarch_long_bit (gdbarch),
1, "unsigned long");
builtin_type->builtin_long_long
= arch_integer_type (gdbarch, gdbarch_long_long_bit (gdbarch),
0, "long long");
builtin_type->builtin_unsigned_long_long
= arch_integer_type (gdbarch, gdbarch_long_long_bit (gdbarch),
1, "unsigned long long");
builtin_type->builtin_half
= arch_float_type (gdbarch, gdbarch_half_bit (gdbarch),
"half", gdbarch_half_format (gdbarch));
builtin_type->builtin_float
= arch_float_type (gdbarch, gdbarch_float_bit (gdbarch),
"float", gdbarch_float_format (gdbarch));
builtin_type->builtin_bfloat16
= arch_float_type (gdbarch, gdbarch_bfloat16_bit (gdbarch),
"bfloat16", gdbarch_bfloat16_format (gdbarch));
builtin_type->builtin_double
= arch_float_type (gdbarch, gdbarch_double_bit (gdbarch),
"double", gdbarch_double_format (gdbarch));
builtin_type->builtin_long_double
= arch_float_type (gdbarch, gdbarch_long_double_bit (gdbarch),
"long double", gdbarch_long_double_format (gdbarch));
builtin_type->builtin_complex
= init_complex_type ("complex", builtin_type->builtin_float);
builtin_type->builtin_double_complex
= init_complex_type ("double complex", builtin_type->builtin_double);
builtin_type->builtin_string
= arch_type (gdbarch, TYPE_CODE_STRING, TARGET_CHAR_BIT, "string");
builtin_type->builtin_bool
= arch_type (gdbarch, TYPE_CODE_BOOL, TARGET_CHAR_BIT, "bool");
/* The following three are about decimal floating point types, which
are 32-bits, 64-bits and 128-bits respectively. */
builtin_type->builtin_decfloat
= arch_decfloat_type (gdbarch, 32, "_Decimal32");
builtin_type->builtin_decdouble
= arch_decfloat_type (gdbarch, 64, "_Decimal64");
builtin_type->builtin_declong
= arch_decfloat_type (gdbarch, 128, "_Decimal128");
/* "True" character types. */
builtin_type->builtin_true_char
= arch_character_type (gdbarch, TARGET_CHAR_BIT, 0, "true character");
builtin_type->builtin_true_unsigned_char
= arch_character_type (gdbarch, TARGET_CHAR_BIT, 1, "true character");
/* Fixed-size integer types. */
builtin_type->builtin_int0
= arch_integer_type (gdbarch, 0, 0, "int0_t");
builtin_type->builtin_int8
= arch_integer_type (gdbarch, 8, 0, "int8_t");
builtin_type->builtin_uint8
= arch_integer_type (gdbarch, 8, 1, "uint8_t");
builtin_type->builtin_int16
= arch_integer_type (gdbarch, 16, 0, "int16_t");
builtin_type->builtin_uint16
= arch_integer_type (gdbarch, 16, 1, "uint16_t");
builtin_type->builtin_int24
= arch_integer_type (gdbarch, 24, 0, "int24_t");
builtin_type->builtin_uint24
= arch_integer_type (gdbarch, 24, 1, "uint24_t");
builtin_type->builtin_int32
= arch_integer_type (gdbarch, 32, 0, "int32_t");
builtin_type->builtin_uint32
= arch_integer_type (gdbarch, 32, 1, "uint32_t");
builtin_type->builtin_int64
= arch_integer_type (gdbarch, 64, 0, "int64_t");
builtin_type->builtin_uint64
= arch_integer_type (gdbarch, 64, 1, "uint64_t");
builtin_type->builtin_int128
= arch_integer_type (gdbarch, 128, 0, "int128_t");
builtin_type->builtin_uint128
= arch_integer_type (gdbarch, 128, 1, "uint128_t");
builtin_type->builtin_int8->set_instance_flags
(builtin_type->builtin_int8->instance_flags ()
| TYPE_INSTANCE_FLAG_NOTTEXT);
builtin_type->builtin_uint8->set_instance_flags
(builtin_type->builtin_uint8->instance_flags ()
| TYPE_INSTANCE_FLAG_NOTTEXT);
/* Wide character types. */
builtin_type->builtin_char16
= arch_integer_type (gdbarch, 16, 1, "char16_t");
builtin_type->builtin_char32
= arch_integer_type (gdbarch, 32, 1, "char32_t");
builtin_type->builtin_wchar
= arch_integer_type (gdbarch, gdbarch_wchar_bit (gdbarch),
!gdbarch_wchar_signed (gdbarch), "wchar_t");
/* Default data/code pointer types. */
builtin_type->builtin_data_ptr
= lookup_pointer_type (builtin_type->builtin_void);
builtin_type->builtin_func_ptr
= lookup_pointer_type (lookup_function_type (builtin_type->builtin_void));
builtin_type->builtin_func_func
= lookup_function_type (builtin_type->builtin_func_ptr);
/* This type represents a GDB internal function. */
builtin_type->internal_fn
= arch_type (gdbarch, TYPE_CODE_INTERNAL_FUNCTION, 0,
"<internal function>");
/* This type represents an xmethod. */
builtin_type->xmethod
= arch_type (gdbarch, TYPE_CODE_XMETHOD, 0, "<xmethod>");
return builtin_type;
}
/* This set of objfile-based types is intended to be used by symbol
readers as basic types. */
static const struct objfile_key<struct objfile_type,
gdb::noop_deleter<struct objfile_type>>
objfile_type_data;
const struct objfile_type *
objfile_type (struct objfile *objfile)
{
struct gdbarch *gdbarch;
struct objfile_type *objfile_type = objfile_type_data.get (objfile);
if (objfile_type)
return objfile_type;
objfile_type = OBSTACK_CALLOC (&objfile->objfile_obstack,
1, struct objfile_type);
/* Use the objfile architecture to determine basic type properties. */
gdbarch = objfile->arch ();
/* Basic types. */
objfile_type->builtin_void
= init_type (objfile, TYPE_CODE_VOID, TARGET_CHAR_BIT, "void");
objfile_type->builtin_char
= init_integer_type (objfile, TARGET_CHAR_BIT,
!gdbarch_char_signed (gdbarch), "char");
objfile_type->builtin_char->set_has_no_signedness (true);
objfile_type->builtin_signed_char
= init_integer_type (objfile, TARGET_CHAR_BIT,
0, "signed char");
objfile_type->builtin_unsigned_char
= init_integer_type (objfile, TARGET_CHAR_BIT,
1, "unsigned char");
objfile_type->builtin_short
= init_integer_type (objfile, gdbarch_short_bit (gdbarch),
0, "short");
objfile_type->builtin_unsigned_short
= init_integer_type (objfile, gdbarch_short_bit (gdbarch),
1, "unsigned short");
objfile_type->builtin_int
= init_integer_type (objfile, gdbarch_int_bit (gdbarch),
0, "int");
objfile_type->builtin_unsigned_int
= init_integer_type (objfile, gdbarch_int_bit (gdbarch),
1, "unsigned int");
objfile_type->builtin_long
= init_integer_type (objfile, gdbarch_long_bit (gdbarch),
0, "long");
objfile_type->builtin_unsigned_long
= init_integer_type (objfile, gdbarch_long_bit (gdbarch),
1, "unsigned long");
objfile_type->builtin_long_long
= init_integer_type (objfile, gdbarch_long_long_bit (gdbarch),
0, "long long");
objfile_type->builtin_unsigned_long_long
= init_integer_type (objfile, gdbarch_long_long_bit (gdbarch),
1, "unsigned long long");
objfile_type->builtin_float
= init_float_type (objfile, gdbarch_float_bit (gdbarch),
"float", gdbarch_float_format (gdbarch));
objfile_type->builtin_double
= init_float_type (objfile, gdbarch_double_bit (gdbarch),
"double", gdbarch_double_format (gdbarch));
objfile_type->builtin_long_double
= init_float_type (objfile, gdbarch_long_double_bit (gdbarch),
"long double", gdbarch_long_double_format (gdbarch));
/* This type represents a type that was unrecognized in symbol read-in. */
objfile_type->builtin_error
= init_type (objfile, TYPE_CODE_ERROR, 0, "<unknown type>");
/* The following set of types is used for symbols with no
debug information. */
objfile_type->nodebug_text_symbol
= init_type (objfile, TYPE_CODE_FUNC, TARGET_CHAR_BIT,
"<text variable, no debug info>");
objfile_type->nodebug_text_gnu_ifunc_symbol
= init_type (objfile, TYPE_CODE_FUNC, TARGET_CHAR_BIT,
"<text gnu-indirect-function variable, no debug info>");
objfile_type->nodebug_text_gnu_ifunc_symbol->set_is_gnu_ifunc (true);
objfile_type->nodebug_got_plt_symbol
= init_pointer_type (objfile, gdbarch_addr_bit (gdbarch),
"<text from jump slot in .got.plt, no debug info>",
objfile_type->nodebug_text_symbol);
objfile_type->nodebug_data_symbol
= init_nodebug_var_type (objfile, "<data variable, no debug info>");
objfile_type->nodebug_unknown_symbol
= init_nodebug_var_type (objfile, "<variable (not text or data), no debug info>");
objfile_type->nodebug_tls_symbol
= init_nodebug_var_type (objfile, "<thread local variable, no debug info>");
/* NOTE: on some targets, addresses and pointers are not necessarily
the same.
The upshot is:
- gdb's `struct type' always describes the target's
representation.
- gdb's `struct value' objects should always hold values in
target form.
- gdb's CORE_ADDR values are addresses in the unified virtual
address space that the assembler and linker work with. Thus,
since target_read_memory takes a CORE_ADDR as an argument, it
can access any memory on the target, even if the processor has
separate code and data address spaces.
In this context, objfile_type->builtin_core_addr is a bit odd:
it's a target type for a value the target will never see. It's
only used to hold the values of (typeless) linker symbols, which
are indeed in the unified virtual address space. */
objfile_type->builtin_core_addr
= init_integer_type (objfile, gdbarch_addr_bit (gdbarch), 1,
"__CORE_ADDR");
objfile_type_data.set (objfile, objfile_type);
return objfile_type;
}
void _initialize_gdbtypes ();
void
_initialize_gdbtypes ()
{
gdbtypes_data = gdbarch_data_register_post_init (gdbtypes_post_init);
add_setshow_zuinteger_cmd ("overload", no_class, &overload_debug,
_("Set debugging of C++ overloading."),
_("Show debugging of C++ overloading."),
_("When enabled, ranking of the "
"functions is displayed."),
NULL,
show_overload_debug,
&setdebuglist, &showdebuglist);
/* Add user knob for controlling resolution of opaque types. */
add_setshow_boolean_cmd ("opaque-type-resolution", class_support,
&opaque_type_resolution,
_("Set resolution of opaque struct/class/union"
" types (if set before loading symbols)."),
_("Show resolution of opaque struct/class/union"
" types (if set before loading symbols)."),
NULL, NULL,
show_opaque_type_resolution,
&setlist, &showlist);
/* Add an option to permit non-strict type checking. */
add_setshow_boolean_cmd ("type", class_support,
&strict_type_checking,
_("Set strict type checking."),
_("Show strict type checking."),
NULL, NULL,
show_strict_type_checking,
&setchecklist, &showchecklist);
}