| /* Low level packing and unpacking of values for GDB, the GNU Debugger. |
| |
| Copyright (C) 1986-2021 Free Software Foundation, Inc. |
| |
| This file is part of GDB. |
| |
| This program is free software; you can redistribute it and/or modify |
| it under the terms of the GNU General Public License as published by |
| the Free Software Foundation; either version 3 of the License, or |
| (at your option) any later version. |
| |
| This program is distributed in the hope that it will be useful, |
| but WITHOUT ANY WARRANTY; without even the implied warranty of |
| MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| GNU General Public License for more details. |
| |
| You should have received a copy of the GNU General Public License |
| along with this program. If not, see <http://www.gnu.org/licenses/>. */ |
| |
| #include "defs.h" |
| #include "arch-utils.h" |
| #include "symtab.h" |
| #include "gdbtypes.h" |
| #include "value.h" |
| #include "gdbcore.h" |
| #include "command.h" |
| #include "gdbcmd.h" |
| #include "target.h" |
| #include "language.h" |
| #include "demangle.h" |
| #include "regcache.h" |
| #include "block.h" |
| #include "target-float.h" |
| #include "objfiles.h" |
| #include "valprint.h" |
| #include "cli/cli-decode.h" |
| #include "extension.h" |
| #include <ctype.h> |
| #include "tracepoint.h" |
| #include "cp-abi.h" |
| #include "user-regs.h" |
| #include <algorithm> |
| #include "completer.h" |
| #include "gdbsupport/selftest.h" |
| #include "gdbsupport/array-view.h" |
| #include "cli/cli-style.h" |
| #include "expop.h" |
| #include "inferior.h" |
| |
| /* Definition of a user function. */ |
| struct internal_function |
| { |
| /* The name of the function. It is a bit odd to have this in the |
| function itself -- the user might use a differently-named |
| convenience variable to hold the function. */ |
| char *name; |
| |
| /* The handler. */ |
| internal_function_fn handler; |
| |
| /* User data for the handler. */ |
| void *cookie; |
| }; |
| |
| /* Defines an [OFFSET, OFFSET + LENGTH) range. */ |
| |
| struct range |
| { |
| /* Lowest offset in the range. */ |
| LONGEST offset; |
| |
| /* Length of the range. */ |
| LONGEST length; |
| |
| /* Returns true if THIS is strictly less than OTHER, useful for |
| searching. We keep ranges sorted by offset and coalesce |
| overlapping and contiguous ranges, so this just compares the |
| starting offset. */ |
| |
| bool operator< (const range &other) const |
| { |
| return offset < other.offset; |
| } |
| |
| /* Returns true if THIS is equal to OTHER. */ |
| bool operator== (const range &other) const |
| { |
| return offset == other.offset && length == other.length; |
| } |
| }; |
| |
| /* Returns true if the ranges defined by [offset1, offset1+len1) and |
| [offset2, offset2+len2) overlap. */ |
| |
| static int |
| ranges_overlap (LONGEST offset1, LONGEST len1, |
| LONGEST offset2, LONGEST len2) |
| { |
| ULONGEST h, l; |
| |
| l = std::max (offset1, offset2); |
| h = std::min (offset1 + len1, offset2 + len2); |
| return (l < h); |
| } |
| |
| /* Returns true if RANGES contains any range that overlaps [OFFSET, |
| OFFSET+LENGTH). */ |
| |
| static int |
| ranges_contain (const std::vector<range> &ranges, LONGEST offset, |
| LONGEST length) |
| { |
| range what; |
| |
| what.offset = offset; |
| what.length = length; |
| |
| /* We keep ranges sorted by offset and coalesce overlapping and |
| contiguous ranges, so to check if a range list contains a given |
| range, we can do a binary search for the position the given range |
| would be inserted if we only considered the starting OFFSET of |
| ranges. We call that position I. Since we also have LENGTH to |
| care for (this is a range afterall), we need to check if the |
| _previous_ range overlaps the I range. E.g., |
| |
| R |
| |---| |
| |---| |---| |------| ... |--| |
| 0 1 2 N |
| |
| I=1 |
| |
| In the case above, the binary search would return `I=1', meaning, |
| this OFFSET should be inserted at position 1, and the current |
| position 1 should be pushed further (and before 2). But, `0' |
| overlaps with R. |
| |
| Then we need to check if the I range overlaps the I range itself. |
| E.g., |
| |
| R |
| |---| |
| |---| |---| |-------| ... |--| |
| 0 1 2 N |
| |
| I=1 |
| */ |
| |
| |
| auto i = std::lower_bound (ranges.begin (), ranges.end (), what); |
| |
| if (i > ranges.begin ()) |
| { |
| const struct range &bef = *(i - 1); |
| |
| if (ranges_overlap (bef.offset, bef.length, offset, length)) |
| return 1; |
| } |
| |
| if (i < ranges.end ()) |
| { |
| const struct range &r = *i; |
| |
| if (ranges_overlap (r.offset, r.length, offset, length)) |
| return 1; |
| } |
| |
| return 0; |
| } |
| |
| static struct cmd_list_element *functionlist; |
| |
| /* Note that the fields in this structure are arranged to save a bit |
| of memory. */ |
| |
| struct value |
| { |
| explicit value (struct type *type_) |
| : modifiable (1), |
| lazy (1), |
| initialized (1), |
| stack (0), |
| is_zero (false), |
| type (type_), |
| enclosing_type (type_) |
| { |
| } |
| |
| ~value () |
| { |
| if (VALUE_LVAL (this) == lval_computed) |
| { |
| const struct lval_funcs *funcs = location.computed.funcs; |
| |
| if (funcs->free_closure) |
| funcs->free_closure (this); |
| } |
| else if (VALUE_LVAL (this) == lval_xcallable) |
| delete location.xm_worker; |
| } |
| |
| DISABLE_COPY_AND_ASSIGN (value); |
| |
| /* Type of value; either not an lval, or one of the various |
| different possible kinds of lval. */ |
| enum lval_type lval = not_lval; |
| |
| /* Is it modifiable? Only relevant if lval != not_lval. */ |
| unsigned int modifiable : 1; |
| |
| /* If zero, contents of this value are in the contents field. If |
| nonzero, contents are in inferior. If the lval field is lval_memory, |
| the contents are in inferior memory at location.address plus offset. |
| The lval field may also be lval_register. |
| |
| WARNING: This field is used by the code which handles watchpoints |
| (see breakpoint.c) to decide whether a particular value can be |
| watched by hardware watchpoints. If the lazy flag is set for |
| some member of a value chain, it is assumed that this member of |
| the chain doesn't need to be watched as part of watching the |
| value itself. This is how GDB avoids watching the entire struct |
| or array when the user wants to watch a single struct member or |
| array element. If you ever change the way lazy flag is set and |
| reset, be sure to consider this use as well! */ |
| unsigned int lazy : 1; |
| |
| /* If value is a variable, is it initialized or not. */ |
| unsigned int initialized : 1; |
| |
| /* If value is from the stack. If this is set, read_stack will be |
| used instead of read_memory to enable extra caching. */ |
| unsigned int stack : 1; |
| |
| /* True if this is a zero value, created by 'value_zero'; false |
| otherwise. */ |
| bool is_zero : 1; |
| |
| /* Location of value (if lval). */ |
| union |
| { |
| /* If lval == lval_memory, this is the address in the inferior */ |
| CORE_ADDR address; |
| |
| /*If lval == lval_register, the value is from a register. */ |
| struct |
| { |
| /* Register number. */ |
| int regnum; |
| /* Frame ID of "next" frame to which a register value is relative. |
| If the register value is found relative to frame F, then the |
| frame id of F->next will be stored in next_frame_id. */ |
| struct frame_id next_frame_id; |
| } reg; |
| |
| /* Pointer to internal variable. */ |
| struct internalvar *internalvar; |
| |
| /* Pointer to xmethod worker. */ |
| struct xmethod_worker *xm_worker; |
| |
| /* If lval == lval_computed, this is a set of function pointers |
| to use to access and describe the value, and a closure pointer |
| for them to use. */ |
| struct |
| { |
| /* Functions to call. */ |
| const struct lval_funcs *funcs; |
| |
| /* Closure for those functions to use. */ |
| void *closure; |
| } computed; |
| } location {}; |
| |
| /* Describes offset of a value within lval of a structure in target |
| addressable memory units. Note also the member embedded_offset |
| below. */ |
| LONGEST offset = 0; |
| |
| /* Only used for bitfields; number of bits contained in them. */ |
| LONGEST bitsize = 0; |
| |
| /* Only used for bitfields; position of start of field. For |
| little-endian targets, it is the position of the LSB. For |
| big-endian targets, it is the position of the MSB. */ |
| LONGEST bitpos = 0; |
| |
| /* The number of references to this value. When a value is created, |
| the value chain holds a reference, so REFERENCE_COUNT is 1. If |
| release_value is called, this value is removed from the chain but |
| the caller of release_value now has a reference to this value. |
| The caller must arrange for a call to value_free later. */ |
| int reference_count = 1; |
| |
| /* Only used for bitfields; the containing value. This allows a |
| single read from the target when displaying multiple |
| bitfields. */ |
| value_ref_ptr parent; |
| |
| /* Type of the value. */ |
| struct type *type; |
| |
| /* If a value represents a C++ object, then the `type' field gives |
| the object's compile-time type. If the object actually belongs |
| to some class derived from `type', perhaps with other base |
| classes and additional members, then `type' is just a subobject |
| of the real thing, and the full object is probably larger than |
| `type' would suggest. |
| |
| If `type' is a dynamic class (i.e. one with a vtable), then GDB |
| can actually determine the object's run-time type by looking at |
| the run-time type information in the vtable. When this |
| information is available, we may elect to read in the entire |
| object, for several reasons: |
| |
| - When printing the value, the user would probably rather see the |
| full object, not just the limited portion apparent from the |
| compile-time type. |
| |
| - If `type' has virtual base classes, then even printing `type' |
| alone may require reaching outside the `type' portion of the |
| object to wherever the virtual base class has been stored. |
| |
| When we store the entire object, `enclosing_type' is the run-time |
| type -- the complete object -- and `embedded_offset' is the |
| offset of `type' within that larger type, in target addressable memory |
| units. The value_contents() macro takes `embedded_offset' into account, |
| so most GDB code continues to see the `type' portion of the value, just |
| as the inferior would. |
| |
| If `type' is a pointer to an object, then `enclosing_type' is a |
| pointer to the object's run-time type, and `pointed_to_offset' is |
| the offset in target addressable memory units from the full object |
| to the pointed-to object -- that is, the value `embedded_offset' would |
| have if we followed the pointer and fetched the complete object. |
| (I don't really see the point. Why not just determine the |
| run-time type when you indirect, and avoid the special case? The |
| contents don't matter until you indirect anyway.) |
| |
| If we're not doing anything fancy, `enclosing_type' is equal to |
| `type', and `embedded_offset' is zero, so everything works |
| normally. */ |
| struct type *enclosing_type; |
| LONGEST embedded_offset = 0; |
| LONGEST pointed_to_offset = 0; |
| |
| /* Actual contents of the value. Target byte-order. NULL or not |
| valid if lazy is nonzero. */ |
| gdb::unique_xmalloc_ptr<gdb_byte> contents; |
| |
| /* Unavailable ranges in CONTENTS. We mark unavailable ranges, |
| rather than available, since the common and default case is for a |
| value to be available. This is filled in at value read time. |
| The unavailable ranges are tracked in bits. Note that a contents |
| bit that has been optimized out doesn't really exist in the |
| program, so it can't be marked unavailable either. */ |
| std::vector<range> unavailable; |
| |
| /* Likewise, but for optimized out contents (a chunk of the value of |
| a variable that does not actually exist in the program). If LVAL |
| is lval_register, this is a register ($pc, $sp, etc., never a |
| program variable) that has not been saved in the frame. Not |
| saved registers and optimized-out program variables values are |
| treated pretty much the same, except not-saved registers have a |
| different string representation and related error strings. */ |
| std::vector<range> optimized_out; |
| }; |
| |
| /* See value.h. */ |
| |
| struct gdbarch * |
| get_value_arch (const struct value *value) |
| { |
| return value_type (value)->arch (); |
| } |
| |
| int |
| value_bits_available (const struct value *value, LONGEST offset, LONGEST length) |
| { |
| gdb_assert (!value->lazy); |
| |
| return !ranges_contain (value->unavailable, offset, length); |
| } |
| |
| int |
| value_bytes_available (const struct value *value, |
| LONGEST offset, LONGEST length) |
| { |
| return value_bits_available (value, |
| offset * TARGET_CHAR_BIT, |
| length * TARGET_CHAR_BIT); |
| } |
| |
| int |
| value_bits_any_optimized_out (const struct value *value, int bit_offset, int bit_length) |
| { |
| gdb_assert (!value->lazy); |
| |
| return ranges_contain (value->optimized_out, bit_offset, bit_length); |
| } |
| |
| int |
| value_entirely_available (struct value *value) |
| { |
| /* We can only tell whether the whole value is available when we try |
| to read it. */ |
| if (value->lazy) |
| value_fetch_lazy (value); |
| |
| if (value->unavailable.empty ()) |
| return 1; |
| return 0; |
| } |
| |
| /* Returns true if VALUE is entirely covered by RANGES. If the value |
| is lazy, it'll be read now. Note that RANGE is a pointer to |
| pointer because reading the value might change *RANGE. */ |
| |
| static int |
| value_entirely_covered_by_range_vector (struct value *value, |
| const std::vector<range> &ranges) |
| { |
| /* We can only tell whether the whole value is optimized out / |
| unavailable when we try to read it. */ |
| if (value->lazy) |
| value_fetch_lazy (value); |
| |
| if (ranges.size () == 1) |
| { |
| const struct range &t = ranges[0]; |
| |
| if (t.offset == 0 |
| && t.length == (TARGET_CHAR_BIT |
| * TYPE_LENGTH (value_enclosing_type (value)))) |
| return 1; |
| } |
| |
| return 0; |
| } |
| |
| int |
| value_entirely_unavailable (struct value *value) |
| { |
| return value_entirely_covered_by_range_vector (value, value->unavailable); |
| } |
| |
| int |
| value_entirely_optimized_out (struct value *value) |
| { |
| return value_entirely_covered_by_range_vector (value, value->optimized_out); |
| } |
| |
| /* Insert into the vector pointed to by VECTORP the bit range starting of |
| OFFSET bits, and extending for the next LENGTH bits. */ |
| |
| static void |
| insert_into_bit_range_vector (std::vector<range> *vectorp, |
| LONGEST offset, LONGEST length) |
| { |
| range newr; |
| |
| /* Insert the range sorted. If there's overlap or the new range |
| would be contiguous with an existing range, merge. */ |
| |
| newr.offset = offset; |
| newr.length = length; |
| |
| /* Do a binary search for the position the given range would be |
| inserted if we only considered the starting OFFSET of ranges. |
| Call that position I. Since we also have LENGTH to care for |
| (this is a range afterall), we need to check if the _previous_ |
| range overlaps the I range. E.g., calling R the new range: |
| |
| #1 - overlaps with previous |
| |
| R |
| |-...-| |
| |---| |---| |------| ... |--| |
| 0 1 2 N |
| |
| I=1 |
| |
| In the case #1 above, the binary search would return `I=1', |
| meaning, this OFFSET should be inserted at position 1, and the |
| current position 1 should be pushed further (and become 2). But, |
| note that `0' overlaps with R, so we want to merge them. |
| |
| A similar consideration needs to be taken if the new range would |
| be contiguous with the previous range: |
| |
| #2 - contiguous with previous |
| |
| R |
| |-...-| |
| |--| |---| |------| ... |--| |
| 0 1 2 N |
| |
| I=1 |
| |
| If there's no overlap with the previous range, as in: |
| |
| #3 - not overlapping and not contiguous |
| |
| R |
| |-...-| |
| |--| |---| |------| ... |--| |
| 0 1 2 N |
| |
| I=1 |
| |
| or if I is 0: |
| |
| #4 - R is the range with lowest offset |
| |
| R |
| |-...-| |
| |--| |---| |------| ... |--| |
| 0 1 2 N |
| |
| I=0 |
| |
| ... we just push the new range to I. |
| |
| All the 4 cases above need to consider that the new range may |
| also overlap several of the ranges that follow, or that R may be |
| contiguous with the following range, and merge. E.g., |
| |
| #5 - overlapping following ranges |
| |
| R |
| |------------------------| |
| |--| |---| |------| ... |--| |
| 0 1 2 N |
| |
| I=0 |
| |
| or: |
| |
| R |
| |-------| |
| |--| |---| |------| ... |--| |
| 0 1 2 N |
| |
| I=1 |
| |
| */ |
| |
| auto i = std::lower_bound (vectorp->begin (), vectorp->end (), newr); |
| if (i > vectorp->begin ()) |
| { |
| struct range &bef = *(i - 1); |
| |
| if (ranges_overlap (bef.offset, bef.length, offset, length)) |
| { |
| /* #1 */ |
| ULONGEST l = std::min (bef.offset, offset); |
| ULONGEST h = std::max (bef.offset + bef.length, offset + length); |
| |
| bef.offset = l; |
| bef.length = h - l; |
| i--; |
| } |
| else if (offset == bef.offset + bef.length) |
| { |
| /* #2 */ |
| bef.length += length; |
| i--; |
| } |
| else |
| { |
| /* #3 */ |
| i = vectorp->insert (i, newr); |
| } |
| } |
| else |
| { |
| /* #4 */ |
| i = vectorp->insert (i, newr); |
| } |
| |
| /* Check whether the ranges following the one we've just added or |
| touched can be folded in (#5 above). */ |
| if (i != vectorp->end () && i + 1 < vectorp->end ()) |
| { |
| int removed = 0; |
| auto next = i + 1; |
| |
| /* Get the range we just touched. */ |
| struct range &t = *i; |
| removed = 0; |
| |
| i = next; |
| for (; i < vectorp->end (); i++) |
| { |
| struct range &r = *i; |
| if (r.offset <= t.offset + t.length) |
| { |
| ULONGEST l, h; |
| |
| l = std::min (t.offset, r.offset); |
| h = std::max (t.offset + t.length, r.offset + r.length); |
| |
| t.offset = l; |
| t.length = h - l; |
| |
| removed++; |
| } |
| else |
| { |
| /* If we couldn't merge this one, we won't be able to |
| merge following ones either, since the ranges are |
| always sorted by OFFSET. */ |
| break; |
| } |
| } |
| |
| if (removed != 0) |
| vectorp->erase (next, next + removed); |
| } |
| } |
| |
| void |
| mark_value_bits_unavailable (struct value *value, |
| LONGEST offset, LONGEST length) |
| { |
| insert_into_bit_range_vector (&value->unavailable, offset, length); |
| } |
| |
| void |
| mark_value_bytes_unavailable (struct value *value, |
| LONGEST offset, LONGEST length) |
| { |
| mark_value_bits_unavailable (value, |
| offset * TARGET_CHAR_BIT, |
| length * TARGET_CHAR_BIT); |
| } |
| |
| /* Find the first range in RANGES that overlaps the range defined by |
| OFFSET and LENGTH, starting at element POS in the RANGES vector, |
| Returns the index into RANGES where such overlapping range was |
| found, or -1 if none was found. */ |
| |
| static int |
| find_first_range_overlap (const std::vector<range> *ranges, int pos, |
| LONGEST offset, LONGEST length) |
| { |
| int i; |
| |
| for (i = pos; i < ranges->size (); i++) |
| { |
| const range &r = (*ranges)[i]; |
| if (ranges_overlap (r.offset, r.length, offset, length)) |
| return i; |
| } |
| |
| return -1; |
| } |
| |
| /* Compare LENGTH_BITS of memory at PTR1 + OFFSET1_BITS with the memory at |
| PTR2 + OFFSET2_BITS. Return 0 if the memory is the same, otherwise |
| return non-zero. |
| |
| It must always be the case that: |
| OFFSET1_BITS % TARGET_CHAR_BIT == OFFSET2_BITS % TARGET_CHAR_BIT |
| |
| It is assumed that memory can be accessed from: |
| PTR + (OFFSET_BITS / TARGET_CHAR_BIT) |
| to: |
| PTR + ((OFFSET_BITS + LENGTH_BITS + TARGET_CHAR_BIT - 1) |
| / TARGET_CHAR_BIT) */ |
| static int |
| memcmp_with_bit_offsets (const gdb_byte *ptr1, size_t offset1_bits, |
| const gdb_byte *ptr2, size_t offset2_bits, |
| size_t length_bits) |
| { |
| gdb_assert (offset1_bits % TARGET_CHAR_BIT |
| == offset2_bits % TARGET_CHAR_BIT); |
| |
| if (offset1_bits % TARGET_CHAR_BIT != 0) |
| { |
| size_t bits; |
| gdb_byte mask, b1, b2; |
| |
| /* The offset from the base pointers PTR1 and PTR2 is not a complete |
| number of bytes. A number of bits up to either the next exact |
| byte boundary, or LENGTH_BITS (which ever is sooner) will be |
| compared. */ |
| bits = TARGET_CHAR_BIT - offset1_bits % TARGET_CHAR_BIT; |
| gdb_assert (bits < sizeof (mask) * TARGET_CHAR_BIT); |
| mask = (1 << bits) - 1; |
| |
| if (length_bits < bits) |
| { |
| mask &= ~(gdb_byte) ((1 << (bits - length_bits)) - 1); |
| bits = length_bits; |
| } |
| |
| /* Now load the two bytes and mask off the bits we care about. */ |
| b1 = *(ptr1 + offset1_bits / TARGET_CHAR_BIT) & mask; |
| b2 = *(ptr2 + offset2_bits / TARGET_CHAR_BIT) & mask; |
| |
| if (b1 != b2) |
| return 1; |
| |
| /* Now update the length and offsets to take account of the bits |
| we've just compared. */ |
| length_bits -= bits; |
| offset1_bits += bits; |
| offset2_bits += bits; |
| } |
| |
| if (length_bits % TARGET_CHAR_BIT != 0) |
| { |
| size_t bits; |
| size_t o1, o2; |
| gdb_byte mask, b1, b2; |
| |
| /* The length is not an exact number of bytes. After the previous |
| IF.. block then the offsets are byte aligned, or the |
| length is zero (in which case this code is not reached). Compare |
| a number of bits at the end of the region, starting from an exact |
| byte boundary. */ |
| bits = length_bits % TARGET_CHAR_BIT; |
| o1 = offset1_bits + length_bits - bits; |
| o2 = offset2_bits + length_bits - bits; |
| |
| gdb_assert (bits < sizeof (mask) * TARGET_CHAR_BIT); |
| mask = ((1 << bits) - 1) << (TARGET_CHAR_BIT - bits); |
| |
| gdb_assert (o1 % TARGET_CHAR_BIT == 0); |
| gdb_assert (o2 % TARGET_CHAR_BIT == 0); |
| |
| b1 = *(ptr1 + o1 / TARGET_CHAR_BIT) & mask; |
| b2 = *(ptr2 + o2 / TARGET_CHAR_BIT) & mask; |
| |
| if (b1 != b2) |
| return 1; |
| |
| length_bits -= bits; |
| } |
| |
| if (length_bits > 0) |
| { |
| /* We've now taken care of any stray "bits" at the start, or end of |
| the region to compare, the remainder can be covered with a simple |
| memcmp. */ |
| gdb_assert (offset1_bits % TARGET_CHAR_BIT == 0); |
| gdb_assert (offset2_bits % TARGET_CHAR_BIT == 0); |
| gdb_assert (length_bits % TARGET_CHAR_BIT == 0); |
| |
| return memcmp (ptr1 + offset1_bits / TARGET_CHAR_BIT, |
| ptr2 + offset2_bits / TARGET_CHAR_BIT, |
| length_bits / TARGET_CHAR_BIT); |
| } |
| |
| /* Length is zero, regions match. */ |
| return 0; |
| } |
| |
| /* Helper struct for find_first_range_overlap_and_match and |
| value_contents_bits_eq. Keep track of which slot of a given ranges |
| vector have we last looked at. */ |
| |
| struct ranges_and_idx |
| { |
| /* The ranges. */ |
| const std::vector<range> *ranges; |
| |
| /* The range we've last found in RANGES. Given ranges are sorted, |
| we can start the next lookup here. */ |
| int idx; |
| }; |
| |
| /* Helper function for value_contents_bits_eq. Compare LENGTH bits of |
| RP1's ranges starting at OFFSET1 bits with LENGTH bits of RP2's |
| ranges starting at OFFSET2 bits. Return true if the ranges match |
| and fill in *L and *H with the overlapping window relative to |
| (both) OFFSET1 or OFFSET2. */ |
| |
| static int |
| find_first_range_overlap_and_match (struct ranges_and_idx *rp1, |
| struct ranges_and_idx *rp2, |
| LONGEST offset1, LONGEST offset2, |
| LONGEST length, ULONGEST *l, ULONGEST *h) |
| { |
| rp1->idx = find_first_range_overlap (rp1->ranges, rp1->idx, |
| offset1, length); |
| rp2->idx = find_first_range_overlap (rp2->ranges, rp2->idx, |
| offset2, length); |
| |
| if (rp1->idx == -1 && rp2->idx == -1) |
| { |
| *l = length; |
| *h = length; |
| return 1; |
| } |
| else if (rp1->idx == -1 || rp2->idx == -1) |
| return 0; |
| else |
| { |
| const range *r1, *r2; |
| ULONGEST l1, h1; |
| ULONGEST l2, h2; |
| |
| r1 = &(*rp1->ranges)[rp1->idx]; |
| r2 = &(*rp2->ranges)[rp2->idx]; |
| |
| /* Get the unavailable windows intersected by the incoming |
| ranges. The first and last ranges that overlap the argument |
| range may be wider than said incoming arguments ranges. */ |
| l1 = std::max (offset1, r1->offset); |
| h1 = std::min (offset1 + length, r1->offset + r1->length); |
| |
| l2 = std::max (offset2, r2->offset); |
| h2 = std::min (offset2 + length, offset2 + r2->length); |
| |
| /* Make them relative to the respective start offsets, so we can |
| compare them for equality. */ |
| l1 -= offset1; |
| h1 -= offset1; |
| |
| l2 -= offset2; |
| h2 -= offset2; |
| |
| /* Different ranges, no match. */ |
| if (l1 != l2 || h1 != h2) |
| return 0; |
| |
| *h = h1; |
| *l = l1; |
| return 1; |
| } |
| } |
| |
| /* Helper function for value_contents_eq. The only difference is that |
| this function is bit rather than byte based. |
| |
| Compare LENGTH bits of VAL1's contents starting at OFFSET1 bits |
| with LENGTH bits of VAL2's contents starting at OFFSET2 bits. |
| Return true if the available bits match. */ |
| |
| static bool |
| value_contents_bits_eq (const struct value *val1, int offset1, |
| const struct value *val2, int offset2, |
| int length) |
| { |
| /* Each array element corresponds to a ranges source (unavailable, |
| optimized out). '1' is for VAL1, '2' for VAL2. */ |
| struct ranges_and_idx rp1[2], rp2[2]; |
| |
| /* See function description in value.h. */ |
| gdb_assert (!val1->lazy && !val2->lazy); |
| |
| /* We shouldn't be trying to compare past the end of the values. */ |
| gdb_assert (offset1 + length |
| <= TYPE_LENGTH (val1->enclosing_type) * TARGET_CHAR_BIT); |
| gdb_assert (offset2 + length |
| <= TYPE_LENGTH (val2->enclosing_type) * TARGET_CHAR_BIT); |
| |
| memset (&rp1, 0, sizeof (rp1)); |
| memset (&rp2, 0, sizeof (rp2)); |
| rp1[0].ranges = &val1->unavailable; |
| rp2[0].ranges = &val2->unavailable; |
| rp1[1].ranges = &val1->optimized_out; |
| rp2[1].ranges = &val2->optimized_out; |
| |
| while (length > 0) |
| { |
| ULONGEST l = 0, h = 0; /* init for gcc -Wall */ |
| int i; |
| |
| for (i = 0; i < 2; i++) |
| { |
| ULONGEST l_tmp, h_tmp; |
| |
| /* The contents only match equal if the invalid/unavailable |
| contents ranges match as well. */ |
| if (!find_first_range_overlap_and_match (&rp1[i], &rp2[i], |
| offset1, offset2, length, |
| &l_tmp, &h_tmp)) |
| return false; |
| |
| /* We're interested in the lowest/first range found. */ |
| if (i == 0 || l_tmp < l) |
| { |
| l = l_tmp; |
| h = h_tmp; |
| } |
| } |
| |
| /* Compare the available/valid contents. */ |
| if (memcmp_with_bit_offsets (val1->contents.get (), offset1, |
| val2->contents.get (), offset2, l) != 0) |
| return false; |
| |
| length -= h; |
| offset1 += h; |
| offset2 += h; |
| } |
| |
| return true; |
| } |
| |
| bool |
| value_contents_eq (const struct value *val1, LONGEST offset1, |
| const struct value *val2, LONGEST offset2, |
| LONGEST length) |
| { |
| return value_contents_bits_eq (val1, offset1 * TARGET_CHAR_BIT, |
| val2, offset2 * TARGET_CHAR_BIT, |
| length * TARGET_CHAR_BIT); |
| } |
| |
| |
| /* The value-history records all the values printed by print commands |
| during this session. */ |
| |
| static std::vector<value_ref_ptr> value_history; |
| |
| |
| /* List of all value objects currently allocated |
| (except for those released by calls to release_value) |
| This is so they can be freed after each command. */ |
| |
| static std::vector<value_ref_ptr> all_values; |
| |
| /* Allocate a lazy value for type TYPE. Its actual content is |
| "lazily" allocated too: the content field of the return value is |
| NULL; it will be allocated when it is fetched from the target. */ |
| |
| struct value * |
| allocate_value_lazy (struct type *type) |
| { |
| struct value *val; |
| |
| /* Call check_typedef on our type to make sure that, if TYPE |
| is a TYPE_CODE_TYPEDEF, its length is set to the length |
| of the target type instead of zero. However, we do not |
| replace the typedef type by the target type, because we want |
| to keep the typedef in order to be able to set the VAL's type |
| description correctly. */ |
| check_typedef (type); |
| |
| val = new struct value (type); |
| |
| /* Values start out on the all_values chain. */ |
| all_values.emplace_back (val); |
| |
| return val; |
| } |
| |
| /* The maximum size, in bytes, that GDB will try to allocate for a value. |
| The initial value of 64k was not selected for any specific reason, it is |
| just a reasonable starting point. */ |
| |
| static int max_value_size = 65536; /* 64k bytes */ |
| |
| /* It is critical that the MAX_VALUE_SIZE is at least as big as the size of |
| LONGEST, otherwise GDB will not be able to parse integer values from the |
| CLI; for example if the MAX_VALUE_SIZE could be set to 1 then GDB would |
| be unable to parse "set max-value-size 2". |
| |
| As we want a consistent GDB experience across hosts with different sizes |
| of LONGEST, this arbitrary minimum value was selected, so long as this |
| is bigger than LONGEST on all GDB supported hosts we're fine. */ |
| |
| #define MIN_VALUE_FOR_MAX_VALUE_SIZE 16 |
| gdb_static_assert (sizeof (LONGEST) <= MIN_VALUE_FOR_MAX_VALUE_SIZE); |
| |
| /* Implement the "set max-value-size" command. */ |
| |
| static void |
| set_max_value_size (const char *args, int from_tty, |
| struct cmd_list_element *c) |
| { |
| gdb_assert (max_value_size == -1 || max_value_size >= 0); |
| |
| if (max_value_size > -1 && max_value_size < MIN_VALUE_FOR_MAX_VALUE_SIZE) |
| { |
| max_value_size = MIN_VALUE_FOR_MAX_VALUE_SIZE; |
| error (_("max-value-size set too low, increasing to %d bytes"), |
| max_value_size); |
| } |
| } |
| |
| /* Implement the "show max-value-size" command. */ |
| |
| static void |
| show_max_value_size (struct ui_file *file, int from_tty, |
| struct cmd_list_element *c, const char *value) |
| { |
| if (max_value_size == -1) |
| fprintf_filtered (file, _("Maximum value size is unlimited.\n")); |
| else |
| fprintf_filtered (file, _("Maximum value size is %d bytes.\n"), |
| max_value_size); |
| } |
| |
| /* Called before we attempt to allocate or reallocate a buffer for the |
| contents of a value. TYPE is the type of the value for which we are |
| allocating the buffer. If the buffer is too large (based on the user |
| controllable setting) then throw an error. If this function returns |
| then we should attempt to allocate the buffer. */ |
| |
| static void |
| check_type_length_before_alloc (const struct type *type) |
| { |
| ULONGEST length = TYPE_LENGTH (type); |
| |
| if (max_value_size > -1 && length > max_value_size) |
| { |
| if (type->name () != NULL) |
| error (_("value of type `%s' requires %s bytes, which is more " |
| "than max-value-size"), type->name (), pulongest (length)); |
| else |
| error (_("value requires %s bytes, which is more than " |
| "max-value-size"), pulongest (length)); |
| } |
| } |
| |
| /* Allocate the contents of VAL if it has not been allocated yet. */ |
| |
| static void |
| allocate_value_contents (struct value *val) |
| { |
| if (!val->contents) |
| { |
| check_type_length_before_alloc (val->enclosing_type); |
| val->contents.reset |
| ((gdb_byte *) xzalloc (TYPE_LENGTH (val->enclosing_type))); |
| } |
| } |
| |
| /* Allocate a value and its contents for type TYPE. */ |
| |
| struct value * |
| allocate_value (struct type *type) |
| { |
| struct value *val = allocate_value_lazy (type); |
| |
| allocate_value_contents (val); |
| val->lazy = 0; |
| return val; |
| } |
| |
| /* Allocate a value that has the correct length |
| for COUNT repetitions of type TYPE. */ |
| |
| struct value * |
| allocate_repeat_value (struct type *type, int count) |
| { |
| /* Despite the fact that we are really creating an array of TYPE here, we |
| use the string lower bound as the array lower bound. This seems to |
| work fine for now. */ |
| int low_bound = current_language->string_lower_bound (); |
| /* FIXME-type-allocation: need a way to free this type when we are |
| done with it. */ |
| struct type *array_type |
| = lookup_array_range_type (type, low_bound, count + low_bound - 1); |
| |
| return allocate_value (array_type); |
| } |
| |
| struct value * |
| allocate_computed_value (struct type *type, |
| const struct lval_funcs *funcs, |
| void *closure) |
| { |
| struct value *v = allocate_value_lazy (type); |
| |
| VALUE_LVAL (v) = lval_computed; |
| v->location.computed.funcs = funcs; |
| v->location.computed.closure = closure; |
| |
| return v; |
| } |
| |
| /* Allocate NOT_LVAL value for type TYPE being OPTIMIZED_OUT. */ |
| |
| struct value * |
| allocate_optimized_out_value (struct type *type) |
| { |
| struct value *retval = allocate_value_lazy (type); |
| |
| mark_value_bytes_optimized_out (retval, 0, TYPE_LENGTH (type)); |
| set_value_lazy (retval, 0); |
| return retval; |
| } |
| |
| /* Accessor methods. */ |
| |
| struct type * |
| value_type (const struct value *value) |
| { |
| return value->type; |
| } |
| void |
| deprecated_set_value_type (struct value *value, struct type *type) |
| { |
| value->type = type; |
| } |
| |
| LONGEST |
| value_offset (const struct value *value) |
| { |
| return value->offset; |
| } |
| void |
| set_value_offset (struct value *value, LONGEST offset) |
| { |
| value->offset = offset; |
| } |
| |
| LONGEST |
| value_bitpos (const struct value *value) |
| { |
| return value->bitpos; |
| } |
| void |
| set_value_bitpos (struct value *value, LONGEST bit) |
| { |
| value->bitpos = bit; |
| } |
| |
| LONGEST |
| value_bitsize (const struct value *value) |
| { |
| return value->bitsize; |
| } |
| void |
| set_value_bitsize (struct value *value, LONGEST bit) |
| { |
| value->bitsize = bit; |
| } |
| |
| struct value * |
| value_parent (const struct value *value) |
| { |
| return value->parent.get (); |
| } |
| |
| /* See value.h. */ |
| |
| void |
| set_value_parent (struct value *value, struct value *parent) |
| { |
| value->parent = value_ref_ptr::new_reference (parent); |
| } |
| |
| gdb_byte * |
| value_contents_raw (struct value *value) |
| { |
| struct gdbarch *arch = get_value_arch (value); |
| int unit_size = gdbarch_addressable_memory_unit_size (arch); |
| |
| allocate_value_contents (value); |
| return value->contents.get () + value->embedded_offset * unit_size; |
| } |
| |
| gdb_byte * |
| value_contents_all_raw (struct value *value) |
| { |
| allocate_value_contents (value); |
| return value->contents.get (); |
| } |
| |
| struct type * |
| value_enclosing_type (const struct value *value) |
| { |
| return value->enclosing_type; |
| } |
| |
| /* Look at value.h for description. */ |
| |
| struct type * |
| value_actual_type (struct value *value, int resolve_simple_types, |
| int *real_type_found) |
| { |
| struct value_print_options opts; |
| struct type *result; |
| |
| get_user_print_options (&opts); |
| |
| if (real_type_found) |
| *real_type_found = 0; |
| result = value_type (value); |
| if (opts.objectprint) |
| { |
| /* If result's target type is TYPE_CODE_STRUCT, proceed to |
| fetch its rtti type. */ |
| if (result->is_pointer_or_reference () |
| && (check_typedef (TYPE_TARGET_TYPE (result))->code () |
| == TYPE_CODE_STRUCT) |
| && !value_optimized_out (value)) |
| { |
| struct type *real_type; |
| |
| real_type = value_rtti_indirect_type (value, NULL, NULL, NULL); |
| if (real_type) |
| { |
| if (real_type_found) |
| *real_type_found = 1; |
| result = real_type; |
| } |
| } |
| else if (resolve_simple_types) |
| { |
| if (real_type_found) |
| *real_type_found = 1; |
| result = value_enclosing_type (value); |
| } |
| } |
| |
| return result; |
| } |
| |
| void |
| error_value_optimized_out (void) |
| { |
| error (_("value has been optimized out")); |
| } |
| |
| static void |
| require_not_optimized_out (const struct value *value) |
| { |
| if (!value->optimized_out.empty ()) |
| { |
| if (value->lval == lval_register) |
| error (_("register has not been saved in frame")); |
| else |
| error_value_optimized_out (); |
| } |
| } |
| |
| static void |
| require_available (const struct value *value) |
| { |
| if (!value->unavailable.empty ()) |
| throw_error (NOT_AVAILABLE_ERROR, _("value is not available")); |
| } |
| |
| const gdb_byte * |
| value_contents_for_printing (struct value *value) |
| { |
| if (value->lazy) |
| value_fetch_lazy (value); |
| return value->contents.get (); |
| } |
| |
| const gdb_byte * |
| value_contents_for_printing_const (const struct value *value) |
| { |
| gdb_assert (!value->lazy); |
| return value->contents.get (); |
| } |
| |
| const gdb_byte * |
| value_contents_all (struct value *value) |
| { |
| const gdb_byte *result = value_contents_for_printing (value); |
| require_not_optimized_out (value); |
| require_available (value); |
| return result; |
| } |
| |
| /* Copy ranges in SRC_RANGE that overlap [SRC_BIT_OFFSET, |
| SRC_BIT_OFFSET+BIT_LENGTH) ranges into *DST_RANGE, adjusted. */ |
| |
| static void |
| ranges_copy_adjusted (std::vector<range> *dst_range, int dst_bit_offset, |
| const std::vector<range> &src_range, int src_bit_offset, |
| int bit_length) |
| { |
| for (const range &r : src_range) |
| { |
| ULONGEST h, l; |
| |
| l = std::max (r.offset, (LONGEST) src_bit_offset); |
| h = std::min (r.offset + r.length, |
| (LONGEST) src_bit_offset + bit_length); |
| |
| if (l < h) |
| insert_into_bit_range_vector (dst_range, |
| dst_bit_offset + (l - src_bit_offset), |
| h - l); |
| } |
| } |
| |
| /* Copy the ranges metadata in SRC that overlaps [SRC_BIT_OFFSET, |
| SRC_BIT_OFFSET+BIT_LENGTH) into DST, adjusted. */ |
| |
| static void |
| value_ranges_copy_adjusted (struct value *dst, int dst_bit_offset, |
| const struct value *src, int src_bit_offset, |
| int bit_length) |
| { |
| ranges_copy_adjusted (&dst->unavailable, dst_bit_offset, |
| src->unavailable, src_bit_offset, |
| bit_length); |
| ranges_copy_adjusted (&dst->optimized_out, dst_bit_offset, |
| src->optimized_out, src_bit_offset, |
| bit_length); |
| } |
| |
| /* Copy LENGTH target addressable memory units of SRC value's (all) contents |
| (value_contents_all) starting at SRC_OFFSET, into DST value's (all) |
| contents, starting at DST_OFFSET. If unavailable contents are |
| being copied from SRC, the corresponding DST contents are marked |
| unavailable accordingly. Neither DST nor SRC may be lazy |
| values. |
| |
| It is assumed the contents of DST in the [DST_OFFSET, |
| DST_OFFSET+LENGTH) range are wholly available. */ |
| |
| static void |
| value_contents_copy_raw (struct value *dst, LONGEST dst_offset, |
| struct value *src, LONGEST src_offset, LONGEST length) |
| { |
| LONGEST src_bit_offset, dst_bit_offset, bit_length; |
| struct gdbarch *arch = get_value_arch (src); |
| int unit_size = gdbarch_addressable_memory_unit_size (arch); |
| |
| /* A lazy DST would make that this copy operation useless, since as |
| soon as DST's contents were un-lazied (by a later value_contents |
| call, say), the contents would be overwritten. A lazy SRC would |
| mean we'd be copying garbage. */ |
| gdb_assert (!dst->lazy && !src->lazy); |
| |
| /* The overwritten DST range gets unavailability ORed in, not |
| replaced. Make sure to remember to implement replacing if it |
| turns out actually necessary. */ |
| gdb_assert (value_bytes_available (dst, dst_offset, length)); |
| gdb_assert (!value_bits_any_optimized_out (dst, |
| TARGET_CHAR_BIT * dst_offset, |
| TARGET_CHAR_BIT * length)); |
| |
| /* Copy the data. */ |
| memcpy (value_contents_all_raw (dst) + dst_offset * unit_size, |
| value_contents_all_raw (src) + src_offset * unit_size, |
| length * unit_size); |
| |
| /* Copy the meta-data, adjusted. */ |
| src_bit_offset = src_offset * unit_size * HOST_CHAR_BIT; |
| dst_bit_offset = dst_offset * unit_size * HOST_CHAR_BIT; |
| bit_length = length * unit_size * HOST_CHAR_BIT; |
| |
| value_ranges_copy_adjusted (dst, dst_bit_offset, |
| src, src_bit_offset, |
| bit_length); |
| } |
| |
| /* Copy LENGTH bytes of SRC value's (all) contents |
| (value_contents_all) starting at SRC_OFFSET byte, into DST value's |
| (all) contents, starting at DST_OFFSET. If unavailable contents |
| are being copied from SRC, the corresponding DST contents are |
| marked unavailable accordingly. DST must not be lazy. If SRC is |
| lazy, it will be fetched now. |
| |
| It is assumed the contents of DST in the [DST_OFFSET, |
| DST_OFFSET+LENGTH) range are wholly available. */ |
| |
| void |
| value_contents_copy (struct value *dst, LONGEST dst_offset, |
| struct value *src, LONGEST src_offset, LONGEST length) |
| { |
| if (src->lazy) |
| value_fetch_lazy (src); |
| |
| value_contents_copy_raw (dst, dst_offset, src, src_offset, length); |
| } |
| |
| int |
| value_lazy (const struct value *value) |
| { |
| return value->lazy; |
| } |
| |
| void |
| set_value_lazy (struct value *value, int val) |
| { |
| value->lazy = val; |
| } |
| |
| int |
| value_stack (const struct value *value) |
| { |
| return value->stack; |
| } |
| |
| void |
| set_value_stack (struct value *value, int val) |
| { |
| value->stack = val; |
| } |
| |
| const gdb_byte * |
| value_contents (struct value *value) |
| { |
| const gdb_byte *result = value_contents_writeable (value); |
| require_not_optimized_out (value); |
| require_available (value); |
| return result; |
| } |
| |
| gdb_byte * |
| value_contents_writeable (struct value *value) |
| { |
| if (value->lazy) |
| value_fetch_lazy (value); |
| return value_contents_raw (value); |
| } |
| |
| int |
| value_optimized_out (struct value *value) |
| { |
| if (value->lazy) |
| { |
| /* See if we can compute the result without fetching the |
| value. */ |
| if (VALUE_LVAL (value) == lval_memory) |
| return false; |
| else if (VALUE_LVAL (value) == lval_computed) |
| { |
| const struct lval_funcs *funcs = value->location.computed.funcs; |
| |
| if (funcs->is_optimized_out != nullptr) |
| return funcs->is_optimized_out (value); |
| } |
| |
| /* Fall back to fetching. */ |
| try |
| { |
| value_fetch_lazy (value); |
| } |
| catch (const gdb_exception_error &ex) |
| { |
| switch (ex.error) |
| { |
| case MEMORY_ERROR: |
| case OPTIMIZED_OUT_ERROR: |
| case NOT_AVAILABLE_ERROR: |
| /* These can normally happen when we try to access an |
| optimized out or unavailable register, either in a |
| physical register or spilled to memory. */ |
| break; |
| default: |
| throw; |
| } |
| } |
| } |
| |
| return !value->optimized_out.empty (); |
| } |
| |
| /* Mark contents of VALUE as optimized out, starting at OFFSET bytes, and |
| the following LENGTH bytes. */ |
| |
| void |
| mark_value_bytes_optimized_out (struct value *value, int offset, int length) |
| { |
| mark_value_bits_optimized_out (value, |
| offset * TARGET_CHAR_BIT, |
| length * TARGET_CHAR_BIT); |
| } |
| |
| /* See value.h. */ |
| |
| void |
| mark_value_bits_optimized_out (struct value *value, |
| LONGEST offset, LONGEST length) |
| { |
| insert_into_bit_range_vector (&value->optimized_out, offset, length); |
| } |
| |
| int |
| value_bits_synthetic_pointer (const struct value *value, |
| LONGEST offset, LONGEST length) |
| { |
| if (value->lval != lval_computed |
| || !value->location.computed.funcs->check_synthetic_pointer) |
| return 0; |
| return value->location.computed.funcs->check_synthetic_pointer (value, |
| offset, |
| length); |
| } |
| |
| LONGEST |
| value_embedded_offset (const struct value *value) |
| { |
| return value->embedded_offset; |
| } |
| |
| void |
| set_value_embedded_offset (struct value *value, LONGEST val) |
| { |
| value->embedded_offset = val; |
| } |
| |
| LONGEST |
| value_pointed_to_offset (const struct value *value) |
| { |
| return value->pointed_to_offset; |
| } |
| |
| void |
| set_value_pointed_to_offset (struct value *value, LONGEST val) |
| { |
| value->pointed_to_offset = val; |
| } |
| |
| const struct lval_funcs * |
| value_computed_funcs (const struct value *v) |
| { |
| gdb_assert (value_lval_const (v) == lval_computed); |
| |
| return v->location.computed.funcs; |
| } |
| |
| void * |
| value_computed_closure (const struct value *v) |
| { |
| gdb_assert (v->lval == lval_computed); |
| |
| return v->location.computed.closure; |
| } |
| |
| enum lval_type * |
| deprecated_value_lval_hack (struct value *value) |
| { |
| return &value->lval; |
| } |
| |
| enum lval_type |
| value_lval_const (const struct value *value) |
| { |
| return value->lval; |
| } |
| |
| CORE_ADDR |
| value_address (const struct value *value) |
| { |
| if (value->lval != lval_memory) |
| return 0; |
| if (value->parent != NULL) |
| return value_address (value->parent.get ()) + value->offset; |
| if (NULL != TYPE_DATA_LOCATION (value_type (value))) |
| { |
| gdb_assert (PROP_CONST == TYPE_DATA_LOCATION_KIND (value_type (value))); |
| return TYPE_DATA_LOCATION_ADDR (value_type (value)); |
| } |
| |
| return value->location.address + value->offset; |
| } |
| |
| CORE_ADDR |
| value_raw_address (const struct value *value) |
| { |
| if (value->lval != lval_memory) |
| return 0; |
| return value->location.address; |
| } |
| |
| void |
| set_value_address (struct value *value, CORE_ADDR addr) |
| { |
| gdb_assert (value->lval == lval_memory); |
| value->location.address = addr; |
| } |
| |
| struct internalvar ** |
| deprecated_value_internalvar_hack (struct value *value) |
| { |
| return &value->location.internalvar; |
| } |
| |
| struct frame_id * |
| deprecated_value_next_frame_id_hack (struct value *value) |
| { |
| gdb_assert (value->lval == lval_register); |
| return &value->location.reg.next_frame_id; |
| } |
| |
| int * |
| deprecated_value_regnum_hack (struct value *value) |
| { |
| gdb_assert (value->lval == lval_register); |
| return &value->location.reg.regnum; |
| } |
| |
| int |
| deprecated_value_modifiable (const struct value *value) |
| { |
| return value->modifiable; |
| } |
| |
| /* Return a mark in the value chain. All values allocated after the |
| mark is obtained (except for those released) are subject to being freed |
| if a subsequent value_free_to_mark is passed the mark. */ |
| struct value * |
| value_mark (void) |
| { |
| if (all_values.empty ()) |
| return nullptr; |
| return all_values.back ().get (); |
| } |
| |
| /* See value.h. */ |
| |
| void |
| value_incref (struct value *val) |
| { |
| val->reference_count++; |
| } |
| |
| /* Release a reference to VAL, which was acquired with value_incref. |
| This function is also called to deallocate values from the value |
| chain. */ |
| |
| void |
| value_decref (struct value *val) |
| { |
| if (val != nullptr) |
| { |
| gdb_assert (val->reference_count > 0); |
| val->reference_count--; |
| if (val->reference_count == 0) |
| delete val; |
| } |
| } |
| |
| /* Free all values allocated since MARK was obtained by value_mark |
| (except for those released). */ |
| void |
| value_free_to_mark (const struct value *mark) |
| { |
| auto iter = std::find (all_values.begin (), all_values.end (), mark); |
| if (iter == all_values.end ()) |
| all_values.clear (); |
| else |
| all_values.erase (iter + 1, all_values.end ()); |
| } |
| |
| /* Remove VAL from the chain all_values |
| so it will not be freed automatically. */ |
| |
| value_ref_ptr |
| release_value (struct value *val) |
| { |
| if (val == nullptr) |
| return value_ref_ptr (); |
| |
| std::vector<value_ref_ptr>::reverse_iterator iter; |
| for (iter = all_values.rbegin (); iter != all_values.rend (); ++iter) |
| { |
| if (*iter == val) |
| { |
| value_ref_ptr result = *iter; |
| all_values.erase (iter.base () - 1); |
| return result; |
| } |
| } |
| |
| /* We must always return an owned reference. Normally this happens |
| because we transfer the reference from the value chain, but in |
| this case the value was not on the chain. */ |
| return value_ref_ptr::new_reference (val); |
| } |
| |
| /* See value.h. */ |
| |
| std::vector<value_ref_ptr> |
| value_release_to_mark (const struct value *mark) |
| { |
| std::vector<value_ref_ptr> result; |
| |
| auto iter = std::find (all_values.begin (), all_values.end (), mark); |
| if (iter == all_values.end ()) |
| std::swap (result, all_values); |
| else |
| { |
| std::move (iter + 1, all_values.end (), std::back_inserter (result)); |
| all_values.erase (iter + 1, all_values.end ()); |
| } |
| std::reverse (result.begin (), result.end ()); |
| return result; |
| } |
| |
| /* Return a copy of the value ARG. |
| It contains the same contents, for same memory address, |
| but it's a different block of storage. */ |
| |
| struct value * |
| value_copy (struct value *arg) |
| { |
| struct type *encl_type = value_enclosing_type (arg); |
| struct value *val; |
| |
| if (value_lazy (arg)) |
| val = allocate_value_lazy (encl_type); |
| else |
| val = allocate_value (encl_type); |
| val->type = arg->type; |
| VALUE_LVAL (val) = VALUE_LVAL (arg); |
| val->location = arg->location; |
| val->offset = arg->offset; |
| val->bitpos = arg->bitpos; |
| val->bitsize = arg->bitsize; |
| val->lazy = arg->lazy; |
| val->embedded_offset = value_embedded_offset (arg); |
| val->pointed_to_offset = arg->pointed_to_offset; |
| val->modifiable = arg->modifiable; |
| val->stack = arg->stack; |
| val->is_zero = arg->is_zero; |
| val->initialized = arg->initialized; |
| if (!value_lazy (val)) |
| { |
| memcpy (value_contents_all_raw (val), value_contents_all_raw (arg), |
| TYPE_LENGTH (value_enclosing_type (arg))); |
| |
| } |
| val->unavailable = arg->unavailable; |
| val->optimized_out = arg->optimized_out; |
| val->parent = arg->parent; |
| if (VALUE_LVAL (val) == lval_computed) |
| { |
| const struct lval_funcs *funcs = val->location.computed.funcs; |
| |
| if (funcs->copy_closure) |
| val->location.computed.closure = funcs->copy_closure (val); |
| } |
| return val; |
| } |
| |
| /* Return a "const" and/or "volatile" qualified version of the value V. |
| If CNST is true, then the returned value will be qualified with |
| "const". |
| if VOLTL is true, then the returned value will be qualified with |
| "volatile". */ |
| |
| struct value * |
| make_cv_value (int cnst, int voltl, struct value *v) |
| { |
| struct type *val_type = value_type (v); |
| struct type *enclosing_type = value_enclosing_type (v); |
| struct value *cv_val = value_copy (v); |
| |
| deprecated_set_value_type (cv_val, |
| make_cv_type (cnst, voltl, val_type, NULL)); |
| set_value_enclosing_type (cv_val, |
| make_cv_type (cnst, voltl, enclosing_type, NULL)); |
| |
| return cv_val; |
| } |
| |
| /* Return a version of ARG that is non-lvalue. */ |
| |
| struct value * |
| value_non_lval (struct value *arg) |
| { |
| if (VALUE_LVAL (arg) != not_lval) |
| { |
| struct type *enc_type = value_enclosing_type (arg); |
| struct value *val = allocate_value (enc_type); |
| |
| memcpy (value_contents_all_raw (val), value_contents_all (arg), |
| TYPE_LENGTH (enc_type)); |
| val->type = arg->type; |
| set_value_embedded_offset (val, value_embedded_offset (arg)); |
| set_value_pointed_to_offset (val, value_pointed_to_offset (arg)); |
| return val; |
| } |
| return arg; |
| } |
| |
| /* Write contents of V at ADDR and set its lval type to be LVAL_MEMORY. */ |
| |
| void |
| value_force_lval (struct value *v, CORE_ADDR addr) |
| { |
| gdb_assert (VALUE_LVAL (v) == not_lval); |
| |
| write_memory (addr, value_contents_raw (v), TYPE_LENGTH (value_type (v))); |
| v->lval = lval_memory; |
| v->location.address = addr; |
| } |
| |
| void |
| set_value_component_location (struct value *component, |
| const struct value *whole) |
| { |
| struct type *type; |
| |
| gdb_assert (whole->lval != lval_xcallable); |
| |
| if (whole->lval == lval_internalvar) |
| VALUE_LVAL (component) = lval_internalvar_component; |
| else |
| VALUE_LVAL (component) = whole->lval; |
| |
| component->location = whole->location; |
| if (whole->lval == lval_computed) |
| { |
| const struct lval_funcs *funcs = whole->location.computed.funcs; |
| |
| if (funcs->copy_closure) |
| component->location.computed.closure = funcs->copy_closure (whole); |
| } |
| |
| /* If the WHOLE value has a dynamically resolved location property then |
| update the address of the COMPONENT. */ |
| type = value_type (whole); |
| if (NULL != TYPE_DATA_LOCATION (type) |
| && TYPE_DATA_LOCATION_KIND (type) == PROP_CONST) |
| set_value_address (component, TYPE_DATA_LOCATION_ADDR (type)); |
| |
| /* Similarly, if the COMPONENT value has a dynamically resolved location |
| property then update its address. */ |
| type = value_type (component); |
| if (NULL != TYPE_DATA_LOCATION (type) |
| && TYPE_DATA_LOCATION_KIND (type) == PROP_CONST) |
| { |
| /* If the COMPONENT has a dynamic location, and is an |
| lval_internalvar_component, then we change it to a lval_memory. |
| |
| Usually a component of an internalvar is created non-lazy, and has |
| its content immediately copied from the parent internalvar. |
| However, for components with a dynamic location, the content of |
| the component is not contained within the parent, but is instead |
| accessed indirectly. Further, the component will be created as a |
| lazy value. |
| |
| By changing the type of the component to lval_memory we ensure |
| that value_fetch_lazy can successfully load the component. |
| |
| This solution isn't ideal, but a real fix would require values to |
| carry around both the parent value contents, and the contents of |
| any dynamic fields within the parent. This is a substantial |
| change to how values work in GDB. */ |
| if (VALUE_LVAL (component) == lval_internalvar_component) |
| { |
| gdb_assert (value_lazy (component)); |
| VALUE_LVAL (component) = lval_memory; |
| } |
| else |
| gdb_assert (VALUE_LVAL (component) == lval_memory); |
| set_value_address (component, TYPE_DATA_LOCATION_ADDR (type)); |
| } |
| } |
| |
| /* Access to the value history. */ |
| |
| /* Record a new value in the value history. |
| Returns the absolute history index of the entry. */ |
| |
| int |
| record_latest_value (struct value *val) |
| { |
| /* We don't want this value to have anything to do with the inferior anymore. |
| In particular, "set $1 = 50" should not affect the variable from which |
| the value was taken, and fast watchpoints should be able to assume that |
| a value on the value history never changes. */ |
| if (value_lazy (val)) |
| value_fetch_lazy (val); |
| /* We preserve VALUE_LVAL so that the user can find out where it was fetched |
| from. This is a bit dubious, because then *&$1 does not just return $1 |
| but the current contents of that location. c'est la vie... */ |
| val->modifiable = 0; |
| |
| value_history.push_back (release_value (val)); |
| |
| return value_history.size (); |
| } |
| |
| /* Return a copy of the value in the history with sequence number NUM. */ |
| |
| struct value * |
| access_value_history (int num) |
| { |
| int absnum = num; |
| |
| if (absnum <= 0) |
| absnum += value_history.size (); |
| |
| if (absnum <= 0) |
| { |
| if (num == 0) |
| error (_("The history is empty.")); |
| else if (num == 1) |
| error (_("There is only one value in the history.")); |
| else |
| error (_("History does not go back to $$%d."), -num); |
| } |
| if (absnum > value_history.size ()) |
| error (_("History has not yet reached $%d."), absnum); |
| |
| absnum--; |
| |
| return value_copy (value_history[absnum].get ()); |
| } |
| |
| static void |
| show_values (const char *num_exp, int from_tty) |
| { |
| int i; |
| struct value *val; |
| static int num = 1; |
| |
| if (num_exp) |
| { |
| /* "show values +" should print from the stored position. |
| "show values <exp>" should print around value number <exp>. */ |
| if (num_exp[0] != '+' || num_exp[1] != '\0') |
| num = parse_and_eval_long (num_exp) - 5; |
| } |
| else |
| { |
| /* "show values" means print the last 10 values. */ |
| num = value_history.size () - 9; |
| } |
| |
| if (num <= 0) |
| num = 1; |
| |
| for (i = num; i < num + 10 && i <= value_history.size (); i++) |
| { |
| struct value_print_options opts; |
| |
| val = access_value_history (i); |
| printf_filtered (("$%d = "), i); |
| get_user_print_options (&opts); |
| value_print (val, gdb_stdout, &opts); |
| printf_filtered (("\n")); |
| } |
| |
| /* The next "show values +" should start after what we just printed. */ |
| num += 10; |
| |
| /* Hitting just return after this command should do the same thing as |
| "show values +". If num_exp is null, this is unnecessary, since |
| "show values +" is not useful after "show values". */ |
| if (from_tty && num_exp) |
| set_repeat_arguments ("+"); |
| } |
| |
| enum internalvar_kind |
| { |
| /* The internal variable is empty. */ |
| INTERNALVAR_VOID, |
| |
| /* The value of the internal variable is provided directly as |
| a GDB value object. */ |
| INTERNALVAR_VALUE, |
| |
| /* A fresh value is computed via a call-back routine on every |
| access to the internal variable. */ |
| INTERNALVAR_MAKE_VALUE, |
| |
| /* The internal variable holds a GDB internal convenience function. */ |
| INTERNALVAR_FUNCTION, |
| |
| /* The variable holds an integer value. */ |
| INTERNALVAR_INTEGER, |
| |
| /* The variable holds a GDB-provided string. */ |
| INTERNALVAR_STRING, |
| }; |
| |
| union internalvar_data |
| { |
| /* A value object used with INTERNALVAR_VALUE. */ |
| struct value *value; |
| |
| /* The call-back routine used with INTERNALVAR_MAKE_VALUE. */ |
| struct |
| { |
| /* The functions to call. */ |
| const struct internalvar_funcs *functions; |
| |
| /* The function's user-data. */ |
| void *data; |
| } make_value; |
| |
| /* The internal function used with INTERNALVAR_FUNCTION. */ |
| struct |
| { |
| struct internal_function *function; |
| /* True if this is the canonical name for the function. */ |
| int canonical; |
| } fn; |
| |
| /* An integer value used with INTERNALVAR_INTEGER. */ |
| struct |
| { |
| /* If type is non-NULL, it will be used as the type to generate |
| a value for this internal variable. If type is NULL, a default |
| integer type for the architecture is used. */ |
| struct type *type; |
| LONGEST val; |
| } integer; |
| |
| /* A string value used with INTERNALVAR_STRING. */ |
| char *string; |
| }; |
| |
| /* Internal variables. These are variables within the debugger |
| that hold values assigned by debugger commands. |
| The user refers to them with a '$' prefix |
| that does not appear in the variable names stored internally. */ |
| |
| struct internalvar |
| { |
| struct internalvar *next; |
| char *name; |
| |
| /* We support various different kinds of content of an internal variable. |
| enum internalvar_kind specifies the kind, and union internalvar_data |
| provides the data associated with this particular kind. */ |
| |
| enum internalvar_kind kind; |
| |
| union internalvar_data u; |
| }; |
| |
| static struct internalvar *internalvars; |
| |
| /* If the variable does not already exist create it and give it the |
| value given. If no value is given then the default is zero. */ |
| static void |
| init_if_undefined_command (const char* args, int from_tty) |
| { |
| struct internalvar *intvar = nullptr; |
| |
| /* Parse the expression - this is taken from set_command(). */ |
| expression_up expr = parse_expression (args); |
| |
| /* Validate the expression. |
| Was the expression an assignment? |
| Or even an expression at all? */ |
| if (expr->first_opcode () != BINOP_ASSIGN) |
| error (_("Init-if-undefined requires an assignment expression.")); |
| |
| /* Extract the variable from the parsed expression. */ |
| expr::assign_operation *assign |
| = dynamic_cast<expr::assign_operation *> (expr->op.get ()); |
| if (assign != nullptr) |
| { |
| expr::operation *lhs = assign->get_lhs (); |
| expr::internalvar_operation *ivarop |
| = dynamic_cast<expr::internalvar_operation *> (lhs); |
| if (ivarop != nullptr) |
| intvar = ivarop->get_internalvar (); |
| } |
| |
| if (intvar == nullptr) |
| error (_("The first parameter to init-if-undefined " |
| "should be a GDB variable.")); |
| |
| /* Only evaluate the expression if the lvalue is void. |
| This may still fail if the expression is invalid. */ |
| if (intvar->kind == INTERNALVAR_VOID) |
| evaluate_expression (expr.get ()); |
| } |
| |
| |
| /* Look up an internal variable with name NAME. NAME should not |
| normally include a dollar sign. |
| |
| If the specified internal variable does not exist, |
| the return value is NULL. */ |
| |
| struct internalvar * |
| lookup_only_internalvar (const char *name) |
| { |
| struct internalvar *var; |
| |
| for (var = internalvars; var; var = var->next) |
| if (strcmp (var->name, name) == 0) |
| return var; |
| |
| return NULL; |
| } |
| |
| /* Complete NAME by comparing it to the names of internal |
| variables. */ |
| |
| void |
| complete_internalvar (completion_tracker &tracker, const char *name) |
| { |
| struct internalvar *var; |
| int len; |
| |
| len = strlen (name); |
| |
| for (var = internalvars; var; var = var->next) |
| if (strncmp (var->name, name, len) == 0) |
| tracker.add_completion (make_unique_xstrdup (var->name)); |
| } |
| |
| /* Create an internal variable with name NAME and with a void value. |
| NAME should not normally include a dollar sign. */ |
| |
| struct internalvar * |
| create_internalvar (const char *name) |
| { |
| struct internalvar *var = XNEW (struct internalvar); |
| |
| var->name = xstrdup (name); |
| var->kind = INTERNALVAR_VOID; |
| var->next = internalvars; |
| internalvars = var; |
| return var; |
| } |
| |
| /* Create an internal variable with name NAME and register FUN as the |
| function that value_of_internalvar uses to create a value whenever |
| this variable is referenced. NAME should not normally include a |
| dollar sign. DATA is passed uninterpreted to FUN when it is |
| called. CLEANUP, if not NULL, is called when the internal variable |
| is destroyed. It is passed DATA as its only argument. */ |
| |
| struct internalvar * |
| create_internalvar_type_lazy (const char *name, |
| const struct internalvar_funcs *funcs, |
| void *data) |
| { |
| struct internalvar *var = create_internalvar (name); |
| |
| var->kind = INTERNALVAR_MAKE_VALUE; |
| var->u.make_value.functions = funcs; |
| var->u.make_value.data = data; |
| return var; |
| } |
| |
| /* See documentation in value.h. */ |
| |
| int |
| compile_internalvar_to_ax (struct internalvar *var, |
| struct agent_expr *expr, |
| struct axs_value *value) |
| { |
| if (var->kind != INTERNALVAR_MAKE_VALUE |
| || var->u.make_value.functions->compile_to_ax == NULL) |
| return 0; |
| |
| var->u.make_value.functions->compile_to_ax (var, expr, value, |
| var->u.make_value.data); |
| return 1; |
| } |
| |
| /* Look up an internal variable with name NAME. NAME should not |
| normally include a dollar sign. |
| |
| If the specified internal variable does not exist, |
| one is created, with a void value. */ |
| |
| struct internalvar * |
| lookup_internalvar (const char *name) |
| { |
| struct internalvar *var; |
| |
| var = lookup_only_internalvar (name); |
| if (var) |
| return var; |
| |
| return create_internalvar (name); |
| } |
| |
| /* Return current value of internal variable VAR. For variables that |
| are not inherently typed, use a value type appropriate for GDBARCH. */ |
| |
| struct value * |
| value_of_internalvar (struct gdbarch *gdbarch, struct internalvar *var) |
| { |
| struct value *val; |
| struct trace_state_variable *tsv; |
| |
| /* If there is a trace state variable of the same name, assume that |
| is what we really want to see. */ |
| tsv = find_trace_state_variable (var->name); |
| if (tsv) |
| { |
| tsv->value_known = target_get_trace_state_variable_value (tsv->number, |
| &(tsv->value)); |
| if (tsv->value_known) |
| val = value_from_longest (builtin_type (gdbarch)->builtin_int64, |
| tsv->value); |
| else |
| val = allocate_value (builtin_type (gdbarch)->builtin_void); |
| return val; |
| } |
| |
| switch (var->kind) |
| { |
| case INTERNALVAR_VOID: |
| val = allocate_value (builtin_type (gdbarch)->builtin_void); |
| break; |
| |
| case INTERNALVAR_FUNCTION: |
| val = allocate_value (builtin_type (gdbarch)->internal_fn); |
| break; |
| |
| case INTERNALVAR_INTEGER: |
| if (!var->u.integer.type) |
| val = value_from_longest (builtin_type (gdbarch)->builtin_int, |
| var->u.integer.val); |
| else |
| val = value_from_longest (var->u.integer.type, var->u.integer.val); |
| break; |
| |
| case INTERNALVAR_STRING: |
| val = value_cstring (var->u.string, strlen (var->u.string), |
| builtin_type (gdbarch)->builtin_char); |
| break; |
| |
| case INTERNALVAR_VALUE: |
| val = value_copy (var->u.value); |
| if (value_lazy (val)) |
| value_fetch_lazy (val); |
| break; |
| |
| case INTERNALVAR_MAKE_VALUE: |
| val = (*var->u.make_value.functions->make_value) (gdbarch, var, |
| var->u.make_value.data); |
| break; |
| |
| default: |
| internal_error (__FILE__, __LINE__, _("bad kind")); |
| } |
| |
| /* Change the VALUE_LVAL to lval_internalvar so that future operations |
| on this value go back to affect the original internal variable. |
| |
| Do not do this for INTERNALVAR_MAKE_VALUE variables, as those have |
| no underlying modifiable state in the internal variable. |
| |
| Likewise, if the variable's value is a computed lvalue, we want |
| references to it to produce another computed lvalue, where |
| references and assignments actually operate through the |
| computed value's functions. |
| |
| This means that internal variables with computed values |
| behave a little differently from other internal variables: |
| assignments to them don't just replace the previous value |
| altogether. At the moment, this seems like the behavior we |
| want. */ |
| |
| if (var->kind != INTERNALVAR_MAKE_VALUE |
| && val->lval != lval_computed) |
| { |
| VALUE_LVAL (val) = lval_internalvar; |
| VALUE_INTERNALVAR (val) = var; |
| } |
| |
| return val; |
| } |
| |
| int |
| get_internalvar_integer (struct internalvar *var, LONGEST *result) |
| { |
| if (var->kind == INTERNALVAR_INTEGER) |
| { |
| *result = var->u.integer.val; |
| return 1; |
| } |
| |
| if (var->kind == INTERNALVAR_VALUE) |
| { |
| struct type *type = check_typedef (value_type (var->u.value)); |
| |
| if (type->code () == TYPE_CODE_INT) |
| { |
| *result = value_as_long (var->u.value); |
| return 1; |
| } |
| } |
| |
| return 0; |
| } |
| |
| static int |
| get_internalvar_function (struct internalvar *var, |
| struct internal_function **result) |
| { |
| switch (var->kind) |
| { |
| case INTERNALVAR_FUNCTION: |
| *result = var->u.fn.function; |
| return 1; |
| |
| default: |
| return 0; |
| } |
| } |
| |
| void |
| set_internalvar_component (struct internalvar *var, |
| LONGEST offset, LONGEST bitpos, |
| LONGEST bitsize, struct value *newval) |
| { |
| gdb_byte *addr; |
| struct gdbarch *arch; |
| int unit_size; |
| |
| switch (var->kind) |
| { |
| case INTERNALVAR_VALUE: |
| addr = value_contents_writeable (var->u.value); |
| arch = get_value_arch (var->u.value); |
| unit_size = gdbarch_addressable_memory_unit_size (arch); |
| |
| if (bitsize) |
| modify_field (value_type (var->u.value), addr + offset, |
| value_as_long (newval), bitpos, bitsize); |
| else |
| memcpy (addr + offset * unit_size, value_contents (newval), |
| TYPE_LENGTH (value_type (newval))); |
| break; |
| |
| default: |
| /* We can never get a component of any other kind. */ |
| internal_error (__FILE__, __LINE__, _("set_internalvar_component")); |
| } |
| } |
| |
| void |
| set_internalvar (struct internalvar *var, struct value *val) |
| { |
| enum internalvar_kind new_kind; |
| union internalvar_data new_data = { 0 }; |
| |
| if (var->kind == INTERNALVAR_FUNCTION && var->u.fn.canonical) |
| error (_("Cannot overwrite convenience function %s"), var->name); |
| |
| /* Prepare new contents. */ |
| switch (check_typedef (value_type (val))->code ()) |
| { |
| case TYPE_CODE_VOID: |
| new_kind = INTERNALVAR_VOID; |
| break; |
| |
| case TYPE_CODE_INTERNAL_FUNCTION: |
| gdb_assert (VALUE_LVAL (val) == lval_internalvar); |
| new_kind = INTERNALVAR_FUNCTION; |
| get_internalvar_function (VALUE_INTERNALVAR (val), |
| &new_data.fn.function); |
| /* Copies created here are never canonical. */ |
| break; |
| |
| default: |
| new_kind = INTERNALVAR_VALUE; |
| struct value *copy = value_copy (val); |
| copy->modifiable = 1; |
| |
| /* Force the value to be fetched from the target now, to avoid problems |
| later when this internalvar is referenced and the target is gone or |
| has changed. */ |
| if (value_lazy (copy)) |
| value_fetch_lazy (copy); |
| |
| /* Release the value from the value chain to prevent it from being |
| deleted by free_all_values. From here on this function should not |
| call error () until new_data is installed into the var->u to avoid |
| leaking memory. */ |
| new_data.value = release_value (copy).release (); |
| |
| /* Internal variables which are created from values with a dynamic |
| location don't need the location property of the origin anymore. |
| The resolved dynamic location is used prior then any other address |
| when accessing the value. |
| If we keep it, we would still refer to the origin value. |
| Remove the location property in case it exist. */ |
| value_type (new_data.value)->remove_dyn_prop (DYN_PROP_DATA_LOCATION); |
| |
| break; |
| } |
| |
| /* Clean up old contents. */ |
| clear_internalvar (var); |
| |
| /* Switch over. */ |
| var->kind = new_kind; |
| var->u = new_data; |
| /* End code which must not call error(). */ |
| } |
| |
| void |
| set_internalvar_integer (struct internalvar *var, LONGEST l) |
| { |
| /* Clean up old contents. */ |
| clear_internalvar (var); |
| |
| var->kind = INTERNALVAR_INTEGER; |
| var->u.integer.type = NULL; |
| var->u.integer.val = l; |
| } |
| |
| void |
| set_internalvar_string (struct internalvar *var, const char *string) |
| { |
| /* Clean up old contents. */ |
| clear_internalvar (var); |
| |
| var->kind = INTERNALVAR_STRING; |
| var->u.string = xstrdup (string); |
| } |
| |
| static void |
| set_internalvar_function (struct internalvar *var, struct internal_function *f) |
| { |
| /* Clean up old contents. */ |
| clear_internalvar (var); |
| |
| var->kind = INTERNALVAR_FUNCTION; |
| var->u.fn.function = f; |
| var->u.fn.canonical = 1; |
| /* Variables installed here are always the canonical version. */ |
| } |
| |
| void |
| clear_internalvar (struct internalvar *var) |
| { |
| /* Clean up old contents. */ |
| switch (var->kind) |
| { |
| case INTERNALVAR_VALUE: |
| value_decref (var->u.value); |
| break; |
| |
| case INTERNALVAR_STRING: |
| xfree (var->u.string); |
| break; |
| |
| case INTERNALVAR_MAKE_VALUE: |
| if (var->u.make_value.functions->destroy != NULL) |
| var->u.make_value.functions->destroy (var->u.make_value.data); |
| break; |
| |
| default: |
| break; |
| } |
| |
| /* Reset to void kind. */ |
| var->kind = INTERNALVAR_VOID; |
| } |
| |
| const char * |
| internalvar_name (const struct internalvar *var) |
| { |
| return var->name; |
| } |
| |
| static struct internal_function * |
| create_internal_function (const char *name, |
| internal_function_fn handler, void *cookie) |
| { |
| struct internal_function *ifn = XNEW (struct internal_function); |
| |
| ifn->name = xstrdup (name); |
| ifn->handler = handler; |
| ifn->cookie = cookie; |
| return ifn; |
| } |
| |
| const char * |
| value_internal_function_name (struct value *val) |
| { |
| struct internal_function *ifn; |
| int result; |
| |
| gdb_assert (VALUE_LVAL (val) == lval_internalvar); |
| result = get_internalvar_function (VALUE_INTERNALVAR (val), &ifn); |
| gdb_assert (result); |
| |
| return ifn->name; |
| } |
| |
| struct value * |
| call_internal_function (struct gdbarch *gdbarch, |
| const struct language_defn *language, |
| struct value *func, int argc, struct value **argv) |
| { |
| struct internal_function *ifn; |
| int result; |
| |
| gdb_assert (VALUE_LVAL (func) == lval_internalvar); |
| result = get_internalvar_function (VALUE_INTERNALVAR (func), &ifn); |
| gdb_assert (result); |
| |
| return (*ifn->handler) (gdbarch, language, ifn->cookie, argc, argv); |
| } |
| |
| /* The 'function' command. This does nothing -- it is just a |
| placeholder to let "help function NAME" work. This is also used as |
| the implementation of the sub-command that is created when |
| registering an internal function. */ |
| static void |
| function_command (const char *command, int from_tty) |
| { |
| /* Do nothing. */ |
| } |
| |
| /* Helper function that does the work for add_internal_function. */ |
| |
| static struct cmd_list_element * |
| do_add_internal_function (const char *name, const char *doc, |
| internal_function_fn handler, void *cookie) |
| { |
| struct internal_function *ifn; |
| struct internalvar *var = lookup_internalvar (name); |
| |
| ifn = create_internal_function (name, handler, cookie); |
| set_internalvar_function (var, ifn); |
| |
| return add_cmd (name, no_class, function_command, doc, &functionlist); |
| } |
| |
| /* See value.h. */ |
| |
| void |
| add_internal_function (const char *name, const char *doc, |
| internal_function_fn handler, void *cookie) |
| { |
| do_add_internal_function (name, doc, handler, cookie); |
| } |
| |
| /* See value.h. */ |
| |
| void |
| add_internal_function (gdb::unique_xmalloc_ptr<char> &&name, |
| gdb::unique_xmalloc_ptr<char> &&doc, |
| internal_function_fn handler, void *cookie) |
| { |
| struct cmd_list_element *cmd |
| = do_add_internal_function (name.get (), doc.get (), handler, cookie); |
| doc.release (); |
| cmd->doc_allocated = 1; |
| name.release (); |
| cmd->name_allocated = 1; |
| } |
| |
| /* Update VALUE before discarding OBJFILE. COPIED_TYPES is used to |
| prevent cycles / duplicates. */ |
| |
| void |
| preserve_one_value (struct value *value, struct objfile *objfile, |
| htab_t copied_types) |
| { |
| if (value->type->objfile_owner () == objfile) |
| value->type = copy_type_recursive (objfile, value->type, copied_types); |
| |
| if (value->enclosing_type->objfile_owner () == objfile) |
| value->enclosing_type = copy_type_recursive (objfile, |
| value->enclosing_type, |
| copied_types); |
| } |
| |
| /* Likewise for internal variable VAR. */ |
| |
| static void |
| preserve_one_internalvar (struct internalvar *var, struct objfile *objfile, |
| htab_t copied_types) |
| { |
| switch (var->kind) |
| { |
| case INTERNALVAR_INTEGER: |
| if (var->u.integer.type |
| && var->u.integer.type->objfile_owner () == objfile) |
| var->u.integer.type |
| = copy_type_recursive (objfile, var->u.integer.type, copied_types); |
| break; |
| |
| case INTERNALVAR_VALUE: |
| preserve_one_value (var->u.value, objfile, copied_types); |
| break; |
| } |
| } |
| |
| /* Update the internal variables and value history when OBJFILE is |
| discarded; we must copy the types out of the objfile. New global types |
| will be created for every convenience variable which currently points to |
| this objfile's types, and the convenience variables will be adjusted to |
| use the new global types. */ |
| |
| void |
| preserve_values (struct objfile *objfile) |
| { |
| struct internalvar *var; |
| |
| /* Create the hash table. We allocate on the objfile's obstack, since |
| it is soon to be deleted. */ |
| htab_up copied_types = create_copied_types_hash (objfile); |
| |
| for (const value_ref_ptr &item : value_history) |
| preserve_one_value (item.get (), objfile, copied_types.get ()); |
| |
| for (var = internalvars; var; var = var->next) |
| preserve_one_internalvar (var, objfile, copied_types.get ()); |
| |
| preserve_ext_lang_values (objfile, copied_types.get ()); |
| } |
| |
| static void |
| show_convenience (const char *ignore, int from_tty) |
| { |
| struct gdbarch *gdbarch = get_current_arch (); |
| struct internalvar *var; |
| int varseen = 0; |
| struct value_print_options opts; |
| |
| get_user_print_options (&opts); |
| for (var = internalvars; var; var = var->next) |
| { |
| |
| if (!varseen) |
| { |
| varseen = 1; |
| } |
| printf_filtered (("$%s = "), var->name); |
| |
| try |
| { |
| struct value *val; |
| |
| val = value_of_internalvar (gdbarch, var); |
| value_print (val, gdb_stdout, &opts); |
| } |
| catch (const gdb_exception_error &ex) |
| { |
| fprintf_styled (gdb_stdout, metadata_style.style (), |
| _("<error: %s>"), ex.what ()); |
| } |
| |
| printf_filtered (("\n")); |
| } |
| if (!varseen) |
| { |
| /* This text does not mention convenience functions on purpose. |
| The user can't create them except via Python, and if Python support |
| is installed this message will never be printed ($_streq will |
| exist). */ |
| printf_unfiltered (_("No debugger convenience variables now defined.\n" |
| "Convenience variables have " |
| "names starting with \"$\";\n" |
| "use \"set\" as in \"set " |
| "$foo = 5\" to define them.\n")); |
| } |
| } |
| |
| |
| /* See value.h. */ |
| |
| struct value * |
| value_from_xmethod (xmethod_worker_up &&worker) |
| { |
| struct value *v; |
| |
| v = allocate_value (builtin_type (target_gdbarch ())->xmethod); |
| v->lval = lval_xcallable; |
| v->location.xm_worker = worker.release (); |
| v->modifiable = 0; |
| |
| return v; |
| } |
| |
| /* Return the type of the result of TYPE_CODE_XMETHOD value METHOD. */ |
| |
| struct type * |
| result_type_of_xmethod (struct value *method, gdb::array_view<value *> argv) |
| { |
| gdb_assert (value_type (method)->code () == TYPE_CODE_XMETHOD |
| && method->lval == lval_xcallable && !argv.empty ()); |
| |
| return method->location.xm_worker->get_result_type (argv[0], argv.slice (1)); |
| } |
| |
| /* Call the xmethod corresponding to the TYPE_CODE_XMETHOD value METHOD. */ |
| |
| struct value * |
| call_xmethod (struct value *method, gdb::array_view<value *> argv) |
| { |
| gdb_assert (value_type (method)->code () == TYPE_CODE_XMETHOD |
| && method->lval == lval_xcallable && !argv.empty ()); |
| |
| return method->location.xm_worker->invoke (argv[0], argv.slice (1)); |
| } |
| |
| /* Extract a value as a C number (either long or double). |
| Knows how to convert fixed values to double, or |
| floating values to long. |
| Does not deallocate the value. */ |
| |
| LONGEST |
| value_as_long (struct value *val) |
| { |
| /* This coerces arrays and functions, which is necessary (e.g. |
| in disassemble_command). It also dereferences references, which |
| I suspect is the most logical thing to do. */ |
| val = coerce_array (val); |
| return unpack_long (value_type (val), value_contents (val)); |
| } |
| |
| /* Extract a value as a C pointer. Does not deallocate the value. |
| Note that val's type may not actually be a pointer; value_as_long |
| handles all the cases. */ |
| CORE_ADDR |
| value_as_address (struct value *val) |
| { |
| struct gdbarch *gdbarch = value_type (val)->arch (); |
| |
| /* Assume a CORE_ADDR can fit in a LONGEST (for now). Not sure |
| whether we want this to be true eventually. */ |
| #if 0 |
| /* gdbarch_addr_bits_remove is wrong if we are being called for a |
| non-address (e.g. argument to "signal", "info break", etc.), or |
| for pointers to char, in which the low bits *are* significant. */ |
| return gdbarch_addr_bits_remove (gdbarch, value_as_long (val)); |
| #else |
| |
| /* There are several targets (IA-64, PowerPC, and others) which |
| don't represent pointers to functions as simply the address of |
| the function's entry point. For example, on the IA-64, a |
| function pointer points to a two-word descriptor, generated by |
| the linker, which contains the function's entry point, and the |
| value the IA-64 "global pointer" register should have --- to |
| support position-independent code. The linker generates |
| descriptors only for those functions whose addresses are taken. |
| |
| On such targets, it's difficult for GDB to convert an arbitrary |
| function address into a function pointer; it has to either find |
| an existing descriptor for that function, or call malloc and |
| build its own. On some targets, it is impossible for GDB to |
| build a descriptor at all: the descriptor must contain a jump |
| instruction; data memory cannot be executed; and code memory |
| cannot be modified. |
| |
| Upon entry to this function, if VAL is a value of type `function' |
| (that is, TYPE_CODE (VALUE_TYPE (val)) == TYPE_CODE_FUNC), then |
| value_address (val) is the address of the function. This is what |
| you'll get if you evaluate an expression like `main'. The call |
| to COERCE_ARRAY below actually does all the usual unary |
| conversions, which includes converting values of type `function' |
| to `pointer to function'. This is the challenging conversion |
| discussed above. Then, `unpack_long' will convert that pointer |
| back into an address. |
| |
| So, suppose the user types `disassemble foo' on an architecture |
| with a strange function pointer representation, on which GDB |
| cannot build its own descriptors, and suppose further that `foo' |
| has no linker-built descriptor. The address->pointer conversion |
| will signal an error and prevent the command from running, even |
| though the next step would have been to convert the pointer |
| directly back into the same address. |
| |
| The following shortcut avoids this whole mess. If VAL is a |
| function, just return its address directly. */ |
| if (value_type (val)->code () == TYPE_CODE_FUNC |
| || value_type (val)->code () == TYPE_CODE_METHOD) |
| return value_address (val); |
| |
| val = coerce_array (val); |
| |
| /* Some architectures (e.g. Harvard), map instruction and data |
| addresses onto a single large unified address space. For |
| instance: An architecture may consider a large integer in the |
| range 0x10000000 .. 0x1000ffff to already represent a data |
| addresses (hence not need a pointer to address conversion) while |
| a small integer would still need to be converted integer to |
| pointer to address. Just assume such architectures handle all |
| integer conversions in a single function. */ |
| |
| /* JimB writes: |
| |
| I think INTEGER_TO_ADDRESS is a good idea as proposed --- but we |
| must admonish GDB hackers to make sure its behavior matches the |
| compiler's, whenever possible. |
| |
| In general, I think GDB should evaluate expressions the same way |
| the compiler does. When the user copies an expression out of |
| their source code and hands it to a `print' command, they should |
| get the same value the compiler would have computed. Any |
| deviation from this rule can cause major confusion and annoyance, |
| and needs to be justified carefully. In other words, GDB doesn't |
| really have the freedom to do these conversions in clever and |
| useful ways. |
| |
| AndrewC pointed out that users aren't complaining about how GDB |
| casts integers to pointers; they are complaining that they can't |
| take an address from a disassembly listing and give it to `x/i'. |
| This is certainly important. |
| |
| Adding an architecture method like integer_to_address() certainly |
| makes it possible for GDB to "get it right" in all circumstances |
| --- the target has complete control over how things get done, so |
| people can Do The Right Thing for their target without breaking |
| anyone else. The standard doesn't specify how integers get |
| converted to pointers; usually, the ABI doesn't either, but |
| ABI-specific code is a more reasonable place to handle it. */ |
| |
| if (!value_type (val)->is_pointer_or_reference () |
| && gdbarch_integer_to_address_p (gdbarch)) |
| return gdbarch_integer_to_address (gdbarch, value_type (val), |
| value_contents (val)); |
| |
| return unpack_long (value_type (val), value_contents (val)); |
| #endif |
| } |
| |
| /* Unpack raw data (copied from debugee, target byte order) at VALADDR |
| as a long, or as a double, assuming the raw data is described |
| by type TYPE. Knows how to convert different sizes of values |
| and can convert between fixed and floating point. We don't assume |
| any alignment for the raw data. Return value is in host byte order. |
| |
| If you want functions and arrays to be coerced to pointers, and |
| references to be dereferenced, call value_as_long() instead. |
| |
| C++: It is assumed that the front-end has taken care of |
| all matters concerning pointers to members. A pointer |
| to member which reaches here is considered to be equivalent |
| to an INT (or some size). After all, it is only an offset. */ |
| |
| LONGEST |
| unpack_long (struct type *type, const gdb_byte *valaddr) |
| { |
| if (is_fixed_point_type (type)) |
| type = type->fixed_point_type_base_type (); |
| |
| enum bfd_endian byte_order = type_byte_order (type); |
| enum type_code code = type->code (); |
| int len = TYPE_LENGTH (type); |
| int nosign = type->is_unsigned (); |
| |
| switch (code) |
| { |
| case TYPE_CODE_TYPEDEF: |
| return unpack_long (check_typedef (type), valaddr); |
| case TYPE_CODE_ENUM: |
| case TYPE_CODE_FLAGS: |
| case TYPE_CODE_BOOL: |
| case TYPE_CODE_INT: |
| case TYPE_CODE_CHAR: |
| case TYPE_CODE_RANGE: |
| case TYPE_CODE_MEMBERPTR: |
| { |
| LONGEST result; |
| |
| if (type->bit_size_differs_p ()) |
| { |
| unsigned bit_off = type->bit_offset (); |
| unsigned bit_size = type->bit_size (); |
| if (bit_size == 0) |
| { |
| /* unpack_bits_as_long doesn't handle this case the |
| way we'd like, so handle it here. */ |
| result = 0; |
| } |
| else |
| result = unpack_bits_as_long (type, valaddr, bit_off, bit_size); |
| } |
| else |
| { |
| if (nosign) |
| result = extract_unsigned_integer (valaddr, len, byte_order); |
| else |
| result = extract_signed_integer (valaddr, len, byte_order); |
| } |
| if (code == TYPE_CODE_RANGE) |
| result += type->bounds ()->bias; |
| return result; |
| } |
| |
| case TYPE_CODE_FLT: |
| case TYPE_CODE_DECFLOAT: |
| return target_float_to_longest (valaddr, type); |
| |
| case TYPE_CODE_FIXED_POINT: |
| { |
| gdb_mpq vq; |
| vq.read_fixed_point (gdb::make_array_view (valaddr, len), |
| byte_order, nosign, |
| type->fixed_point_scaling_factor ()); |
| |
| gdb_mpz vz; |
| mpz_tdiv_q (vz.val, mpq_numref (vq.val), mpq_denref (vq.val)); |
| return vz.as_integer<LONGEST> (); |
| } |
| |
| case TYPE_CODE_PTR: |
| case TYPE_CODE_REF: |
| case TYPE_CODE_RVALUE_REF: |
| /* Assume a CORE_ADDR can fit in a LONGEST (for now). Not sure |
| whether we want this to be true eventually. */ |
| return extract_typed_address (valaddr, type); |
| |
| default: |
| error (_("Value can't be converted to integer.")); |
| } |
| } |
| |
| /* Unpack raw data (copied from debugee, target byte order) at VALADDR |
| as a CORE_ADDR, assuming the raw data is described by type TYPE. |
| We don't assume any alignment for the raw data. Return value is in |
| host byte order. |
| |
| If you want functions and arrays to be coerced to pointers, and |
| references to be dereferenced, call value_as_address() instead. |
| |
| C++: It is assumed that the front-end has taken care of |
| all matters concerning pointers to members. A pointer |
| to member which reaches here is considered to be equivalent |
| to an INT (or some size). After all, it is only an offset. */ |
| |
| CORE_ADDR |
| unpack_pointer (struct type *type, const gdb_byte *valaddr) |
| { |
| /* Assume a CORE_ADDR can fit in a LONGEST (for now). Not sure |
| whether we want this to be true eventually. */ |
| return unpack_long (type, valaddr); |
| } |
| |
| bool |
| is_floating_value (struct value *val) |
| { |
| struct type *type = check_typedef (value_type (val)); |
| |
| if (is_floating_type (type)) |
| { |
| if (!target_float_is_valid (value_contents (val), type)) |
| error (_("Invalid floating value found in program.")); |
| return true; |
| } |
| |
| return false; |
| } |
| |
| |
| /* Get the value of the FIELDNO'th field (which must be static) of |
| TYPE. */ |
| |
| struct value * |
| value_static_field (struct type *type, int fieldno) |
| { |
| struct value *retval; |
| |
| switch (TYPE_FIELD_LOC_KIND (type, fieldno)) |
| { |
| case FIELD_LOC_KIND_PHYSADDR: |
| retval = value_at_lazy (type->field (fieldno).type (), |
| TYPE_FIELD_STATIC_PHYSADDR (type, fieldno)); |
| break; |
| case FIELD_LOC_KIND_PHYSNAME: |
| { |
| const char *phys_name = TYPE_FIELD_STATIC_PHYSNAME (type, fieldno); |
| /* type->field (fieldno).name (); */ |
| struct block_symbol sym = lookup_symbol (phys_name, 0, VAR_DOMAIN, 0); |
| |
| if (sym.symbol == NULL) |
| { |
| /* With some compilers, e.g. HP aCC, static data members are |
| reported as non-debuggable symbols. */ |
| struct bound_minimal_symbol msym |
| = lookup_minimal_symbol (phys_name, NULL, NULL); |
| struct type *field_type = type->field (fieldno).type (); |
| |
| if (!msym.minsym) |
| retval = allocate_optimized_out_value (field_type); |
| else |
| retval = value_at_lazy (field_type, BMSYMBOL_VALUE_ADDRESS (msym)); |
| } |
| else |
| retval = value_of_variable (sym.symbol, sym.block); |
| break; |
| } |
| default: |
| gdb_assert_not_reached ("unexpected field location kind"); |
| } |
| |
| return retval; |
| } |
| |
| /* Change the enclosing type of a value object VAL to NEW_ENCL_TYPE. |
| You have to be careful here, since the size of the data area for the value |
| is set by the length of the enclosing type. So if NEW_ENCL_TYPE is bigger |
| than the old enclosing type, you have to allocate more space for the |
| data. */ |
| |
| void |
| set_value_enclosing_type (struct value *val, struct type *new_encl_type) |
| { |
| if (TYPE_LENGTH (new_encl_type) > TYPE_LENGTH (value_enclosing_type (val))) |
| { |
| check_type_length_before_alloc (new_encl_type); |
| val->contents |
| .reset ((gdb_byte *) xrealloc (val->contents.release (), |
| TYPE_LENGTH (new_encl_type))); |
| } |
| |
| val->enclosing_type = new_encl_type; |
| } |
| |
| /* Given a value ARG1 (offset by OFFSET bytes) |
| of a struct or union type ARG_TYPE, |
| extract and return the value of one of its (non-static) fields. |
| FIELDNO says which field. */ |
| |
| struct value * |
| value_primitive_field (struct value *arg1, LONGEST offset, |
| int fieldno, struct type *arg_type) |
| { |
| struct value *v; |
| struct type *type; |
| struct gdbarch *arch = get_value_arch (arg1); |
| int unit_size = gdbarch_addressable_memory_unit_size (arch); |
| |
| arg_type = check_typedef (arg_type); |
| type = arg_type->field (fieldno).type (); |
| |
| /* Call check_typedef on our type to make sure that, if TYPE |
| is a TYPE_CODE_TYPEDEF, its length is set to the length |
| of the target type instead of zero. However, we do not |
| replace the typedef type by the target type, because we want |
| to keep the typedef in order to be able to print the type |
| description correctly. */ |
| check_typedef (type); |
| |
| if (TYPE_FIELD_BITSIZE (arg_type, fieldno)) |
| { |
| /* Handle packed fields. |
| |
| Create a new value for the bitfield, with bitpos and bitsize |
| set. If possible, arrange offset and bitpos so that we can |
| do a single aligned read of the size of the containing type. |
| Otherwise, adjust offset to the byte containing the first |
| bit. Assume that the address, offset, and embedded offset |
| are sufficiently aligned. */ |
| |
| LONGEST bitpos = TYPE_FIELD_BITPOS (arg_type, fieldno); |
| LONGEST container_bitsize = TYPE_LENGTH (type) * 8; |
| |
| v = allocate_value_lazy (type); |
| v->bitsize = TYPE_FIELD_BITSIZE (arg_type, fieldno); |
| if ((bitpos % container_bitsize) + v->bitsize <= container_bitsize |
| && TYPE_LENGTH (type) <= (int) sizeof (LONGEST)) |
| v->bitpos = bitpos % container_bitsize; |
| else |
| v->bitpos = bitpos % 8; |
| v->offset = (value_embedded_offset (arg1) |
| + offset |
| + (bitpos - v->bitpos) / 8); |
| set_value_parent (v, arg1); |
| if (!value_lazy (arg1)) |
| value_fetch_lazy (v); |
| } |
| else if (fieldno < TYPE_N_BASECLASSES (arg_type)) |
| { |
| /* This field is actually a base subobject, so preserve the |
| entire object's contents for later references to virtual |
| bases, etc. */ |
| LONGEST boffset; |
| |
| /* Lazy register values with offsets are not supported. */ |
| if (VALUE_LVAL (arg1) == lval_register && value_lazy (arg1)) |
| value_fetch_lazy (arg1); |
| |
| /* We special case virtual inheritance here because this |
| requires access to the contents, which we would rather avoid |
| for references to ordinary fields of unavailable values. */ |
| if (BASETYPE_VIA_VIRTUAL (arg_type, fieldno)) |
| boffset = baseclass_offset (arg_type, fieldno, |
| value_contents (arg1), |
| value_embedded_offset (arg1), |
| value_address (arg1), |
| arg1); |
| else |
| boffset = TYPE_FIELD_BITPOS (arg_type, fieldno) / 8; |
| |
| if (value_lazy (arg1)) |
| v = allocate_value_lazy (value_enclosing_type (arg1)); |
| else |
| { |
| v = allocate_value (value_enclosing_type (arg1)); |
| value_contents_copy_raw (v, 0, arg1, 0, |
| TYPE_LENGTH (value_enclosing_type (arg1))); |
| } |
| v->type = type; |
| v->offset = value_offset (arg1); |
| v->embedded_offset = offset + value_embedded_offset (arg1) + boffset; |
| } |
| else if (NULL != TYPE_DATA_LOCATION (type)) |
| { |
| /* Field is a dynamic data member. */ |
| |
| gdb_assert (0 == offset); |
| /* We expect an already resolved data location. */ |
| gdb_assert (PROP_CONST == TYPE_DATA_LOCATION_KIND (type)); |
| /* For dynamic data types defer memory allocation |
| until we actual access the value. */ |
| v = allocate_value_lazy (type); |
| } |
| else |
| { |
| /* Plain old data member */ |
| offset += (TYPE_FIELD_BITPOS (arg_type, fieldno) |
| / (HOST_CHAR_BIT * unit_size)); |
| |
| /* Lazy register values with offsets are not supported. */ |
| if (VALUE_LVAL (arg1) == lval_register && value_lazy (arg1)) |
| value_fetch_lazy (arg1); |
| |
| if (value_lazy (arg1)) |
| v = allocate_value_lazy (type); |
| else |
| { |
| v = allocate_value (type); |
| value_contents_copy_raw (v, value_embedded_offset (v), |
| arg1, value_embedded_offset (arg1) + offset, |
| type_length_units (type)); |
| } |
| v->offset = (value_offset (arg1) + offset |
| + value_embedded_offset (arg1)); |
| } |
| set_value_component_location (v, arg1); |
| return v; |
| } |
| |
| /* Given a value ARG1 of a struct or union type, |
| extract and return the value of one of its (non-static) fields. |
| FIELDNO says which field. */ |
| |
| struct value * |
| value_field (struct value *arg1, int fieldno) |
| { |
| return value_primitive_field (arg1, 0, fieldno, value_type (arg1)); |
| } |
| |
| /* Return a non-virtual function as a value. |
| F is the list of member functions which contains the desired method. |
| J is an index into F which provides the desired method. |
| |
| We only use the symbol for its address, so be happy with either a |
| full symbol or a minimal symbol. */ |
| |
| struct value * |
| value_fn_field (struct value **arg1p, struct fn_field *f, |
| int j, struct type *type, |
| LONGEST offset) |
| { |
| struct value *v; |
| struct type *ftype = TYPE_FN_FIELD_TYPE (f, j); |
| const char *physname = TYPE_FN_FIELD_PHYSNAME (f, j); |
| struct symbol *sym; |
| struct bound_minimal_symbol msym; |
| |
| sym = lookup_symbol (physname, 0, VAR_DOMAIN, 0).symbol; |
| if (sym != NULL) |
| { |
| memset (&msym, 0, sizeof (msym)); |
| } |
| else |
| { |
| gdb_assert (sym == NULL); |
| msym = lookup_bound_minimal_symbol (physname); |
| if (msym.minsym == NULL) |
| return NULL; |
| } |
| |
| v = allocate_value (ftype); |
| VALUE_LVAL (v) = lval_memory; |
| if (sym) |
| { |
| set_value_address (v, BLOCK_ENTRY_PC (SYMBOL_BLOCK_VALUE (sym))); |
| } |
| else |
| { |
| /* The minimal symbol might point to a function descriptor; |
| resolve it to the actual code address instead. */ |
| struct objfile *objfile = msym.objfile; |
| struct gdbarch *gdbarch = objfile->arch (); |
| |
| set_value_address (v, |
| gdbarch_convert_from_func_ptr_addr |
| (gdbarch, BMSYMBOL_VALUE_ADDRESS (msym), |
| current_inferior ()->top_target ())); |
| } |
| |
| if (arg1p) |
| { |
| if (type != value_type (*arg1p)) |
| *arg1p = value_ind (value_cast (lookup_pointer_type (type), |
| value_addr (*arg1p))); |
| |
| /* Move the `this' pointer according to the offset. |
| VALUE_OFFSET (*arg1p) += offset; */ |
| } |
| |
| return v; |
| } |
| |
| |
| |
| /* See value.h. */ |
| |
| LONGEST |
| unpack_bits_as_long (struct type *field_type, const gdb_byte *valaddr, |
| LONGEST bitpos, LONGEST bitsize) |
| { |
| enum bfd_endian byte_order = type_byte_order (field_type); |
| ULONGEST val; |
| ULONGEST valmask; |
| int lsbcount; |
| LONGEST bytes_read; |
| LONGEST read_offset; |
| |
| /* Read the minimum number of bytes required; there may not be |
| enough bytes to read an entire ULONGEST. */ |
| field_type = check_typedef (field_type); |
| if (bitsize) |
| bytes_read = ((bitpos % 8) + bitsize + 7) / 8; |
| else |
| { |
| bytes_read = TYPE_LENGTH (field_type); |
| bitsize = 8 * bytes_read; |
| } |
| |
| read_offset = bitpos / 8; |
| |
| val = extract_unsigned_integer (valaddr + read_offset, |
| bytes_read, byte_order); |
| |
| /* Extract bits. See comment above. */ |
| |
| if (byte_order == BFD_ENDIAN_BIG) |
| lsbcount = (bytes_read * 8 - bitpos % 8 - bitsize); |
| else |
| lsbcount = (bitpos % 8); |
| val >>= lsbcount; |
| |
| /* If the field does not entirely fill a LONGEST, then zero the sign bits. |
| If the field is signed, and is negative, then sign extend. */ |
| |
| if (bitsize < 8 * (int) sizeof (val)) |
| { |
| valmask = (((ULONGEST) 1) << bitsize) - 1; |
| val &= valmask; |
| if (!field_type->is_unsigned ()) |
| { |
| if (val & (valmask ^ (valmask >> 1))) |
| { |
| val |= ~valmask; |
| } |
| } |
| } |
| |
| return val; |
| } |
| |
| /* Unpack a field FIELDNO of the specified TYPE, from the object at |
| VALADDR + EMBEDDED_OFFSET. VALADDR points to the contents of |
| ORIGINAL_VALUE, which must not be NULL. See |
| unpack_value_bits_as_long for more details. */ |
| |
| int |
| unpack_value_field_as_long (struct type *type, const gdb_byte *valaddr, |
| LONGEST embedded_offset, int fieldno, |
| const struct value *val, LONGEST *result) |
| { |
| int bitpos = TYPE_FIELD_BITPOS (type, fieldno); |
| int bitsize = TYPE_FIELD_BITSIZE (type, fieldno); |
| struct type *field_type = type->field (fieldno).type (); |
| int bit_offset; |
| |
| gdb_assert (val != NULL); |
| |
| bit_offset = embedded_offset * TARGET_CHAR_BIT + bitpos; |
| if (value_bits_any_optimized_out (val, bit_offset, bitsize) |
| || !value_bits_available (val, bit_offset, bitsize)) |
| return 0; |
| |
| *result = unpack_bits_as_long (field_type, valaddr + embedded_offset, |
| bitpos, bitsize); |
| return 1; |
| } |
| |
| /* Unpack a field FIELDNO of the specified TYPE, from the anonymous |
| object at VALADDR. See unpack_bits_as_long for more details. */ |
| |
| LONGEST |
| unpack_field_as_long (struct type *type, const gdb_byte *valaddr, int fieldno) |
| { |
| int bitpos = TYPE_FIELD_BITPOS (type, fieldno); |
| int bitsize = TYPE_FIELD_BITSIZE (type, fieldno); |
| struct type *field_type = type->field (fieldno).type (); |
| |
| return unpack_bits_as_long (field_type, valaddr, bitpos, bitsize); |
| } |
| |
| /* Unpack a bitfield of BITSIZE bits found at BITPOS in the object at |
| VALADDR + EMBEDDEDOFFSET that has the type of DEST_VAL and store |
| the contents in DEST_VAL, zero or sign extending if the type of |
| DEST_VAL is wider than BITSIZE. VALADDR points to the contents of |
| VAL. If the VAL's contents required to extract the bitfield from |
| are unavailable/optimized out, DEST_VAL is correspondingly |
| marked unavailable/optimized out. */ |
| |
| void |
| unpack_value_bitfield (struct value *dest_val, |
| LONGEST bitpos, LONGEST bitsize, |
| const gdb_byte *valaddr, LONGEST embedded_offset, |
| const struct value *val) |
| { |
| enum bfd_endian byte_order; |
| int src_bit_offset; |
| int dst_bit_offset; |
| struct type *field_type = value_type (dest_val); |
| |
| byte_order = type_byte_order (field_type); |
| |
| /* First, unpack and sign extend the bitfield as if it was wholly |
| valid. Optimized out/unavailable bits are read as zero, but |
| that's OK, as they'll end up marked below. If the VAL is |
| wholly-invalid we may have skipped allocating its contents, |
| though. See allocate_optimized_out_value. */ |
| if (valaddr != NULL) |
| { |
| LONGEST num; |
| |
| num = unpack_bits_as_long (field_type, valaddr + embedded_offset, |
| bitpos, bitsize); |
| store_signed_integer (value_contents_raw (dest_val), |
| TYPE_LENGTH (field_type), byte_order, num); |
| } |
| |
| /* Now copy the optimized out / unavailability ranges to the right |
| bits. */ |
| src_bit_offset = embedded_offset * TARGET_CHAR_BIT + bitpos; |
| if (byte_order == BFD_ENDIAN_BIG) |
| dst_bit_offset = TYPE_LENGTH (field_type) * TARGET_CHAR_BIT - bitsize; |
| else |
| dst_bit_offset = 0; |
| value_ranges_copy_adjusted (dest_val, dst_bit_offset, |
| val, src_bit_offset, bitsize); |
| } |
| |
| /* Return a new value with type TYPE, which is FIELDNO field of the |
| object at VALADDR + EMBEDDEDOFFSET. VALADDR points to the contents |
| of VAL. If the VAL's contents required to extract the bitfield |
| from are unavailable/optimized out, the new value is |
| correspondingly marked unavailable/optimized out. */ |
| |
| struct value * |
| value_field_bitfield (struct type *type, int fieldno, |
| const gdb_byte *valaddr, |
| LONGEST embedded_offset, const struct value *val) |
| { |
| int bitpos = TYPE_FIELD_BITPOS (type, fieldno); |
| int bitsize = TYPE_FIELD_BITSIZE (type, fieldno); |
| struct value *res_val = allocate_value (type->field (fieldno).type ()); |
| |
| unpack_value_bitfield (res_val, bitpos, bitsize, |
| valaddr, embedded_offset, val); |
| |
| return res_val; |
| } |
| |
| /* Modify the value of a bitfield. ADDR points to a block of memory in |
| target byte order; the bitfield starts in the byte pointed to. FIELDVAL |
| is the desired value of the field, in host byte order. BITPOS and BITSIZE |
| indicate which bits (in target bit order) comprise the bitfield. |
| Requires 0 < BITSIZE <= lbits, 0 <= BITPOS % 8 + BITSIZE <= lbits, and |
| 0 <= BITPOS, where lbits is the size of a LONGEST in bits. */ |
| |
| void |
| |