| /* elf.c -- Get debug data from an ELF file for backtraces. |
| Copyright (C) 2012-2024 Free Software Foundation, Inc. |
| Written by Ian Lance Taylor, Google. |
| |
| Redistribution and use in source and binary forms, with or without |
| modification, are permitted provided that the following conditions are |
| met: |
| |
| (1) Redistributions of source code must retain the above copyright |
| notice, this list of conditions and the following disclaimer. |
| |
| (2) Redistributions in binary form must reproduce the above copyright |
| notice, this list of conditions and the following disclaimer in |
| the documentation and/or other materials provided with the |
| distribution. |
| |
| (3) The name of the author may not be used to |
| endorse or promote products derived from this software without |
| specific prior written permission. |
| |
| THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
| IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED |
| WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
| DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, |
| INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES |
| (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR |
| SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) |
| HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, |
| STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING |
| IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| POSSIBILITY OF SUCH DAMAGE. */ |
| |
| #include "config.h" |
| |
| #include <errno.h> |
| #include <stdlib.h> |
| #include <string.h> |
| #include <sys/types.h> |
| #include <sys/stat.h> |
| #include <unistd.h> |
| |
| #ifdef HAVE_DL_ITERATE_PHDR |
| #ifdef HAVE_LINK_H |
| #include <link.h> |
| #endif |
| #ifdef HAVE_SYS_LINK_H |
| #include <sys/link.h> |
| #endif |
| #endif |
| |
| #include "backtrace.h" |
| #include "internal.h" |
| |
| #ifndef S_ISLNK |
| #ifndef S_IFLNK |
| #define S_IFLNK 0120000 |
| #endif |
| #ifndef S_IFMT |
| #define S_IFMT 0170000 |
| #endif |
| #define S_ISLNK(m) (((m) & S_IFMT) == S_IFLNK) |
| #endif |
| |
| #ifndef __GNUC__ |
| #define __builtin_prefetch(p, r, l) |
| #define unlikely(x) (x) |
| #else |
| #define unlikely(x) __builtin_expect(!!(x), 0) |
| #endif |
| |
| #if !defined(HAVE_DECL_STRNLEN) || !HAVE_DECL_STRNLEN |
| |
| /* If strnlen is not declared, provide our own version. */ |
| |
| static size_t |
| xstrnlen (const char *s, size_t maxlen) |
| { |
| size_t i; |
| |
| for (i = 0; i < maxlen; ++i) |
| if (s[i] == '\0') |
| break; |
| return i; |
| } |
| |
| #define strnlen xstrnlen |
| |
| #endif |
| |
| #ifndef HAVE_LSTAT |
| |
| /* Dummy version of lstat for systems that don't have it. */ |
| |
| static int |
| xlstat (const char *path ATTRIBUTE_UNUSED, struct stat *st ATTRIBUTE_UNUSED) |
| { |
| return -1; |
| } |
| |
| #define lstat xlstat |
| |
| #endif |
| |
| #ifndef HAVE_READLINK |
| |
| /* Dummy version of readlink for systems that don't have it. */ |
| |
| static ssize_t |
| xreadlink (const char *path ATTRIBUTE_UNUSED, char *buf ATTRIBUTE_UNUSED, |
| size_t bufsz ATTRIBUTE_UNUSED) |
| { |
| return -1; |
| } |
| |
| #define readlink xreadlink |
| |
| #endif |
| |
| #ifndef HAVE_DL_ITERATE_PHDR |
| |
| /* Dummy version of dl_iterate_phdr for systems that don't have it. */ |
| |
| #define dl_phdr_info x_dl_phdr_info |
| #define dl_iterate_phdr x_dl_iterate_phdr |
| |
| struct dl_phdr_info |
| { |
| uintptr_t dlpi_addr; |
| const char *dlpi_name; |
| }; |
| |
| static int |
| dl_iterate_phdr (int (*callback) (struct dl_phdr_info *, |
| size_t, void *) ATTRIBUTE_UNUSED, |
| void *data ATTRIBUTE_UNUSED) |
| { |
| return 0; |
| } |
| |
| #endif /* ! defined (HAVE_DL_ITERATE_PHDR) */ |
| |
| /* The configure script must tell us whether we are 32-bit or 64-bit |
| ELF. We could make this code test and support either possibility, |
| but there is no point. This code only works for the currently |
| running executable, which means that we know the ELF mode at |
| configure time. */ |
| |
| #if BACKTRACE_ELF_SIZE != 32 && BACKTRACE_ELF_SIZE != 64 |
| #error "Unknown BACKTRACE_ELF_SIZE" |
| #endif |
| |
| /* <link.h> might #include <elf.h> which might define our constants |
| with slightly different values. Undefine them to be safe. */ |
| |
| #undef EI_NIDENT |
| #undef EI_MAG0 |
| #undef EI_MAG1 |
| #undef EI_MAG2 |
| #undef EI_MAG3 |
| #undef EI_CLASS |
| #undef EI_DATA |
| #undef EI_VERSION |
| #undef ELF_MAG0 |
| #undef ELF_MAG1 |
| #undef ELF_MAG2 |
| #undef ELF_MAG3 |
| #undef ELFCLASS32 |
| #undef ELFCLASS64 |
| #undef ELFDATA2LSB |
| #undef ELFDATA2MSB |
| #undef EV_CURRENT |
| #undef ET_DYN |
| #undef EM_PPC64 |
| #undef EF_PPC64_ABI |
| #undef SHN_LORESERVE |
| #undef SHN_XINDEX |
| #undef SHN_UNDEF |
| #undef SHT_PROGBITS |
| #undef SHT_SYMTAB |
| #undef SHT_STRTAB |
| #undef SHT_DYNSYM |
| #undef SHF_COMPRESSED |
| #undef STT_OBJECT |
| #undef STT_FUNC |
| #undef NT_GNU_BUILD_ID |
| #undef ELFCOMPRESS_ZLIB |
| #undef ELFCOMPRESS_ZSTD |
| |
| /* Basic types. */ |
| |
| typedef uint16_t b_elf_half; /* Elf_Half. */ |
| typedef uint32_t b_elf_word; /* Elf_Word. */ |
| typedef int32_t b_elf_sword; /* Elf_Sword. */ |
| |
| #if BACKTRACE_ELF_SIZE == 32 |
| |
| typedef uint32_t b_elf_addr; /* Elf_Addr. */ |
| typedef uint32_t b_elf_off; /* Elf_Off. */ |
| |
| typedef uint32_t b_elf_wxword; /* 32-bit Elf_Word, 64-bit ELF_Xword. */ |
| |
| #else |
| |
| typedef uint64_t b_elf_addr; /* Elf_Addr. */ |
| typedef uint64_t b_elf_off; /* Elf_Off. */ |
| typedef uint64_t b_elf_xword; /* Elf_Xword. */ |
| typedef int64_t b_elf_sxword; /* Elf_Sxword. */ |
| |
| typedef uint64_t b_elf_wxword; /* 32-bit Elf_Word, 64-bit ELF_Xword. */ |
| |
| #endif |
| |
| /* Data structures and associated constants. */ |
| |
| #define EI_NIDENT 16 |
| |
| typedef struct { |
| unsigned char e_ident[EI_NIDENT]; /* ELF "magic number" */ |
| b_elf_half e_type; /* Identifies object file type */ |
| b_elf_half e_machine; /* Specifies required architecture */ |
| b_elf_word e_version; /* Identifies object file version */ |
| b_elf_addr e_entry; /* Entry point virtual address */ |
| b_elf_off e_phoff; /* Program header table file offset */ |
| b_elf_off e_shoff; /* Section header table file offset */ |
| b_elf_word e_flags; /* Processor-specific flags */ |
| b_elf_half e_ehsize; /* ELF header size in bytes */ |
| b_elf_half e_phentsize; /* Program header table entry size */ |
| b_elf_half e_phnum; /* Program header table entry count */ |
| b_elf_half e_shentsize; /* Section header table entry size */ |
| b_elf_half e_shnum; /* Section header table entry count */ |
| b_elf_half e_shstrndx; /* Section header string table index */ |
| } b_elf_ehdr; /* Elf_Ehdr. */ |
| |
| #define EI_MAG0 0 |
| #define EI_MAG1 1 |
| #define EI_MAG2 2 |
| #define EI_MAG3 3 |
| #define EI_CLASS 4 |
| #define EI_DATA 5 |
| #define EI_VERSION 6 |
| |
| #define ELFMAG0 0x7f |
| #define ELFMAG1 'E' |
| #define ELFMAG2 'L' |
| #define ELFMAG3 'F' |
| |
| #define ELFCLASS32 1 |
| #define ELFCLASS64 2 |
| |
| #define ELFDATA2LSB 1 |
| #define ELFDATA2MSB 2 |
| |
| #define EV_CURRENT 1 |
| |
| #define ET_DYN 3 |
| |
| #define EM_PPC64 21 |
| #define EF_PPC64_ABI 3 |
| |
| typedef struct { |
| b_elf_word sh_name; /* Section name, index in string tbl */ |
| b_elf_word sh_type; /* Type of section */ |
| b_elf_wxword sh_flags; /* Miscellaneous section attributes */ |
| b_elf_addr sh_addr; /* Section virtual addr at execution */ |
| b_elf_off sh_offset; /* Section file offset */ |
| b_elf_wxword sh_size; /* Size of section in bytes */ |
| b_elf_word sh_link; /* Index of another section */ |
| b_elf_word sh_info; /* Additional section information */ |
| b_elf_wxword sh_addralign; /* Section alignment */ |
| b_elf_wxword sh_entsize; /* Entry size if section holds table */ |
| } b_elf_shdr; /* Elf_Shdr. */ |
| |
| #define SHN_UNDEF 0x0000 /* Undefined section */ |
| #define SHN_LORESERVE 0xFF00 /* Begin range of reserved indices */ |
| #define SHN_XINDEX 0xFFFF /* Section index is held elsewhere */ |
| |
| #define SHT_PROGBITS 1 |
| #define SHT_SYMTAB 2 |
| #define SHT_STRTAB 3 |
| #define SHT_DYNSYM 11 |
| |
| #define SHF_COMPRESSED 0x800 |
| |
| #if BACKTRACE_ELF_SIZE == 32 |
| |
| typedef struct |
| { |
| b_elf_word st_name; /* Symbol name, index in string tbl */ |
| b_elf_addr st_value; /* Symbol value */ |
| b_elf_word st_size; /* Symbol size */ |
| unsigned char st_info; /* Symbol binding and type */ |
| unsigned char st_other; /* Visibility and other data */ |
| b_elf_half st_shndx; /* Symbol section index */ |
| } b_elf_sym; /* Elf_Sym. */ |
| |
| #else /* BACKTRACE_ELF_SIZE != 32 */ |
| |
| typedef struct |
| { |
| b_elf_word st_name; /* Symbol name, index in string tbl */ |
| unsigned char st_info; /* Symbol binding and type */ |
| unsigned char st_other; /* Visibility and other data */ |
| b_elf_half st_shndx; /* Symbol section index */ |
| b_elf_addr st_value; /* Symbol value */ |
| b_elf_xword st_size; /* Symbol size */ |
| } b_elf_sym; /* Elf_Sym. */ |
| |
| #endif /* BACKTRACE_ELF_SIZE != 32 */ |
| |
| #define STT_OBJECT 1 |
| #define STT_FUNC 2 |
| |
| typedef struct |
| { |
| uint32_t namesz; |
| uint32_t descsz; |
| uint32_t type; |
| char name[1]; |
| } b_elf_note; |
| |
| #define NT_GNU_BUILD_ID 3 |
| |
| #if BACKTRACE_ELF_SIZE == 32 |
| |
| typedef struct |
| { |
| b_elf_word ch_type; /* Compresstion algorithm */ |
| b_elf_word ch_size; /* Uncompressed size */ |
| b_elf_word ch_addralign; /* Alignment for uncompressed data */ |
| } b_elf_chdr; /* Elf_Chdr */ |
| |
| #else /* BACKTRACE_ELF_SIZE != 32 */ |
| |
| typedef struct |
| { |
| b_elf_word ch_type; /* Compression algorithm */ |
| b_elf_word ch_reserved; /* Reserved */ |
| b_elf_xword ch_size; /* Uncompressed size */ |
| b_elf_xword ch_addralign; /* Alignment for uncompressed data */ |
| } b_elf_chdr; /* Elf_Chdr */ |
| |
| #endif /* BACKTRACE_ELF_SIZE != 32 */ |
| |
| #define ELFCOMPRESS_ZLIB 1 |
| #define ELFCOMPRESS_ZSTD 2 |
| |
| /* Names of sections, indexed by enum dwarf_section in internal.h. */ |
| |
| static const char * const dwarf_section_names[DEBUG_MAX] = |
| { |
| ".debug_info", |
| ".debug_line", |
| ".debug_abbrev", |
| ".debug_ranges", |
| ".debug_str", |
| ".debug_addr", |
| ".debug_str_offsets", |
| ".debug_line_str", |
| ".debug_rnglists" |
| }; |
| |
| /* Information we gather for the sections we care about. */ |
| |
| struct debug_section_info |
| { |
| /* Section file offset. */ |
| off_t offset; |
| /* Section size. */ |
| size_t size; |
| /* Section contents, after read from file. */ |
| const unsigned char *data; |
| /* Whether the SHF_COMPRESSED flag is set for the section. */ |
| int compressed; |
| }; |
| |
| /* Information we keep for an ELF symbol. */ |
| |
| struct elf_symbol |
| { |
| /* The name of the symbol. */ |
| const char *name; |
| /* The address of the symbol. */ |
| uintptr_t address; |
| /* The size of the symbol. */ |
| size_t size; |
| }; |
| |
| /* Information to pass to elf_syminfo. */ |
| |
| struct elf_syminfo_data |
| { |
| /* Symbols for the next module. */ |
| struct elf_syminfo_data *next; |
| /* The ELF symbols, sorted by address. */ |
| struct elf_symbol *symbols; |
| /* The number of symbols. */ |
| size_t count; |
| }; |
| |
| /* A view that works for either a file or memory. */ |
| |
| struct elf_view |
| { |
| struct backtrace_view view; |
| int release; /* If non-zero, must call backtrace_release_view. */ |
| }; |
| |
| /* Information about PowerPC64 ELFv1 .opd section. */ |
| |
| struct elf_ppc64_opd_data |
| { |
| /* Address of the .opd section. */ |
| b_elf_addr addr; |
| /* Section data. */ |
| const char *data; |
| /* Size of the .opd section. */ |
| size_t size; |
| /* Corresponding section view. */ |
| struct elf_view view; |
| }; |
| |
| /* Create a view of SIZE bytes from DESCRIPTOR/MEMORY at OFFSET. */ |
| |
| static int |
| elf_get_view (struct backtrace_state *state, int descriptor, |
| const unsigned char *memory, size_t memory_size, off_t offset, |
| uint64_t size, backtrace_error_callback error_callback, |
| void *data, struct elf_view *view) |
| { |
| if (memory == NULL) |
| { |
| view->release = 1; |
| return backtrace_get_view (state, descriptor, offset, size, |
| error_callback, data, &view->view); |
| } |
| else |
| { |
| if ((uint64_t) offset + size > (uint64_t) memory_size) |
| { |
| error_callback (data, "out of range for in-memory file", 0); |
| return 0; |
| } |
| view->view.data = (const void *) (memory + offset); |
| view->view.base = NULL; |
| view->view.len = size; |
| view->release = 0; |
| return 1; |
| } |
| } |
| |
| /* Release a view read by elf_get_view. */ |
| |
| static void |
| elf_release_view (struct backtrace_state *state, struct elf_view *view, |
| backtrace_error_callback error_callback, void *data) |
| { |
| if (view->release) |
| backtrace_release_view (state, &view->view, error_callback, data); |
| } |
| |
| /* Compute the CRC-32 of BUF/LEN. This uses the CRC used for |
| .gnu_debuglink files. */ |
| |
| static uint32_t |
| elf_crc32 (uint32_t crc, const unsigned char *buf, size_t len) |
| { |
| static const uint32_t crc32_table[256] = |
| { |
| 0x00000000, 0x77073096, 0xee0e612c, 0x990951ba, 0x076dc419, |
| 0x706af48f, 0xe963a535, 0x9e6495a3, 0x0edb8832, 0x79dcb8a4, |
| 0xe0d5e91e, 0x97d2d988, 0x09b64c2b, 0x7eb17cbd, 0xe7b82d07, |
| 0x90bf1d91, 0x1db71064, 0x6ab020f2, 0xf3b97148, 0x84be41de, |
| 0x1adad47d, 0x6ddde4eb, 0xf4d4b551, 0x83d385c7, 0x136c9856, |
| 0x646ba8c0, 0xfd62f97a, 0x8a65c9ec, 0x14015c4f, 0x63066cd9, |
| 0xfa0f3d63, 0x8d080df5, 0x3b6e20c8, 0x4c69105e, 0xd56041e4, |
| 0xa2677172, 0x3c03e4d1, 0x4b04d447, 0xd20d85fd, 0xa50ab56b, |
| 0x35b5a8fa, 0x42b2986c, 0xdbbbc9d6, 0xacbcf940, 0x32d86ce3, |
| 0x45df5c75, 0xdcd60dcf, 0xabd13d59, 0x26d930ac, 0x51de003a, |
| 0xc8d75180, 0xbfd06116, 0x21b4f4b5, 0x56b3c423, 0xcfba9599, |
| 0xb8bda50f, 0x2802b89e, 0x5f058808, 0xc60cd9b2, 0xb10be924, |
| 0x2f6f7c87, 0x58684c11, 0xc1611dab, 0xb6662d3d, 0x76dc4190, |
| 0x01db7106, 0x98d220bc, 0xefd5102a, 0x71b18589, 0x06b6b51f, |
| 0x9fbfe4a5, 0xe8b8d433, 0x7807c9a2, 0x0f00f934, 0x9609a88e, |
| 0xe10e9818, 0x7f6a0dbb, 0x086d3d2d, 0x91646c97, 0xe6635c01, |
| 0x6b6b51f4, 0x1c6c6162, 0x856530d8, 0xf262004e, 0x6c0695ed, |
| 0x1b01a57b, 0x8208f4c1, 0xf50fc457, 0x65b0d9c6, 0x12b7e950, |
| 0x8bbeb8ea, 0xfcb9887c, 0x62dd1ddf, 0x15da2d49, 0x8cd37cf3, |
| 0xfbd44c65, 0x4db26158, 0x3ab551ce, 0xa3bc0074, 0xd4bb30e2, |
| 0x4adfa541, 0x3dd895d7, 0xa4d1c46d, 0xd3d6f4fb, 0x4369e96a, |
| 0x346ed9fc, 0xad678846, 0xda60b8d0, 0x44042d73, 0x33031de5, |
| 0xaa0a4c5f, 0xdd0d7cc9, 0x5005713c, 0x270241aa, 0xbe0b1010, |
| 0xc90c2086, 0x5768b525, 0x206f85b3, 0xb966d409, 0xce61e49f, |
| 0x5edef90e, 0x29d9c998, 0xb0d09822, 0xc7d7a8b4, 0x59b33d17, |
| 0x2eb40d81, 0xb7bd5c3b, 0xc0ba6cad, 0xedb88320, 0x9abfb3b6, |
| 0x03b6e20c, 0x74b1d29a, 0xead54739, 0x9dd277af, 0x04db2615, |
| 0x73dc1683, 0xe3630b12, 0x94643b84, 0x0d6d6a3e, 0x7a6a5aa8, |
| 0xe40ecf0b, 0x9309ff9d, 0x0a00ae27, 0x7d079eb1, 0xf00f9344, |
| 0x8708a3d2, 0x1e01f268, 0x6906c2fe, 0xf762575d, 0x806567cb, |
| 0x196c3671, 0x6e6b06e7, 0xfed41b76, 0x89d32be0, 0x10da7a5a, |
| 0x67dd4acc, 0xf9b9df6f, 0x8ebeeff9, 0x17b7be43, 0x60b08ed5, |
| 0xd6d6a3e8, 0xa1d1937e, 0x38d8c2c4, 0x4fdff252, 0xd1bb67f1, |
| 0xa6bc5767, 0x3fb506dd, 0x48b2364b, 0xd80d2bda, 0xaf0a1b4c, |
| 0x36034af6, 0x41047a60, 0xdf60efc3, 0xa867df55, 0x316e8eef, |
| 0x4669be79, 0xcb61b38c, 0xbc66831a, 0x256fd2a0, 0x5268e236, |
| 0xcc0c7795, 0xbb0b4703, 0x220216b9, 0x5505262f, 0xc5ba3bbe, |
| 0xb2bd0b28, 0x2bb45a92, 0x5cb36a04, 0xc2d7ffa7, 0xb5d0cf31, |
| 0x2cd99e8b, 0x5bdeae1d, 0x9b64c2b0, 0xec63f226, 0x756aa39c, |
| 0x026d930a, 0x9c0906a9, 0xeb0e363f, 0x72076785, 0x05005713, |
| 0x95bf4a82, 0xe2b87a14, 0x7bb12bae, 0x0cb61b38, 0x92d28e9b, |
| 0xe5d5be0d, 0x7cdcefb7, 0x0bdbdf21, 0x86d3d2d4, 0xf1d4e242, |
| 0x68ddb3f8, 0x1fda836e, 0x81be16cd, 0xf6b9265b, 0x6fb077e1, |
| 0x18b74777, 0x88085ae6, 0xff0f6a70, 0x66063bca, 0x11010b5c, |
| 0x8f659eff, 0xf862ae69, 0x616bffd3, 0x166ccf45, 0xa00ae278, |
| 0xd70dd2ee, 0x4e048354, 0x3903b3c2, 0xa7672661, 0xd06016f7, |
| 0x4969474d, 0x3e6e77db, 0xaed16a4a, 0xd9d65adc, 0x40df0b66, |
| 0x37d83bf0, 0xa9bcae53, 0xdebb9ec5, 0x47b2cf7f, 0x30b5ffe9, |
| 0xbdbdf21c, 0xcabac28a, 0x53b39330, 0x24b4a3a6, 0xbad03605, |
| 0xcdd70693, 0x54de5729, 0x23d967bf, 0xb3667a2e, 0xc4614ab8, |
| 0x5d681b02, 0x2a6f2b94, 0xb40bbe37, 0xc30c8ea1, 0x5a05df1b, |
| 0x2d02ef8d |
| }; |
| const unsigned char *end; |
| |
| crc = ~crc; |
| for (end = buf + len; buf < end; ++ buf) |
| crc = crc32_table[(crc ^ *buf) & 0xff] ^ (crc >> 8); |
| return ~crc; |
| } |
| |
| /* Return the CRC-32 of the entire file open at DESCRIPTOR. */ |
| |
| static uint32_t |
| elf_crc32_file (struct backtrace_state *state, int descriptor, |
| backtrace_error_callback error_callback, void *data) |
| { |
| struct stat st; |
| struct backtrace_view file_view; |
| uint32_t ret; |
| |
| if (fstat (descriptor, &st) < 0) |
| { |
| error_callback (data, "fstat", errno); |
| return 0; |
| } |
| |
| if (!backtrace_get_view (state, descriptor, 0, st.st_size, error_callback, |
| data, &file_view)) |
| return 0; |
| |
| ret = elf_crc32 (0, (const unsigned char *) file_view.data, st.st_size); |
| |
| backtrace_release_view (state, &file_view, error_callback, data); |
| |
| return ret; |
| } |
| |
| /* A dummy callback function used when we can't find a symbol |
| table. */ |
| |
| static void |
| elf_nosyms (struct backtrace_state *state ATTRIBUTE_UNUSED, |
| uintptr_t addr ATTRIBUTE_UNUSED, |
| backtrace_syminfo_callback callback ATTRIBUTE_UNUSED, |
| backtrace_error_callback error_callback, void *data) |
| { |
| error_callback (data, "no symbol table in ELF executable", -1); |
| } |
| |
| /* A callback function used when we can't find any debug info. */ |
| |
| static int |
| elf_nodebug (struct backtrace_state *state, uintptr_t pc, |
| backtrace_full_callback callback, |
| backtrace_error_callback error_callback, void *data) |
| { |
| if (state->syminfo_fn != NULL && state->syminfo_fn != elf_nosyms) |
| { |
| struct backtrace_call_full bdata; |
| |
| /* Fetch symbol information so that we can least get the |
| function name. */ |
| |
| bdata.full_callback = callback; |
| bdata.full_error_callback = error_callback; |
| bdata.full_data = data; |
| bdata.ret = 0; |
| state->syminfo_fn (state, pc, backtrace_syminfo_to_full_callback, |
| backtrace_syminfo_to_full_error_callback, &bdata); |
| return bdata.ret; |
| } |
| |
| error_callback (data, "no debug info in ELF executable", -1); |
| return 0; |
| } |
| |
| /* Compare struct elf_symbol for qsort. */ |
| |
| static int |
| elf_symbol_compare (const void *v1, const void *v2) |
| { |
| const struct elf_symbol *e1 = (const struct elf_symbol *) v1; |
| const struct elf_symbol *e2 = (const struct elf_symbol *) v2; |
| |
| if (e1->address < e2->address) |
| return -1; |
| else if (e1->address > e2->address) |
| return 1; |
| else |
| return 0; |
| } |
| |
| /* Compare an ADDR against an elf_symbol for bsearch. We allocate one |
| extra entry in the array so that this can look safely at the next |
| entry. */ |
| |
| static int |
| elf_symbol_search (const void *vkey, const void *ventry) |
| { |
| const uintptr_t *key = (const uintptr_t *) vkey; |
| const struct elf_symbol *entry = (const struct elf_symbol *) ventry; |
| uintptr_t addr; |
| |
| addr = *key; |
| if (addr < entry->address) |
| return -1; |
| else if (addr >= entry->address + entry->size) |
| return 1; |
| else |
| return 0; |
| } |
| |
| /* Initialize the symbol table info for elf_syminfo. */ |
| |
| static int |
| elf_initialize_syminfo (struct backtrace_state *state, |
| uintptr_t base_address, |
| const unsigned char *symtab_data, size_t symtab_size, |
| const unsigned char *strtab, size_t strtab_size, |
| backtrace_error_callback error_callback, |
| void *data, struct elf_syminfo_data *sdata, |
| struct elf_ppc64_opd_data *opd) |
| { |
| size_t sym_count; |
| const b_elf_sym *sym; |
| size_t elf_symbol_count; |
| size_t elf_symbol_size; |
| struct elf_symbol *elf_symbols; |
| size_t i; |
| unsigned int j; |
| |
| sym_count = symtab_size / sizeof (b_elf_sym); |
| |
| /* We only care about function symbols. Count them. */ |
| sym = (const b_elf_sym *) symtab_data; |
| elf_symbol_count = 0; |
| for (i = 0; i < sym_count; ++i, ++sym) |
| { |
| int info; |
| |
| info = sym->st_info & 0xf; |
| if ((info == STT_FUNC || info == STT_OBJECT) |
| && sym->st_shndx != SHN_UNDEF) |
| ++elf_symbol_count; |
| } |
| |
| elf_symbol_size = elf_symbol_count * sizeof (struct elf_symbol); |
| elf_symbols = ((struct elf_symbol *) |
| backtrace_alloc (state, elf_symbol_size, error_callback, |
| data)); |
| if (elf_symbols == NULL) |
| return 0; |
| |
| sym = (const b_elf_sym *) symtab_data; |
| j = 0; |
| for (i = 0; i < sym_count; ++i, ++sym) |
| { |
| int info; |
| |
| info = sym->st_info & 0xf; |
| if (info != STT_FUNC && info != STT_OBJECT) |
| continue; |
| if (sym->st_shndx == SHN_UNDEF) |
| continue; |
| if (sym->st_name >= strtab_size) |
| { |
| error_callback (data, "symbol string index out of range", 0); |
| backtrace_free (state, elf_symbols, elf_symbol_size, error_callback, |
| data); |
| return 0; |
| } |
| elf_symbols[j].name = (const char *) strtab + sym->st_name; |
| /* Special case PowerPC64 ELFv1 symbols in .opd section, if the symbol |
| is a function descriptor, read the actual code address from the |
| descriptor. */ |
| if (opd |
| && sym->st_value >= opd->addr |
| && sym->st_value < opd->addr + opd->size) |
| elf_symbols[j].address |
| = *(const b_elf_addr *) (opd->data + (sym->st_value - opd->addr)); |
| else |
| elf_symbols[j].address = sym->st_value; |
| elf_symbols[j].address += base_address; |
| elf_symbols[j].size = sym->st_size; |
| ++j; |
| } |
| |
| backtrace_qsort (elf_symbols, elf_symbol_count, sizeof (struct elf_symbol), |
| elf_symbol_compare); |
| |
| sdata->next = NULL; |
| sdata->symbols = elf_symbols; |
| sdata->count = elf_symbol_count; |
| |
| return 1; |
| } |
| |
| /* Add EDATA to the list in STATE. */ |
| |
| static void |
| elf_add_syminfo_data (struct backtrace_state *state, |
| struct elf_syminfo_data *edata) |
| { |
| if (!state->threaded) |
| { |
| struct elf_syminfo_data **pp; |
| |
| for (pp = (struct elf_syminfo_data **) (void *) &state->syminfo_data; |
| *pp != NULL; |
| pp = &(*pp)->next) |
| ; |
| *pp = edata; |
| } |
| else |
| { |
| while (1) |
| { |
| struct elf_syminfo_data **pp; |
| |
| pp = (struct elf_syminfo_data **) (void *) &state->syminfo_data; |
| |
| while (1) |
| { |
| struct elf_syminfo_data *p; |
| |
| p = backtrace_atomic_load_pointer (pp); |
| |
| if (p == NULL) |
| break; |
| |
| pp = &p->next; |
| } |
| |
| if (__sync_bool_compare_and_swap (pp, NULL, edata)) |
| break; |
| } |
| } |
| } |
| |
| /* Return the symbol name and value for an ADDR. */ |
| |
| static void |
| elf_syminfo (struct backtrace_state *state, uintptr_t addr, |
| backtrace_syminfo_callback callback, |
| backtrace_error_callback error_callback ATTRIBUTE_UNUSED, |
| void *data) |
| { |
| struct elf_syminfo_data *edata; |
| struct elf_symbol *sym = NULL; |
| |
| if (!state->threaded) |
| { |
| for (edata = (struct elf_syminfo_data *) state->syminfo_data; |
| edata != NULL; |
| edata = edata->next) |
| { |
| sym = ((struct elf_symbol *) |
| bsearch (&addr, edata->symbols, edata->count, |
| sizeof (struct elf_symbol), elf_symbol_search)); |
| if (sym != NULL) |
| break; |
| } |
| } |
| else |
| { |
| struct elf_syminfo_data **pp; |
| |
| pp = (struct elf_syminfo_data **) (void *) &state->syminfo_data; |
| while (1) |
| { |
| edata = backtrace_atomic_load_pointer (pp); |
| if (edata == NULL) |
| break; |
| |
| sym = ((struct elf_symbol *) |
| bsearch (&addr, edata->symbols, edata->count, |
| sizeof (struct elf_symbol), elf_symbol_search)); |
| if (sym != NULL) |
| break; |
| |
| pp = &edata->next; |
| } |
| } |
| |
| if (sym == NULL) |
| callback (data, addr, NULL, 0, 0); |
| else |
| callback (data, addr, sym->name, sym->address, sym->size); |
| } |
| |
| /* Return whether FILENAME is a symlink. */ |
| |
| static int |
| elf_is_symlink (const char *filename) |
| { |
| struct stat st; |
| |
| if (lstat (filename, &st) < 0) |
| return 0; |
| return S_ISLNK (st.st_mode); |
| } |
| |
| /* Return the results of reading the symlink FILENAME in a buffer |
| allocated by backtrace_alloc. Return the length of the buffer in |
| *LEN. */ |
| |
| static char * |
| elf_readlink (struct backtrace_state *state, const char *filename, |
| backtrace_error_callback error_callback, void *data, |
| size_t *plen) |
| { |
| size_t len; |
| char *buf; |
| |
| len = 128; |
| while (1) |
| { |
| ssize_t rl; |
| |
| buf = backtrace_alloc (state, len, error_callback, data); |
| if (buf == NULL) |
| return NULL; |
| rl = readlink (filename, buf, len); |
| if (rl < 0) |
| { |
| backtrace_free (state, buf, len, error_callback, data); |
| return NULL; |
| } |
| if ((size_t) rl < len - 1) |
| { |
| buf[rl] = '\0'; |
| *plen = len; |
| return buf; |
| } |
| backtrace_free (state, buf, len, error_callback, data); |
| len *= 2; |
| } |
| } |
| |
| #define SYSTEM_BUILD_ID_DIR "/usr/lib/debug/.build-id/" |
| |
| /* Open a separate debug info file, using the build ID to find it. |
| Returns an open file descriptor, or -1. |
| |
| The GDB manual says that the only place gdb looks for a debug file |
| when the build ID is known is in /usr/lib/debug/.build-id. */ |
| |
| static int |
| elf_open_debugfile_by_buildid (struct backtrace_state *state, |
| const char *buildid_data, size_t buildid_size, |
| backtrace_error_callback error_callback, |
| void *data) |
| { |
| const char * const prefix = SYSTEM_BUILD_ID_DIR; |
| const size_t prefix_len = strlen (prefix); |
| const char * const suffix = ".debug"; |
| const size_t suffix_len = strlen (suffix); |
| size_t len; |
| char *bd_filename; |
| char *t; |
| size_t i; |
| int ret; |
| int does_not_exist; |
| |
| len = prefix_len + buildid_size * 2 + suffix_len + 2; |
| bd_filename = backtrace_alloc (state, len, error_callback, data); |
| if (bd_filename == NULL) |
| return -1; |
| |
| t = bd_filename; |
| memcpy (t, prefix, prefix_len); |
| t += prefix_len; |
| for (i = 0; i < buildid_size; i++) |
| { |
| unsigned char b; |
| unsigned char nib; |
| |
| b = (unsigned char) buildid_data[i]; |
| nib = (b & 0xf0) >> 4; |
| *t++ = nib < 10 ? '0' + nib : 'a' + nib - 10; |
| nib = b & 0x0f; |
| *t++ = nib < 10 ? '0' + nib : 'a' + nib - 10; |
| if (i == 0) |
| *t++ = '/'; |
| } |
| memcpy (t, suffix, suffix_len); |
| t[suffix_len] = '\0'; |
| |
| ret = backtrace_open (bd_filename, error_callback, data, &does_not_exist); |
| |
| backtrace_free (state, bd_filename, len, error_callback, data); |
| |
| /* gdb checks that the debuginfo file has the same build ID note. |
| That seems kind of pointless to me--why would it have the right |
| name but not the right build ID?--so skipping the check. */ |
| |
| return ret; |
| } |
| |
| /* Try to open a file whose name is PREFIX (length PREFIX_LEN) |
| concatenated with PREFIX2 (length PREFIX2_LEN) concatenated with |
| DEBUGLINK_NAME. Returns an open file descriptor, or -1. */ |
| |
| static int |
| elf_try_debugfile (struct backtrace_state *state, const char *prefix, |
| size_t prefix_len, const char *prefix2, size_t prefix2_len, |
| const char *debuglink_name, |
| backtrace_error_callback error_callback, void *data) |
| { |
| size_t debuglink_len; |
| size_t try_len; |
| char *try; |
| int does_not_exist; |
| int ret; |
| |
| debuglink_len = strlen (debuglink_name); |
| try_len = prefix_len + prefix2_len + debuglink_len + 1; |
| try = backtrace_alloc (state, try_len, error_callback, data); |
| if (try == NULL) |
| return -1; |
| |
| memcpy (try, prefix, prefix_len); |
| memcpy (try + prefix_len, prefix2, prefix2_len); |
| memcpy (try + prefix_len + prefix2_len, debuglink_name, debuglink_len); |
| try[prefix_len + prefix2_len + debuglink_len] = '\0'; |
| |
| ret = backtrace_open (try, error_callback, data, &does_not_exist); |
| |
| backtrace_free (state, try, try_len, error_callback, data); |
| |
| return ret; |
| } |
| |
| /* Find a separate debug info file, using the debuglink section data |
| to find it. Returns an open file descriptor, or -1. */ |
| |
| static int |
| elf_find_debugfile_by_debuglink (struct backtrace_state *state, |
| const char *filename, |
| const char *debuglink_name, |
| backtrace_error_callback error_callback, |
| void *data) |
| { |
| int ret; |
| char *alc; |
| size_t alc_len; |
| const char *slash; |
| int ddescriptor; |
| const char *prefix; |
| size_t prefix_len; |
| |
| /* Resolve symlinks in FILENAME. Since FILENAME is fairly likely to |
| be /proc/self/exe, symlinks are common. We don't try to resolve |
| the whole path name, just the base name. */ |
| ret = -1; |
| alc = NULL; |
| alc_len = 0; |
| while (elf_is_symlink (filename)) |
| { |
| char *new_buf; |
| size_t new_len; |
| |
| new_buf = elf_readlink (state, filename, error_callback, data, &new_len); |
| if (new_buf == NULL) |
| break; |
| |
| if (new_buf[0] == '/') |
| filename = new_buf; |
| else |
| { |
| slash = strrchr (filename, '/'); |
| if (slash == NULL) |
| filename = new_buf; |
| else |
| { |
| size_t clen; |
| char *c; |
| |
| slash++; |
| clen = slash - filename + strlen (new_buf) + 1; |
| c = backtrace_alloc (state, clen, error_callback, data); |
| if (c == NULL) |
| goto done; |
| |
| memcpy (c, filename, slash - filename); |
| memcpy (c + (slash - filename), new_buf, strlen (new_buf)); |
| c[slash - filename + strlen (new_buf)] = '\0'; |
| backtrace_free (state, new_buf, new_len, error_callback, data); |
| filename = c; |
| new_buf = c; |
| new_len = clen; |
| } |
| } |
| |
| if (alc != NULL) |
| backtrace_free (state, alc, alc_len, error_callback, data); |
| alc = new_buf; |
| alc_len = new_len; |
| } |
| |
| /* Look for DEBUGLINK_NAME in the same directory as FILENAME. */ |
| |
| slash = strrchr (filename, '/'); |
| if (slash == NULL) |
| { |
| prefix = ""; |
| prefix_len = 0; |
| } |
| else |
| { |
| slash++; |
| prefix = filename; |
| prefix_len = slash - filename; |
| } |
| |
| ddescriptor = elf_try_debugfile (state, prefix, prefix_len, "", 0, |
| debuglink_name, error_callback, data); |
| if (ddescriptor >= 0) |
| { |
| ret = ddescriptor; |
| goto done; |
| } |
| |
| /* Look for DEBUGLINK_NAME in a .debug subdirectory of FILENAME. */ |
| |
| ddescriptor = elf_try_debugfile (state, prefix, prefix_len, ".debug/", |
| strlen (".debug/"), debuglink_name, |
| error_callback, data); |
| if (ddescriptor >= 0) |
| { |
| ret = ddescriptor; |
| goto done; |
| } |
| |
| /* Look for DEBUGLINK_NAME in /usr/lib/debug. */ |
| |
| ddescriptor = elf_try_debugfile (state, "/usr/lib/debug/", |
| strlen ("/usr/lib/debug/"), prefix, |
| prefix_len, debuglink_name, |
| error_callback, data); |
| if (ddescriptor >= 0) |
| ret = ddescriptor; |
| |
| done: |
| if (alc != NULL && alc_len > 0) |
| backtrace_free (state, alc, alc_len, error_callback, data); |
| return ret; |
| } |
| |
| /* Open a separate debug info file, using the debuglink section data |
| to find it. Returns an open file descriptor, or -1. */ |
| |
| static int |
| elf_open_debugfile_by_debuglink (struct backtrace_state *state, |
| const char *filename, |
| const char *debuglink_name, |
| uint32_t debuglink_crc, |
| backtrace_error_callback error_callback, |
| void *data) |
| { |
| int ddescriptor; |
| |
| ddescriptor = elf_find_debugfile_by_debuglink (state, filename, |
| debuglink_name, |
| error_callback, data); |
| if (ddescriptor < 0) |
| return -1; |
| |
| if (debuglink_crc != 0) |
| { |
| uint32_t got_crc; |
| |
| got_crc = elf_crc32_file (state, ddescriptor, error_callback, data); |
| if (got_crc != debuglink_crc) |
| { |
| backtrace_close (ddescriptor, error_callback, data); |
| return -1; |
| } |
| } |
| |
| return ddescriptor; |
| } |
| |
| /* A function useful for setting a breakpoint for an inflation failure |
| when this code is compiled with -g. */ |
| |
| static void |
| elf_uncompress_failed(void) |
| { |
| } |
| |
| /* *PVAL is the current value being read from the stream, and *PBITS |
| is the number of valid bits. Ensure that *PVAL holds at least 15 |
| bits by reading additional bits from *PPIN, up to PINEND, as |
| needed. Updates *PPIN, *PVAL and *PBITS. Returns 1 on success, 0 |
| on error. */ |
| |
| static int |
| elf_fetch_bits (const unsigned char **ppin, const unsigned char *pinend, |
| uint64_t *pval, unsigned int *pbits) |
| { |
| unsigned int bits; |
| const unsigned char *pin; |
| uint64_t val; |
| uint32_t next; |
| |
| bits = *pbits; |
| if (bits >= 15) |
| return 1; |
| pin = *ppin; |
| val = *pval; |
| |
| if (unlikely (pinend - pin < 4)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| #if defined(__BYTE_ORDER__) && defined(__ORDER_LITTLE_ENDIAN__) \ |
| && defined(__ORDER_BIG_ENDIAN__) \ |
| && (__BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ \ |
| || __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__) |
| /* We've ensured that PIN is aligned. */ |
| next = *(const uint32_t *)pin; |
| |
| #if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ |
| next = __builtin_bswap32 (next); |
| #endif |
| #else |
| next = pin[0] | (pin[1] << 8) | (pin[2] << 16) | (pin[3] << 24); |
| #endif |
| |
| val |= (uint64_t)next << bits; |
| bits += 32; |
| pin += 4; |
| |
| /* We will need the next four bytes soon. */ |
| __builtin_prefetch (pin, 0, 0); |
| |
| *ppin = pin; |
| *pval = val; |
| *pbits = bits; |
| return 1; |
| } |
| |
| /* This is like elf_fetch_bits, but it fetchs the bits backward, and ensures at |
| least 16 bits. This is for zstd. */ |
| |
| static int |
| elf_fetch_bits_backward (const unsigned char **ppin, |
| const unsigned char *pinend, |
| uint64_t *pval, unsigned int *pbits) |
| { |
| unsigned int bits; |
| const unsigned char *pin; |
| uint64_t val; |
| uint32_t next; |
| |
| bits = *pbits; |
| if (bits >= 16) |
| return 1; |
| pin = *ppin; |
| val = *pval; |
| |
| if (unlikely (pin <= pinend)) |
| { |
| if (bits == 0) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| return 1; |
| } |
| |
| pin -= 4; |
| |
| #if defined(__BYTE_ORDER__) && defined(__ORDER_LITTLE_ENDIAN__) \ |
| && defined(__ORDER_BIG_ENDIAN__) \ |
| && (__BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ \ |
| || __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__) |
| /* We've ensured that PIN is aligned. */ |
| next = *(const uint32_t *)pin; |
| |
| #if __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ |
| next = __builtin_bswap32 (next); |
| #endif |
| #else |
| next = pin[0] | (pin[1] << 8) | (pin[2] << 16) | (pin[3] << 24); |
| #endif |
| |
| val <<= 32; |
| val |= next; |
| bits += 32; |
| |
| if (unlikely (pin < pinend)) |
| { |
| val >>= (pinend - pin) * 8; |
| bits -= (pinend - pin) * 8; |
| } |
| |
| *ppin = pin; |
| *pval = val; |
| *pbits = bits; |
| return 1; |
| } |
| |
| /* Initialize backward fetching when the bitstream starts with a 1 bit in the |
| last byte in memory (which is the first one that we read). This is used by |
| zstd decompression. Returns 1 on success, 0 on error. */ |
| |
| static int |
| elf_fetch_backward_init (const unsigned char **ppin, |
| const unsigned char *pinend, |
| uint64_t *pval, unsigned int *pbits) |
| { |
| const unsigned char *pin; |
| unsigned int stream_start; |
| uint64_t val; |
| unsigned int bits; |
| |
| pin = *ppin; |
| stream_start = (unsigned int)*pin; |
| if (unlikely (stream_start == 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| val = 0; |
| bits = 0; |
| |
| /* Align to a 32-bit boundary. */ |
| while ((((uintptr_t)pin) & 3) != 0) |
| { |
| val <<= 8; |
| val |= (uint64_t)*pin; |
| bits += 8; |
| --pin; |
| } |
| |
| val <<= 8; |
| val |= (uint64_t)*pin; |
| bits += 8; |
| |
| *ppin = pin; |
| *pval = val; |
| *pbits = bits; |
| if (!elf_fetch_bits_backward (ppin, pinend, pval, pbits)) |
| return 0; |
| |
| *pbits -= __builtin_clz (stream_start) - (sizeof (unsigned int) - 1) * 8 + 1; |
| |
| if (!elf_fetch_bits_backward (ppin, pinend, pval, pbits)) |
| return 0; |
| |
| return 1; |
| } |
| |
| /* Huffman code tables, like the rest of the zlib format, are defined |
| by RFC 1951. We store a Huffman code table as a series of tables |
| stored sequentially in memory. Each entry in a table is 16 bits. |
| The first, main, table has 256 entries. It is followed by a set of |
| secondary tables of length 2 to 128 entries. The maximum length of |
| a code sequence in the deflate format is 15 bits, so that is all we |
| need. Each secondary table has an index, which is the offset of |
| the table in the overall memory storage. |
| |
| The deflate format says that all codes of a given bit length are |
| lexicographically consecutive. Perhaps we could have 130 values |
| that require a 15-bit code, perhaps requiring three secondary |
| tables of size 128. I don't know if this is actually possible, but |
| it suggests that the maximum size required for secondary tables is |
| 3 * 128 + 3 * 64 ... == 768. The zlib enough program reports 660 |
| as the maximum. We permit 768, since in addition to the 256 for |
| the primary table, with two bytes per entry, and with the two |
| tables we need, that gives us a page. |
| |
| A single table entry needs to store a value or (for the main table |
| only) the index and size of a secondary table. Values range from 0 |
| to 285, inclusive. Secondary table indexes, per above, range from |
| 0 to 510. For a value we need to store the number of bits we need |
| to determine that value (one value may appear multiple times in the |
| table), which is 1 to 8. For a secondary table we need to store |
| the number of bits used to index into the table, which is 1 to 7. |
| And of course we need 1 bit to decide whether we have a value or a |
| secondary table index. So each entry needs 9 bits for value/table |
| index, 3 bits for size, 1 bit what it is. For simplicity we use 16 |
| bits per entry. */ |
| |
| /* Number of entries we allocate to for one code table. We get a page |
| for the two code tables we need. */ |
| |
| #define ZLIB_HUFFMAN_TABLE_SIZE (1024) |
| |
| /* Bit masks and shifts for the values in the table. */ |
| |
| #define ZLIB_HUFFMAN_VALUE_MASK 0x01ff |
| #define ZLIB_HUFFMAN_BITS_SHIFT 9 |
| #define ZLIB_HUFFMAN_BITS_MASK 0x7 |
| #define ZLIB_HUFFMAN_SECONDARY_SHIFT 12 |
| |
| /* For working memory while inflating we need two code tables, we need |
| an array of code lengths (max value 15, so we use unsigned char), |
| and an array of unsigned shorts used while building a table. The |
| latter two arrays must be large enough to hold the maximum number |
| of code lengths, which RFC 1951 defines as 286 + 30. */ |
| |
| #define ZLIB_TABLE_SIZE \ |
| (2 * ZLIB_HUFFMAN_TABLE_SIZE * sizeof (uint16_t) \ |
| + (286 + 30) * sizeof (uint16_t) \ |
| + (286 + 30) * sizeof (unsigned char)) |
| |
| #define ZLIB_TABLE_CODELEN_OFFSET \ |
| (2 * ZLIB_HUFFMAN_TABLE_SIZE * sizeof (uint16_t) \ |
| + (286 + 30) * sizeof (uint16_t)) |
| |
| #define ZLIB_TABLE_WORK_OFFSET \ |
| (2 * ZLIB_HUFFMAN_TABLE_SIZE * sizeof (uint16_t)) |
| |
| #ifdef BACKTRACE_GENERATE_FIXED_HUFFMAN_TABLE |
| |
| /* Used by the main function that generates the fixed table to learn |
| the table size. */ |
| static size_t final_next_secondary; |
| |
| #endif |
| |
| /* Build a Huffman code table from an array of lengths in CODES of |
| length CODES_LEN. The table is stored into *TABLE. ZDEBUG_TABLE |
| is the same as for elf_zlib_inflate, used to find some work space. |
| Returns 1 on success, 0 on error. */ |
| |
| static int |
| elf_zlib_inflate_table (unsigned char *codes, size_t codes_len, |
| uint16_t *zdebug_table, uint16_t *table) |
| { |
| uint16_t count[16]; |
| uint16_t start[16]; |
| uint16_t prev[16]; |
| uint16_t firstcode[7]; |
| uint16_t *next; |
| size_t i; |
| size_t j; |
| unsigned int code; |
| size_t next_secondary; |
| |
| /* Count the number of code of each length. Set NEXT[val] to be the |
| next value after VAL with the same bit length. */ |
| |
| next = (uint16_t *) (((unsigned char *) zdebug_table) |
| + ZLIB_TABLE_WORK_OFFSET); |
| |
| memset (&count[0], 0, 16 * sizeof (uint16_t)); |
| for (i = 0; i < codes_len; ++i) |
| { |
| if (unlikely (codes[i] >= 16)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (count[codes[i]] == 0) |
| { |
| start[codes[i]] = i; |
| prev[codes[i]] = i; |
| } |
| else |
| { |
| next[prev[codes[i]]] = i; |
| prev[codes[i]] = i; |
| } |
| |
| ++count[codes[i]]; |
| } |
| |
| /* For each length, fill in the table for the codes of that |
| length. */ |
| |
| memset (table, 0, ZLIB_HUFFMAN_TABLE_SIZE * sizeof (uint16_t)); |
| |
| /* Handle the values that do not require a secondary table. */ |
| |
| code = 0; |
| for (j = 1; j <= 8; ++j) |
| { |
| unsigned int jcnt; |
| unsigned int val; |
| |
| jcnt = count[j]; |
| if (jcnt == 0) |
| continue; |
| |
| if (unlikely (jcnt > (1U << j))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* There are JCNT values that have this length, the values |
| starting from START[j] continuing through NEXT[VAL]. Those |
| values are assigned consecutive values starting at CODE. */ |
| |
| val = start[j]; |
| for (i = 0; i < jcnt; ++i) |
| { |
| uint16_t tval; |
| size_t ind; |
| unsigned int incr; |
| |
| /* In the compressed bit stream, the value VAL is encoded as |
| J bits with the value C. */ |
| |
| if (unlikely ((val & ~ZLIB_HUFFMAN_VALUE_MASK) != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| tval = val | ((j - 1) << ZLIB_HUFFMAN_BITS_SHIFT); |
| |
| /* The table lookup uses 8 bits. If J is less than 8, we |
| don't know what the other bits will be. We need to fill |
| in all possibilities in the table. Since the Huffman |
| code is unambiguous, those entries can't be used for any |
| other code. */ |
| |
| for (ind = code; ind < 0x100; ind += 1 << j) |
| { |
| if (unlikely (table[ind] != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| table[ind] = tval; |
| } |
| |
| /* Advance to the next value with this length. */ |
| if (i + 1 < jcnt) |
| val = next[val]; |
| |
| /* The Huffman codes are stored in the bitstream with the |
| most significant bit first, as is required to make them |
| unambiguous. The effect is that when we read them from |
| the bitstream we see the bit sequence in reverse order: |
| the most significant bit of the Huffman code is the least |
| significant bit of the value we read from the bitstream. |
| That means that to make our table lookups work, we need |
| to reverse the bits of CODE. Since reversing bits is |
| tedious and in general requires using a table, we instead |
| increment CODE in reverse order. That is, if the number |
| of bits we are currently using, here named J, is 3, we |
| count as 000, 100, 010, 110, 001, 101, 011, 111, which is |
| to say the numbers from 0 to 7 but with the bits |
| reversed. Going to more bits, aka incrementing J, |
| effectively just adds more zero bits as the beginning, |
| and as such does not change the numeric value of CODE. |
| |
| To increment CODE of length J in reverse order, find the |
| most significant zero bit and set it to one while |
| clearing all higher bits. In other words, add 1 modulo |
| 2^J, only reversed. */ |
| |
| incr = 1U << (j - 1); |
| while ((code & incr) != 0) |
| incr >>= 1; |
| if (incr == 0) |
| code = 0; |
| else |
| { |
| code &= incr - 1; |
| code += incr; |
| } |
| } |
| } |
| |
| /* Handle the values that require a secondary table. */ |
| |
| /* Set FIRSTCODE, the number at which the codes start, for each |
| length. */ |
| |
| for (j = 9; j < 16; j++) |
| { |
| unsigned int jcnt; |
| unsigned int k; |
| |
| jcnt = count[j]; |
| if (jcnt == 0) |
| continue; |
| |
| /* There are JCNT values that have this length, the values |
| starting from START[j]. Those values are assigned |
| consecutive values starting at CODE. */ |
| |
| firstcode[j - 9] = code; |
| |
| /* Reverse add JCNT to CODE modulo 2^J. */ |
| for (k = 0; k < j; ++k) |
| { |
| if ((jcnt & (1U << k)) != 0) |
| { |
| unsigned int m; |
| unsigned int bit; |
| |
| bit = 1U << (j - k - 1); |
| for (m = 0; m < j - k; ++m, bit >>= 1) |
| { |
| if ((code & bit) == 0) |
| { |
| code += bit; |
| break; |
| } |
| code &= ~bit; |
| } |
| jcnt &= ~(1U << k); |
| } |
| } |
| if (unlikely (jcnt != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| |
| /* For J from 9 to 15, inclusive, we store COUNT[J] consecutive |
| values starting at START[J] with consecutive codes starting at |
| FIRSTCODE[J - 9]. In the primary table we need to point to the |
| secondary table, and the secondary table will be indexed by J - 9 |
| bits. We count down from 15 so that we install the larger |
| secondary tables first, as the smaller ones may be embedded in |
| the larger ones. */ |
| |
| next_secondary = 0; /* Index of next secondary table (after primary). */ |
| for (j = 15; j >= 9; j--) |
| { |
| unsigned int jcnt; |
| unsigned int val; |
| size_t primary; /* Current primary index. */ |
| size_t secondary; /* Offset to current secondary table. */ |
| size_t secondary_bits; /* Bit size of current secondary table. */ |
| |
| jcnt = count[j]; |
| if (jcnt == 0) |
| continue; |
| |
| val = start[j]; |
| code = firstcode[j - 9]; |
| primary = 0x100; |
| secondary = 0; |
| secondary_bits = 0; |
| for (i = 0; i < jcnt; ++i) |
| { |
| uint16_t tval; |
| size_t ind; |
| unsigned int incr; |
| |
| if ((code & 0xff) != primary) |
| { |
| uint16_t tprimary; |
| |
| /* Fill in a new primary table entry. */ |
| |
| primary = code & 0xff; |
| |
| tprimary = table[primary]; |
| if (tprimary == 0) |
| { |
| /* Start a new secondary table. */ |
| |
| if (unlikely ((next_secondary & ZLIB_HUFFMAN_VALUE_MASK) |
| != next_secondary)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| secondary = next_secondary; |
| secondary_bits = j - 8; |
| next_secondary += 1 << secondary_bits; |
| table[primary] = (secondary |
| + ((j - 8) << ZLIB_HUFFMAN_BITS_SHIFT) |
| + (1U << ZLIB_HUFFMAN_SECONDARY_SHIFT)); |
| } |
| else |
| { |
| /* There is an existing entry. It had better be a |
| secondary table with enough bits. */ |
| if (unlikely ((tprimary |
| & (1U << ZLIB_HUFFMAN_SECONDARY_SHIFT)) |
| == 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| secondary = tprimary & ZLIB_HUFFMAN_VALUE_MASK; |
| secondary_bits = ((tprimary >> ZLIB_HUFFMAN_BITS_SHIFT) |
| & ZLIB_HUFFMAN_BITS_MASK); |
| if (unlikely (secondary_bits < j - 8)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| } |
| |
| /* Fill in secondary table entries. */ |
| |
| tval = val | ((j - 8) << ZLIB_HUFFMAN_BITS_SHIFT); |
| |
| for (ind = code >> 8; |
| ind < (1U << secondary_bits); |
| ind += 1U << (j - 8)) |
| { |
| if (unlikely (table[secondary + 0x100 + ind] != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| table[secondary + 0x100 + ind] = tval; |
| } |
| |
| if (i + 1 < jcnt) |
| val = next[val]; |
| |
| incr = 1U << (j - 1); |
| while ((code & incr) != 0) |
| incr >>= 1; |
| if (incr == 0) |
| code = 0; |
| else |
| { |
| code &= incr - 1; |
| code += incr; |
| } |
| } |
| } |
| |
| #ifdef BACKTRACE_GENERATE_FIXED_HUFFMAN_TABLE |
| final_next_secondary = next_secondary; |
| #endif |
| |
| return 1; |
| } |
| |
| #ifdef BACKTRACE_GENERATE_FIXED_HUFFMAN_TABLE |
| |
| /* Used to generate the fixed Huffman table for block type 1. */ |
| |
| #include <stdio.h> |
| |
| static uint16_t table[ZLIB_TABLE_SIZE]; |
| static unsigned char codes[288]; |
| |
| int |
| main () |
| { |
| size_t i; |
| |
| for (i = 0; i <= 143; ++i) |
| codes[i] = 8; |
| for (i = 144; i <= 255; ++i) |
| codes[i] = 9; |
| for (i = 256; i <= 279; ++i) |
| codes[i] = 7; |
| for (i = 280; i <= 287; ++i) |
| codes[i] = 8; |
| if (!elf_zlib_inflate_table (&codes[0], 288, &table[0], &table[0])) |
| { |
| fprintf (stderr, "elf_zlib_inflate_table failed\n"); |
| exit (EXIT_FAILURE); |
| } |
| |
| printf ("static const uint16_t elf_zlib_default_table[%#zx] =\n", |
| final_next_secondary + 0x100); |
| printf ("{\n"); |
| for (i = 0; i < final_next_secondary + 0x100; i += 8) |
| { |
| size_t j; |
| |
| printf (" "); |
| for (j = i; j < final_next_secondary + 0x100 && j < i + 8; ++j) |
| printf (" %#x,", table[j]); |
| printf ("\n"); |
| } |
| printf ("};\n"); |
| printf ("\n"); |
| |
| for (i = 0; i < 32; ++i) |
| codes[i] = 5; |
| if (!elf_zlib_inflate_table (&codes[0], 32, &table[0], &table[0])) |
| { |
| fprintf (stderr, "elf_zlib_inflate_table failed\n"); |
| exit (EXIT_FAILURE); |
| } |
| |
| printf ("static const uint16_t elf_zlib_default_dist_table[%#zx] =\n", |
| final_next_secondary + 0x100); |
| printf ("{\n"); |
| for (i = 0; i < final_next_secondary + 0x100; i += 8) |
| { |
| size_t j; |
| |
| printf (" "); |
| for (j = i; j < final_next_secondary + 0x100 && j < i + 8; ++j) |
| printf (" %#x,", table[j]); |
| printf ("\n"); |
| } |
| printf ("};\n"); |
| |
| return 0; |
| } |
| |
| #endif |
| |
| /* The fixed tables generated by the #ifdef'ed out main function |
| above. */ |
| |
| static const uint16_t elf_zlib_default_table[0x170] = |
| { |
| 0xd00, 0xe50, 0xe10, 0xf18, 0xd10, 0xe70, 0xe30, 0x1230, |
| 0xd08, 0xe60, 0xe20, 0x1210, 0xe00, 0xe80, 0xe40, 0x1250, |
| 0xd04, 0xe58, 0xe18, 0x1200, 0xd14, 0xe78, 0xe38, 0x1240, |
| 0xd0c, 0xe68, 0xe28, 0x1220, 0xe08, 0xe88, 0xe48, 0x1260, |
| 0xd02, 0xe54, 0xe14, 0xf1c, 0xd12, 0xe74, 0xe34, 0x1238, |
| 0xd0a, 0xe64, 0xe24, 0x1218, 0xe04, 0xe84, 0xe44, 0x1258, |
| 0xd06, 0xe5c, 0xe1c, 0x1208, 0xd16, 0xe7c, 0xe3c, 0x1248, |
| 0xd0e, 0xe6c, 0xe2c, 0x1228, 0xe0c, 0xe8c, 0xe4c, 0x1268, |
| 0xd01, 0xe52, 0xe12, 0xf1a, 0xd11, 0xe72, 0xe32, 0x1234, |
| 0xd09, 0xe62, 0xe22, 0x1214, 0xe02, 0xe82, 0xe42, 0x1254, |
| 0xd05, 0xe5a, 0xe1a, 0x1204, 0xd15, 0xe7a, 0xe3a, 0x1244, |
| 0xd0d, 0xe6a, 0xe2a, 0x1224, 0xe0a, 0xe8a, 0xe4a, 0x1264, |
| 0xd03, 0xe56, 0xe16, 0xf1e, 0xd13, 0xe76, 0xe36, 0x123c, |
| 0xd0b, 0xe66, 0xe26, 0x121c, 0xe06, 0xe86, 0xe46, 0x125c, |
| 0xd07, 0xe5e, 0xe1e, 0x120c, 0xd17, 0xe7e, 0xe3e, 0x124c, |
| 0xd0f, 0xe6e, 0xe2e, 0x122c, 0xe0e, 0xe8e, 0xe4e, 0x126c, |
| 0xd00, 0xe51, 0xe11, 0xf19, 0xd10, 0xe71, 0xe31, 0x1232, |
| 0xd08, 0xe61, 0xe21, 0x1212, 0xe01, 0xe81, 0xe41, 0x1252, |
| 0xd04, 0xe59, 0xe19, 0x1202, 0xd14, 0xe79, 0xe39, 0x1242, |
| 0xd0c, 0xe69, 0xe29, 0x1222, 0xe09, 0xe89, 0xe49, 0x1262, |
| 0xd02, 0xe55, 0xe15, 0xf1d, 0xd12, 0xe75, 0xe35, 0x123a, |
| 0xd0a, 0xe65, 0xe25, 0x121a, 0xe05, 0xe85, 0xe45, 0x125a, |
| 0xd06, 0xe5d, 0xe1d, 0x120a, 0xd16, 0xe7d, 0xe3d, 0x124a, |
| 0xd0e, 0xe6d, 0xe2d, 0x122a, 0xe0d, 0xe8d, 0xe4d, 0x126a, |
| 0xd01, 0xe53, 0xe13, 0xf1b, 0xd11, 0xe73, 0xe33, 0x1236, |
| 0xd09, 0xe63, 0xe23, 0x1216, 0xe03, 0xe83, 0xe43, 0x1256, |
| 0xd05, 0xe5b, 0xe1b, 0x1206, 0xd15, 0xe7b, 0xe3b, 0x1246, |
| 0xd0d, 0xe6b, 0xe2b, 0x1226, 0xe0b, 0xe8b, 0xe4b, 0x1266, |
| 0xd03, 0xe57, 0xe17, 0xf1f, 0xd13, 0xe77, 0xe37, 0x123e, |
| 0xd0b, 0xe67, 0xe27, 0x121e, 0xe07, 0xe87, 0xe47, 0x125e, |
| 0xd07, 0xe5f, 0xe1f, 0x120e, 0xd17, 0xe7f, 0xe3f, 0x124e, |
| 0xd0f, 0xe6f, 0xe2f, 0x122e, 0xe0f, 0xe8f, 0xe4f, 0x126e, |
| 0x290, 0x291, 0x292, 0x293, 0x294, 0x295, 0x296, 0x297, |
| 0x298, 0x299, 0x29a, 0x29b, 0x29c, 0x29d, 0x29e, 0x29f, |
| 0x2a0, 0x2a1, 0x2a2, 0x2a3, 0x2a4, 0x2a5, 0x2a6, 0x2a7, |
| 0x2a8, 0x2a9, 0x2aa, 0x2ab, 0x2ac, 0x2ad, 0x2ae, 0x2af, |
| 0x2b0, 0x2b1, 0x2b2, 0x2b3, 0x2b4, 0x2b5, 0x2b6, 0x2b7, |
| 0x2b8, 0x2b9, 0x2ba, 0x2bb, 0x2bc, 0x2bd, 0x2be, 0x2bf, |
| 0x2c0, 0x2c1, 0x2c2, 0x2c3, 0x2c4, 0x2c5, 0x2c6, 0x2c7, |
| 0x2c8, 0x2c9, 0x2ca, 0x2cb, 0x2cc, 0x2cd, 0x2ce, 0x2cf, |
| 0x2d0, 0x2d1, 0x2d2, 0x2d3, 0x2d4, 0x2d5, 0x2d6, 0x2d7, |
| 0x2d8, 0x2d9, 0x2da, 0x2db, 0x2dc, 0x2dd, 0x2de, 0x2df, |
| 0x2e0, 0x2e1, 0x2e2, 0x2e3, 0x2e4, 0x2e5, 0x2e6, 0x2e7, |
| 0x2e8, 0x2e9, 0x2ea, 0x2eb, 0x2ec, 0x2ed, 0x2ee, 0x2ef, |
| 0x2f0, 0x2f1, 0x2f2, 0x2f3, 0x2f4, 0x2f5, 0x2f6, 0x2f7, |
| 0x2f8, 0x2f9, 0x2fa, 0x2fb, 0x2fc, 0x2fd, 0x2fe, 0x2ff, |
| }; |
| |
| static const uint16_t elf_zlib_default_dist_table[0x100] = |
| { |
| 0x800, 0x810, 0x808, 0x818, 0x804, 0x814, 0x80c, 0x81c, |
| 0x802, 0x812, 0x80a, 0x81a, 0x806, 0x816, 0x80e, 0x81e, |
| 0x801, 0x811, 0x809, 0x819, 0x805, 0x815, 0x80d, 0x81d, |
| 0x803, 0x813, 0x80b, 0x81b, 0x807, 0x817, 0x80f, 0x81f, |
| 0x800, 0x810, 0x808, 0x818, 0x804, 0x814, 0x80c, 0x81c, |
| 0x802, 0x812, 0x80a, 0x81a, 0x806, 0x816, 0x80e, 0x81e, |
| 0x801, 0x811, 0x809, 0x819, 0x805, 0x815, 0x80d, 0x81d, |
| 0x803, 0x813, 0x80b, 0x81b, 0x807, 0x817, 0x80f, 0x81f, |
| 0x800, 0x810, 0x808, 0x818, 0x804, 0x814, 0x80c, 0x81c, |
| 0x802, 0x812, 0x80a, 0x81a, 0x806, 0x816, 0x80e, 0x81e, |
| 0x801, 0x811, 0x809, 0x819, 0x805, 0x815, 0x80d, 0x81d, |
| 0x803, 0x813, 0x80b, 0x81b, 0x807, 0x817, 0x80f, 0x81f, |
| 0x800, 0x810, 0x808, 0x818, 0x804, 0x814, 0x80c, 0x81c, |
| 0x802, 0x812, 0x80a, 0x81a, 0x806, 0x816, 0x80e, 0x81e, |
| 0x801, 0x811, 0x809, 0x819, 0x805, 0x815, 0x80d, 0x81d, |
| 0x803, 0x813, 0x80b, 0x81b, 0x807, 0x817, 0x80f, 0x81f, |
| 0x800, 0x810, 0x808, 0x818, 0x804, 0x814, 0x80c, 0x81c, |
| 0x802, 0x812, 0x80a, 0x81a, 0x806, 0x816, 0x80e, 0x81e, |
| 0x801, 0x811, 0x809, 0x819, 0x805, 0x815, 0x80d, 0x81d, |
| 0x803, 0x813, 0x80b, 0x81b, 0x807, 0x817, 0x80f, 0x81f, |
| 0x800, 0x810, 0x808, 0x818, 0x804, 0x814, 0x80c, 0x81c, |
| 0x802, 0x812, 0x80a, 0x81a, 0x806, 0x816, 0x80e, 0x81e, |
| 0x801, 0x811, 0x809, 0x819, 0x805, 0x815, 0x80d, 0x81d, |
| 0x803, 0x813, 0x80b, 0x81b, 0x807, 0x817, 0x80f, 0x81f, |
| 0x800, 0x810, 0x808, 0x818, 0x804, 0x814, 0x80c, 0x81c, |
| 0x802, 0x812, 0x80a, 0x81a, 0x806, 0x816, 0x80e, 0x81e, |
| 0x801, 0x811, 0x809, 0x819, 0x805, 0x815, 0x80d, 0x81d, |
| 0x803, 0x813, 0x80b, 0x81b, 0x807, 0x817, 0x80f, 0x81f, |
| 0x800, 0x810, 0x808, 0x818, 0x804, 0x814, 0x80c, 0x81c, |
| 0x802, 0x812, 0x80a, 0x81a, 0x806, 0x816, 0x80e, 0x81e, |
| 0x801, 0x811, 0x809, 0x819, 0x805, 0x815, 0x80d, 0x81d, |
| 0x803, 0x813, 0x80b, 0x81b, 0x807, 0x817, 0x80f, 0x81f, |
| }; |
| |
| /* Inflate a zlib stream from PIN/SIN to POUT/SOUT. Return 1 on |
| success, 0 on some error parsing the stream. */ |
| |
| static int |
| elf_zlib_inflate (const unsigned char *pin, size_t sin, uint16_t *zdebug_table, |
| unsigned char *pout, size_t sout) |
| { |
| unsigned char *porigout; |
| const unsigned char *pinend; |
| unsigned char *poutend; |
| |
| /* We can apparently see multiple zlib streams concatenated |
| together, so keep going as long as there is something to read. |
| The last 4 bytes are the checksum. */ |
| porigout = pout; |
| pinend = pin + sin; |
| poutend = pout + sout; |
| while ((pinend - pin) > 4) |
| { |
| uint64_t val; |
| unsigned int bits; |
| int last; |
| |
| /* Read the two byte zlib header. */ |
| |
| if (unlikely ((pin[0] & 0xf) != 8)) /* 8 is zlib encoding. */ |
| { |
| /* Unknown compression method. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| if (unlikely ((pin[0] >> 4) > 7)) |
| { |
| /* Window size too large. Other than this check, we don't |
| care about the window size. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| if (unlikely ((pin[1] & 0x20) != 0)) |
| { |
| /* Stream expects a predefined dictionary, but we have no |
| dictionary. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| val = (pin[0] << 8) | pin[1]; |
| if (unlikely (val % 31 != 0)) |
| { |
| /* Header check failure. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| pin += 2; |
| |
| /* Align PIN to a 32-bit boundary. */ |
| |
| val = 0; |
| bits = 0; |
| while ((((uintptr_t) pin) & 3) != 0) |
| { |
| val |= (uint64_t)*pin << bits; |
| bits += 8; |
| ++pin; |
| } |
| |
| /* Read blocks until one is marked last. */ |
| |
| last = 0; |
| |
| while (!last) |
| { |
| unsigned int type; |
| const uint16_t *tlit; |
| const uint16_t *tdist; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| last = val & 1; |
| type = (val >> 1) & 3; |
| val >>= 3; |
| bits -= 3; |
| |
| if (unlikely (type == 3)) |
| { |
| /* Invalid block type. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (type == 0) |
| { |
| uint16_t len; |
| uint16_t lenc; |
| |
| /* An uncompressed block. */ |
| |
| /* If we've read ahead more than a byte, back up. */ |
| while (bits >= 8) |
| { |
| --pin; |
| bits -= 8; |
| } |
| |
| val = 0; |
| bits = 0; |
| if (unlikely ((pinend - pin) < 4)) |
| { |
| /* Missing length. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| len = pin[0] | (pin[1] << 8); |
| lenc = pin[2] | (pin[3] << 8); |
| pin += 4; |
| lenc = ~lenc; |
| if (unlikely (len != lenc)) |
| { |
| /* Corrupt data. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| if (unlikely (len > (unsigned int) (pinend - pin) |
| || len > (unsigned int) (poutend - pout))) |
| { |
| /* Not enough space in buffers. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| memcpy (pout, pin, len); |
| pout += len; |
| pin += len; |
| |
| /* Align PIN. */ |
| while ((((uintptr_t) pin) & 3) != 0) |
| { |
| val |= (uint64_t)*pin << bits; |
| bits += 8; |
| ++pin; |
| } |
| |
| /* Go around to read the next block. */ |
| continue; |
| } |
| |
| if (type == 1) |
| { |
| tlit = elf_zlib_default_table; |
| tdist = elf_zlib_default_dist_table; |
| } |
| else |
| { |
| unsigned int nlit; |
| unsigned int ndist; |
| unsigned int nclen; |
| unsigned char codebits[19]; |
| unsigned char *plenbase; |
| unsigned char *plen; |
| unsigned char *plenend; |
| |
| /* Read a Huffman encoding table. The various magic |
| numbers here are from RFC 1951. */ |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| nlit = (val & 0x1f) + 257; |
| val >>= 5; |
| ndist = (val & 0x1f) + 1; |
| val >>= 5; |
| nclen = (val & 0xf) + 4; |
| val >>= 4; |
| bits -= 14; |
| if (unlikely (nlit > 286 || ndist > 30)) |
| { |
| /* Values out of range. */ |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Read and build the table used to compress the |
| literal, length, and distance codes. */ |
| |
| memset(&codebits[0], 0, 19); |
| |
| /* There are always at least 4 elements in the |
| table. */ |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| codebits[16] = val & 7; |
| codebits[17] = (val >> 3) & 7; |
| codebits[18] = (val >> 6) & 7; |
| codebits[0] = (val >> 9) & 7; |
| val >>= 12; |
| bits -= 12; |
| |
| if (nclen == 4) |
| goto codebitsdone; |
| |
| codebits[8] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 5) |
| goto codebitsdone; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| codebits[7] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 6) |
| goto codebitsdone; |
| |
| codebits[9] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 7) |
| goto codebitsdone; |
| |
| codebits[6] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 8) |
| goto codebitsdone; |
| |
| codebits[10] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 9) |
| goto codebitsdone; |
| |
| codebits[5] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 10) |
| goto codebitsdone; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| codebits[11] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 11) |
| goto codebitsdone; |
| |
| codebits[4] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 12) |
| goto codebitsdone; |
| |
| codebits[12] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 13) |
| goto codebitsdone; |
| |
| codebits[3] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 14) |
| goto codebitsdone; |
| |
| codebits[13] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 15) |
| goto codebitsdone; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| codebits[2] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 16) |
| goto codebitsdone; |
| |
| codebits[14] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 17) |
| goto codebitsdone; |
| |
| codebits[1] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| if (nclen == 18) |
| goto codebitsdone; |
| |
| codebits[15] = val & 7; |
| val >>= 3; |
| bits -= 3; |
| |
| codebitsdone: |
| |
| if (!elf_zlib_inflate_table (codebits, 19, zdebug_table, |
| zdebug_table)) |
| return 0; |
| |
| /* Read the compressed bit lengths of the literal, |
| length, and distance codes. We have allocated space |
| at the end of zdebug_table to hold them. */ |
| |
| plenbase = (((unsigned char *) zdebug_table) |
| + ZLIB_TABLE_CODELEN_OFFSET); |
| plen = plenbase; |
| plenend = plen + nlit + ndist; |
| while (plen < plenend) |
| { |
| uint16_t t; |
| unsigned int b; |
| uint16_t v; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| t = zdebug_table[val & 0xff]; |
| |
| /* The compression here uses bit lengths up to 7, so |
| a secondary table is never necessary. */ |
| if (unlikely ((t & (1U << ZLIB_HUFFMAN_SECONDARY_SHIFT)) |
| != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| b = (t >> ZLIB_HUFFMAN_BITS_SHIFT) & ZLIB_HUFFMAN_BITS_MASK; |
| val >>= b + 1; |
| bits -= b + 1; |
| |
| v = t & ZLIB_HUFFMAN_VALUE_MASK; |
| if (v < 16) |
| *plen++ = v; |
| else if (v == 16) |
| { |
| unsigned int c; |
| unsigned int prev; |
| |
| /* Copy previous entry 3 to 6 times. */ |
| |
| if (unlikely (plen == plenbase)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* We used up to 7 bits since the last |
| elf_fetch_bits, so we have at least 8 bits |
| available here. */ |
| |
| c = 3 + (val & 0x3); |
| val >>= 2; |
| bits -= 2; |
| if (unlikely ((unsigned int) (plenend - plen) < c)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| prev = plen[-1]; |
| switch (c) |
| { |
| case 6: |
| *plen++ = prev; |
| ATTRIBUTE_FALLTHROUGH; |
| case 5: |
| *plen++ = prev; |
| ATTRIBUTE_FALLTHROUGH; |
| case 4: |
| *plen++ = prev; |
| } |
| *plen++ = prev; |
| *plen++ = prev; |
| *plen++ = prev; |
| } |
| else if (v == 17) |
| { |
| unsigned int c; |
| |
| /* Store zero 3 to 10 times. */ |
| |
| /* We used up to 7 bits since the last |
| elf_fetch_bits, so we have at least 8 bits |
| available here. */ |
| |
| c = 3 + (val & 0x7); |
| val >>= 3; |
| bits -= 3; |
| if (unlikely ((unsigned int) (plenend - plen) < c)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| switch (c) |
| { |
| case 10: |
| *plen++ = 0; |
| ATTRIBUTE_FALLTHROUGH; |
| case 9: |
| *plen++ = 0; |
| ATTRIBUTE_FALLTHROUGH; |
| case 8: |
| *plen++ = 0; |
| ATTRIBUTE_FALLTHROUGH; |
| case 7: |
| *plen++ = 0; |
| ATTRIBUTE_FALLTHROUGH; |
| case 6: |
| *plen++ = 0; |
| ATTRIBUTE_FALLTHROUGH; |
| case 5: |
| *plen++ = 0; |
| ATTRIBUTE_FALLTHROUGH; |
| case 4: |
| *plen++ = 0; |
| } |
| *plen++ = 0; |
| *plen++ = 0; |
| *plen++ = 0; |
| } |
| else if (v == 18) |
| { |
| unsigned int c; |
| |
| /* Store zero 11 to 138 times. */ |
| |
| /* We used up to 7 bits since the last |
| elf_fetch_bits, so we have at least 8 bits |
| available here. */ |
| |
| c = 11 + (val & 0x7f); |
| val >>= 7; |
| bits -= 7; |
| if (unlikely ((unsigned int) (plenend - plen) < c)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| memset (plen, 0, c); |
| plen += c; |
| } |
| else |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| |
| /* Make sure that the stop code can appear. */ |
| |
| plen = plenbase; |
| if (unlikely (plen[256] == 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Build the decompression tables. */ |
| |
| if (!elf_zlib_inflate_table (plen, nlit, zdebug_table, |
| zdebug_table)) |
| return 0; |
| if (!elf_zlib_inflate_table (plen + nlit, ndist, zdebug_table, |
| (zdebug_table |
| + ZLIB_HUFFMAN_TABLE_SIZE))) |
| return 0; |
| tlit = zdebug_table; |
| tdist = zdebug_table + ZLIB_HUFFMAN_TABLE_SIZE; |
| } |
| |
| /* Inflate values until the end of the block. This is the |
| main loop of the inflation code. */ |
| |
| while (1) |
| { |
| uint16_t t; |
| unsigned int b; |
| uint16_t v; |
| unsigned int lit; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| t = tlit[val & 0xff]; |
| b = (t >> ZLIB_HUFFMAN_BITS_SHIFT) & ZLIB_HUFFMAN_BITS_MASK; |
| v = t & ZLIB_HUFFMAN_VALUE_MASK; |
| |
| if ((t & (1U << ZLIB_HUFFMAN_SECONDARY_SHIFT)) == 0) |
| { |
| lit = v; |
| val >>= b + 1; |
| bits -= b + 1; |
| } |
| else |
| { |
| t = tlit[v + 0x100 + ((val >> 8) & ((1U << b) - 1))]; |
| b = (t >> ZLIB_HUFFMAN_BITS_SHIFT) & ZLIB_HUFFMAN_BITS_MASK; |
| lit = t & ZLIB_HUFFMAN_VALUE_MASK; |
| val >>= b + 8; |
| bits -= b + 8; |
| } |
| |
| if (lit < 256) |
| { |
| if (unlikely (pout == poutend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| *pout++ = lit; |
| |
| /* We will need to write the next byte soon. We ask |
| for high temporal locality because we will write |
| to the whole cache line soon. */ |
| __builtin_prefetch (pout, 1, 3); |
| } |
| else if (lit == 256) |
| { |
| /* The end of the block. */ |
| break; |
| } |
| else |
| { |
| unsigned int dist; |
| unsigned int len; |
| |
| /* Convert lit into a length. */ |
| |
| if (lit < 265) |
| len = lit - 257 + 3; |
| else if (lit == 285) |
| len = 258; |
| else if (unlikely (lit > 285)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| else |
| { |
| unsigned int extra; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| /* This is an expression for the table of length |
| codes in RFC 1951 3.2.5. */ |
| lit -= 265; |
| extra = (lit >> 2) + 1; |
| len = (lit & 3) << extra; |
| len += 11; |
| len += ((1U << (extra - 1)) - 1) << 3; |
| len += val & ((1U << extra) - 1); |
| val >>= extra; |
| bits -= extra; |
| } |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| t = tdist[val & 0xff]; |
| b = (t >> ZLIB_HUFFMAN_BITS_SHIFT) & ZLIB_HUFFMAN_BITS_MASK; |
| v = t & ZLIB_HUFFMAN_VALUE_MASK; |
| |
| if ((t & (1U << ZLIB_HUFFMAN_SECONDARY_SHIFT)) == 0) |
| { |
| dist = v; |
| val >>= b + 1; |
| bits -= b + 1; |
| } |
| else |
| { |
| t = tdist[v + 0x100 + ((val >> 8) & ((1U << b) - 1))]; |
| b = ((t >> ZLIB_HUFFMAN_BITS_SHIFT) |
| & ZLIB_HUFFMAN_BITS_MASK); |
| dist = t & ZLIB_HUFFMAN_VALUE_MASK; |
| val >>= b + 8; |
| bits -= b + 8; |
| } |
| |
| /* Convert dist to a distance. */ |
| |
| if (dist == 0) |
| { |
| /* A distance of 1. A common case, meaning |
| repeat the last character LEN times. */ |
| |
| if (unlikely (pout == porigout)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely ((unsigned int) (poutend - pout) < len)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| memset (pout, pout[-1], len); |
| pout += len; |
| } |
| else if (unlikely (dist > 29)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| else |
| { |
| if (dist < 4) |
| dist = dist + 1; |
| else |
| { |
| unsigned int extra; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| /* This is an expression for the table of |
| distance codes in RFC 1951 3.2.5. */ |
| dist -= 4; |
| extra = (dist >> 1) + 1; |
| dist = (dist & 1) << extra; |
| dist += 5; |
| dist += ((1U << (extra - 1)) - 1) << 2; |
| dist += val & ((1U << extra) - 1); |
| val >>= extra; |
| bits -= extra; |
| } |
| |
| /* Go back dist bytes, and copy len bytes from |
| there. */ |
| |
| if (unlikely ((unsigned int) (pout - porigout) < dist)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely ((unsigned int) (poutend - pout) < len)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (dist >= len) |
| { |
| memcpy (pout, pout - dist, len); |
| pout += len; |
| } |
| else |
| { |
| while (len > 0) |
| { |
| unsigned int copy; |
| |
| copy = len < dist ? len : dist; |
| memcpy (pout, pout - dist, copy); |
| len -= copy; |
| pout += copy; |
| } |
| } |
| } |
| } |
| } |
| } |
| } |
| |
| /* We should have filled the output buffer. */ |
| if (unlikely (pout != poutend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| return 1; |
| } |
| |
| /* Verify the zlib checksum. The checksum is in the 4 bytes at |
| CHECKBYTES, and the uncompressed data is at UNCOMPRESSED / |
| UNCOMPRESSED_SIZE. Returns 1 on success, 0 on failure. */ |
| |
| static int |
| elf_zlib_verify_checksum (const unsigned char *checkbytes, |
| const unsigned char *uncompressed, |
| size_t uncompressed_size) |
| { |
| unsigned int i; |
| unsigned int cksum; |
| const unsigned char *p; |
| uint32_t s1; |
| uint32_t s2; |
| size_t hsz; |
| |
| cksum = 0; |
| for (i = 0; i < 4; i++) |
| cksum = (cksum << 8) | checkbytes[i]; |
| |
| s1 = 1; |
| s2 = 0; |
| |
| /* Minimize modulo operations. */ |
| |
| p = uncompressed; |
| hsz = uncompressed_size; |
| while (hsz >= 5552) |
| { |
| for (i = 0; i < 5552; i += 16) |
| { |
| /* Manually unroll loop 16 times. */ |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| } |
| hsz -= 5552; |
| s1 %= 65521; |
| s2 %= 65521; |
| } |
| |
| while (hsz >= 16) |
| { |
| /* Manually unroll loop 16 times. */ |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| |
| hsz -= 16; |
| } |
| |
| for (i = 0; i < hsz; ++i) |
| { |
| s1 = s1 + *p++; |
| s2 = s2 + s1; |
| } |
| |
| s1 %= 65521; |
| s2 %= 65521; |
| |
| if (unlikely ((s2 << 16) + s1 != cksum)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| return 1; |
| } |
| |
| /* Inflate a zlib stream from PIN/SIN to POUT/SOUT, and verify the |
| checksum. Return 1 on success, 0 on error. */ |
| |
| static int |
| elf_zlib_inflate_and_verify (const unsigned char *pin, size_t sin, |
| uint16_t *zdebug_table, unsigned char *pout, |
| size_t sout) |
| { |
| if (!elf_zlib_inflate (pin, sin, zdebug_table, pout, sout)) |
| return 0; |
| if (!elf_zlib_verify_checksum (pin + sin - 4, pout, sout)) |
| return 0; |
| return 1; |
| } |
| |
| /* For working memory during zstd compression, we need |
| - a literal length FSE table: 512 64-bit values == 4096 bytes |
| - a match length FSE table: 512 64-bit values == 4096 bytes |
| - a offset FSE table: 256 64-bit values == 2048 bytes |
| - a Huffman tree: 2048 uint16_t values == 4096 bytes |
| - scratch space, one of |
| - to build an FSE table: 512 uint16_t values == 1024 bytes |
| - to build a Huffman tree: 512 uint16_t + 256 uint32_t == 2048 bytes |
| */ |
| |
| #define ZSTD_TABLE_SIZE \ |
| (2 * 512 * sizeof (struct elf_zstd_fse_baseline_entry) \ |
| + 256 * sizeof (struct elf_zstd_fse_baseline_entry) \ |
| + 2048 * sizeof (uint16_t) \ |
| + 512 * sizeof (uint16_t) + 256 * sizeof (uint32_t)) |
| |
| #define ZSTD_TABLE_LITERAL_FSE_OFFSET (0) |
| |
| #define ZSTD_TABLE_MATCH_FSE_OFFSET \ |
| (512 * sizeof (struct elf_zstd_fse_baseline_entry)) |
| |
| #define ZSTD_TABLE_OFFSET_FSE_OFFSET \ |
| (ZSTD_TABLE_MATCH_FSE_OFFSET \ |
| + 512 * sizeof (struct elf_zstd_fse_baseline_entry)) |
| |
| #define ZSTD_TABLE_HUFFMAN_OFFSET \ |
| (ZSTD_TABLE_OFFSET_FSE_OFFSET \ |
| + 256 * sizeof (struct elf_zstd_fse_baseline_entry)) |
| |
| #define ZSTD_TABLE_WORK_OFFSET \ |
| (ZSTD_TABLE_HUFFMAN_OFFSET + 2048 * sizeof (uint16_t)) |
| |
| /* An entry in a zstd FSE table. */ |
| |
| struct elf_zstd_fse_entry |
| { |
| /* The value that this FSE entry represents. */ |
| unsigned char symbol; |
| /* The number of bits to read to determine the next state. */ |
| unsigned char bits; |
| /* Add the bits to this base to get the next state. */ |
| uint16_t base; |
| }; |
| |
| static int |
| elf_zstd_build_fse (const int16_t *, int, uint16_t *, int, |
| struct elf_zstd_fse_entry *); |
| |
| /* Read a zstd FSE table and build the decoding table in *TABLE, updating *PPIN |
| as it reads. ZDEBUG_TABLE is scratch space; it must be enough for 512 |
| uint16_t values (1024 bytes). MAXIDX is the maximum number of symbols |
| permitted. *TABLE_BITS is the maximum number of bits for symbols in the |
| table: the size of *TABLE is at least 1 << *TABLE_BITS. This updates |
| *TABLE_BITS to the actual number of bits. Returns 1 on success, 0 on |
| error. */ |
| |
| static int |
| elf_zstd_read_fse (const unsigned char **ppin, const unsigned char *pinend, |
| uint16_t *zdebug_table, int maxidx, |
| struct elf_zstd_fse_entry *table, int *table_bits) |
| { |
| const unsigned char *pin; |
| int16_t *norm; |
| uint16_t *next; |
| uint64_t val; |
| unsigned int bits; |
| int accuracy_log; |
| uint32_t remaining; |
| uint32_t threshold; |
| int bits_needed; |
| int idx; |
| int prev0; |
| |
| pin = *ppin; |
| |
| norm = (int16_t *) zdebug_table; |
| next = zdebug_table + 256; |
| |
| if (unlikely (pin + 3 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Align PIN to a 32-bit boundary. */ |
| |
| val = 0; |
| bits = 0; |
| while ((((uintptr_t) pin) & 3) != 0) |
| { |
| val |= (uint64_t)*pin << bits; |
| bits += 8; |
| ++pin; |
| } |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| accuracy_log = (val & 0xf) + 5; |
| if (accuracy_log > *table_bits) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| *table_bits = accuracy_log; |
| val >>= 4; |
| bits -= 4; |
| |
| /* This code is mostly copied from the reference implementation. */ |
| |
| /* The number of remaining probabilities, plus 1. This sets the number of |
| bits that need to be read for the next value. */ |
| remaining = (1 << accuracy_log) + 1; |
| |
| /* The current difference between small and large values, which depends on |
| the number of remaining values. Small values use one less bit. */ |
| threshold = 1 << accuracy_log; |
| |
| /* The number of bits used to compute threshold. */ |
| bits_needed = accuracy_log + 1; |
| |
| /* The next character value. */ |
| idx = 0; |
| |
| /* Whether the last count was 0. */ |
| prev0 = 0; |
| |
| while (remaining > 1 && idx <= maxidx) |
| { |
| uint32_t max; |
| int32_t count; |
| |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| |
| if (prev0) |
| { |
| int zidx; |
| |
| /* Previous count was 0, so there is a 2-bit repeat flag. If the |
| 2-bit flag is 0b11, it adds 3 and then there is another repeat |
| flag. */ |
| zidx = idx; |
| while ((val & 0xfff) == 0xfff) |
| { |
| zidx += 3 * 6; |
| val >>= 12; |
| bits -= 12; |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| } |
| while ((val & 3) == 3) |
| { |
| zidx += 3; |
| val >>= 2; |
| bits -= 2; |
| if (!elf_fetch_bits (&pin, pinend, &val, &bits)) |
| return 0; |
| } |
| /* We have at least 13 bits here, don't need to fetch. */ |
| zidx += val & 3; |
| val >>= 2; |
| bits -= 2; |
| |
| if (unlikely (zidx > maxidx)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| for (; idx < zidx; idx++) |
| norm[idx] = 0; |
| |
| prev0 = 0; |
| continue; |
| } |
| |
| max = (2 * threshold - 1) - remaining; |
| if ((val & (threshold - 1)) < max) |
| { |
| /* A small value. */ |
| count = (int32_t) ((uint32_t) val & (threshold - 1)); |
| val >>= bits_needed - 1; |
| bits -= bits_needed - 1; |
| } |
| else |
| { |
| /* A large value. */ |
| count = (int32_t) ((uint32_t) val & (2 * threshold - 1)); |
| if (count >= (int32_t) threshold) |
| count -= (int32_t) max; |
| val >>= bits_needed; |
| bits -= bits_needed; |
| } |
| |
| count--; |
| if (count >= 0) |
| remaining -= count; |
| else |
| remaining--; |
| if (unlikely (idx >= 256)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| norm[idx] = (int16_t) count; |
| ++idx; |
| |
| prev0 = count == 0; |
| |
| while (remaining < threshold) |
| { |
| bits_needed--; |
| threshold >>= 1; |
| } |
| } |
| |
| if (unlikely (remaining != 1)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* If we've read ahead more than a byte, back up. */ |
| while (bits >= 8) |
| { |
| --pin; |
| bits -= 8; |
| } |
| |
| *ppin = pin; |
| |
| for (; idx <= maxidx; idx++) |
| norm[idx] = 0; |
| |
| return elf_zstd_build_fse (norm, idx, next, *table_bits, table); |
| } |
| |
| /* Build the FSE decoding table from a list of probabilities. This reads from |
| NORM of length IDX, uses NEXT as scratch space, and writes to *TABLE, whose |
| size is TABLE_BITS. */ |
| |
| static int |
| elf_zstd_build_fse (const int16_t *norm, int idx, uint16_t *next, |
| int table_bits, struct elf_zstd_fse_entry *table) |
| { |
| int table_size; |
| int high_threshold; |
| int i; |
| int pos; |
| int step; |
| int mask; |
| |
| table_size = 1 << table_bits; |
| high_threshold = table_size - 1; |
| for (i = 0; i < idx; i++) |
| { |
| int16_t n; |
| |
| n = norm[i]; |
| if (n >= 0) |
| next[i] = (uint16_t) n; |
| else |
| { |
| table[high_threshold].symbol = (unsigned char) i; |
| high_threshold--; |
| next[i] = 1; |
| } |
| } |
| |
| pos = 0; |
| step = (table_size >> 1) + (table_size >> 3) + 3; |
| mask = table_size - 1; |
| for (i = 0; i < idx; i++) |
| { |
| int n; |
| int j; |
| |
| n = (int) norm[i]; |
| for (j = 0; j < n; j++) |
| { |
| table[pos].symbol = (unsigned char) i; |
| pos = (pos + step) & mask; |
| while (unlikely (pos > high_threshold)) |
| pos = (pos + step) & mask; |
| } |
| } |
| if (unlikely (pos != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| for (i = 0; i < table_size; i++) |
| { |
| unsigned char sym; |
| uint16_t next_state; |
| int high_bit; |
| int bits; |
| |
| sym = table[i].symbol; |
| next_state = next[sym]; |
| ++next[sym]; |
| |
| if (next_state == 0) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| high_bit = 31 - __builtin_clz (next_state); |
| |
| bits = table_bits - high_bit; |
| table[i].bits = (unsigned char) bits; |
| table[i].base = (uint16_t) ((next_state << bits) - table_size); |
| } |
| |
| return 1; |
| } |
| |
| /* Encode the baseline and bits into a single 32-bit value. */ |
| |
| #define ZSTD_ENCODE_BASELINE_BITS(baseline, basebits) \ |
| ((uint32_t)(baseline) | ((uint32_t)(basebits) << 24)) |
| |
| #define ZSTD_DECODE_BASELINE(baseline_basebits) \ |
| ((uint32_t)(baseline_basebits) & 0xffffff) |
| |
| #define ZSTD_DECODE_BASEBITS(baseline_basebits) \ |
| ((uint32_t)(baseline_basebits) >> 24) |
| |
| /* Given a literal length code, we need to read a number of bits and add that |
| to a baseline. For states 0 to 15 the baseline is the state and the number |
| of bits is zero. */ |
| |
| #define ZSTD_LITERAL_LENGTH_BASELINE_OFFSET (16) |
| |
| static const uint32_t elf_zstd_literal_length_base[] = |
| { |
| ZSTD_ENCODE_BASELINE_BITS(16, 1), |
| ZSTD_ENCODE_BASELINE_BITS(18, 1), |
| ZSTD_ENCODE_BASELINE_BITS(20, 1), |
| ZSTD_ENCODE_BASELINE_BITS(22, 1), |
| ZSTD_ENCODE_BASELINE_BITS(24, 2), |
| ZSTD_ENCODE_BASELINE_BITS(28, 2), |
| ZSTD_ENCODE_BASELINE_BITS(32, 3), |
| ZSTD_ENCODE_BASELINE_BITS(40, 3), |
| ZSTD_ENCODE_BASELINE_BITS(48, 4), |
| ZSTD_ENCODE_BASELINE_BITS(64, 6), |
| ZSTD_ENCODE_BASELINE_BITS(128, 7), |
| ZSTD_ENCODE_BASELINE_BITS(256, 8), |
| ZSTD_ENCODE_BASELINE_BITS(512, 9), |
| ZSTD_ENCODE_BASELINE_BITS(1024, 10), |
| ZSTD_ENCODE_BASELINE_BITS(2048, 11), |
| ZSTD_ENCODE_BASELINE_BITS(4096, 12), |
| ZSTD_ENCODE_BASELINE_BITS(8192, 13), |
| ZSTD_ENCODE_BASELINE_BITS(16384, 14), |
| ZSTD_ENCODE_BASELINE_BITS(32768, 15), |
| ZSTD_ENCODE_BASELINE_BITS(65536, 16) |
| }; |
| |
| /* The same applies to match length codes. For states 0 to 31 the baseline is |
| the state + 3 and the number of bits is zero. */ |
| |
| #define ZSTD_MATCH_LENGTH_BASELINE_OFFSET (32) |
| |
| static const uint32_t elf_zstd_match_length_base[] = |
| { |
| ZSTD_ENCODE_BASELINE_BITS(35, 1), |
| ZSTD_ENCODE_BASELINE_BITS(37, 1), |
| ZSTD_ENCODE_BASELINE_BITS(39, 1), |
| ZSTD_ENCODE_BASELINE_BITS(41, 1), |
| ZSTD_ENCODE_BASELINE_BITS(43, 2), |
| ZSTD_ENCODE_BASELINE_BITS(47, 2), |
| ZSTD_ENCODE_BASELINE_BITS(51, 3), |
| ZSTD_ENCODE_BASELINE_BITS(59, 3), |
| ZSTD_ENCODE_BASELINE_BITS(67, 4), |
| ZSTD_ENCODE_BASELINE_BITS(83, 4), |
| ZSTD_ENCODE_BASELINE_BITS(99, 5), |
| ZSTD_ENCODE_BASELINE_BITS(131, 7), |
| ZSTD_ENCODE_BASELINE_BITS(259, 8), |
| ZSTD_ENCODE_BASELINE_BITS(515, 9), |
| ZSTD_ENCODE_BASELINE_BITS(1027, 10), |
| ZSTD_ENCODE_BASELINE_BITS(2051, 11), |
| ZSTD_ENCODE_BASELINE_BITS(4099, 12), |
| ZSTD_ENCODE_BASELINE_BITS(8195, 13), |
| ZSTD_ENCODE_BASELINE_BITS(16387, 14), |
| ZSTD_ENCODE_BASELINE_BITS(32771, 15), |
| ZSTD_ENCODE_BASELINE_BITS(65539, 16) |
| }; |
| |
| /* An entry in an FSE table used for literal/match/length values. For these we |
| have to map the symbol to a baseline value, and we have to read zero or more |
| bits and add that value to the baseline value. Rather than look the values |
| up in a separate table, we grow the FSE table so that we get better memory |
| caching. */ |
| |
| struct elf_zstd_fse_baseline_entry |
| { |
| /* The baseline for the value that this FSE entry represents.. */ |
| uint32_t baseline; |
| /* The number of bits to read to add to the baseline. */ |
| unsigned char basebits; |
| /* The number of bits to read to determine the next state. */ |
| unsigned char bits; |
| /* Add the bits to this base to get the next state. */ |
| uint16_t base; |
| }; |
| |
| /* Convert the literal length FSE table FSE_TABLE to an FSE baseline table at |
| BASELINE_TABLE. Note that FSE_TABLE and BASELINE_TABLE will overlap. */ |
| |
| static int |
| elf_zstd_make_literal_baseline_fse ( |
| const struct elf_zstd_fse_entry *fse_table, |
| int table_bits, |
| struct elf_zstd_fse_baseline_entry *baseline_table) |
| { |
| size_t count; |
| const struct elf_zstd_fse_entry *pfse; |
| struct elf_zstd_fse_baseline_entry *pbaseline; |
| |
| /* Convert backward to avoid overlap. */ |
| |
| count = 1U << table_bits; |
| pfse = fse_table + count; |
| pbaseline = baseline_table + count; |
| while (pfse > fse_table) |
| { |
| unsigned char symbol; |
| unsigned char bits; |
| uint16_t base; |
| |
| --pfse; |
| --pbaseline; |
| symbol = pfse->symbol; |
| bits = pfse->bits; |
| base = pfse->base; |
| if (symbol < ZSTD_LITERAL_LENGTH_BASELINE_OFFSET) |
| { |
| pbaseline->baseline = (uint32_t)symbol; |
| pbaseline->basebits = 0; |
| } |
| else |
| { |
| unsigned int idx; |
| uint32_t basebits; |
| |
| if (unlikely (symbol > 35)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| idx = symbol - ZSTD_LITERAL_LENGTH_BASELINE_OFFSET; |
| basebits = elf_zstd_literal_length_base[idx]; |
| pbaseline->baseline = ZSTD_DECODE_BASELINE(basebits); |
| pbaseline->basebits = ZSTD_DECODE_BASEBITS(basebits); |
| } |
| pbaseline->bits = bits; |
| pbaseline->base = base; |
| } |
| |
| return 1; |
| } |
| |
| /* Convert the offset length FSE table FSE_TABLE to an FSE baseline table at |
| BASELINE_TABLE. Note that FSE_TABLE and BASELINE_TABLE will overlap. */ |
| |
| static int |
| elf_zstd_make_offset_baseline_fse ( |
| const struct elf_zstd_fse_entry *fse_table, |
| int table_bits, |
| struct elf_zstd_fse_baseline_entry *baseline_table) |
| { |
| size_t count; |
| const struct elf_zstd_fse_entry *pfse; |
| struct elf_zstd_fse_baseline_entry *pbaseline; |
| |
| /* Convert backward to avoid overlap. */ |
| |
| count = 1U << table_bits; |
| pfse = fse_table + count; |
| pbaseline = baseline_table + count; |
| while (pfse > fse_table) |
| { |
| unsigned char symbol; |
| unsigned char bits; |
| uint16_t base; |
| |
| --pfse; |
| --pbaseline; |
| symbol = pfse->symbol; |
| bits = pfse->bits; |
| base = pfse->base; |
| if (unlikely (symbol > 31)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* The simple way to write this is |
| |
| pbaseline->baseline = (uint32_t)1 << symbol; |
| pbaseline->basebits = symbol; |
| |
| That will give us an offset value that corresponds to the one |
| described in the RFC. However, for offset values > 3, we have to |
| subtract 3. And for offset values 1, 2, 3 we use a repeated offset. |
| The baseline is always a power of 2, and is never 0, so for these low |
| values we will see one entry that is baseline 1, basebits 0, and one |
| entry that is baseline 2, basebits 1. All other entries will have |
| baseline >= 4 and basebits >= 2. |
| |
| So we can check for RFC offset <= 3 by checking for basebits <= 1. |
| And that means that we can subtract 3 here and not worry about doing |
| it in the hot loop. */ |
| |
| pbaseline->baseline = (uint32_t)1 << symbol; |
| if (symbol >= 2) |
| pbaseline->baseline -= 3; |
| pbaseline->basebits = symbol; |
| pbaseline->bits = bits; |
| pbaseline->base = base; |
| } |
| |
| return 1; |
| } |
| |
| /* Convert the match length FSE table FSE_TABLE to an FSE baseline table at |
| BASELINE_TABLE. Note that FSE_TABLE and BASELINE_TABLE will overlap. */ |
| |
| static int |
| elf_zstd_make_match_baseline_fse ( |
| const struct elf_zstd_fse_entry *fse_table, |
| int table_bits, |
| struct elf_zstd_fse_baseline_entry *baseline_table) |
| { |
| size_t count; |
| const struct elf_zstd_fse_entry *pfse; |
| struct elf_zstd_fse_baseline_entry *pbaseline; |
| |
| /* Convert backward to avoid overlap. */ |
| |
| count = 1U << table_bits; |
| pfse = fse_table + count; |
| pbaseline = baseline_table + count; |
| while (pfse > fse_table) |
| { |
| unsigned char symbol; |
| unsigned char bits; |
| uint16_t base; |
| |
| --pfse; |
| --pbaseline; |
| symbol = pfse->symbol; |
| bits = pfse->bits; |
| base = pfse->base; |
| if (symbol < ZSTD_MATCH_LENGTH_BASELINE_OFFSET) |
| { |
| pbaseline->baseline = (uint32_t)symbol + 3; |
| pbaseline->basebits = 0; |
| } |
| else |
| { |
| unsigned int idx; |
| uint32_t basebits; |
| |
| if (unlikely (symbol > 52)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| idx = symbol - ZSTD_MATCH_LENGTH_BASELINE_OFFSET; |
| basebits = elf_zstd_match_length_base[idx]; |
| pbaseline->baseline = ZSTD_DECODE_BASELINE(basebits); |
| pbaseline->basebits = ZSTD_DECODE_BASEBITS(basebits); |
| } |
| pbaseline->bits = bits; |
| pbaseline->base = base; |
| } |
| |
| return 1; |
| } |
| |
| #ifdef BACKTRACE_GENERATE_ZSTD_FSE_TABLES |
| |
| /* Used to generate the predefined FSE decoding tables for zstd. */ |
| |
| #include <stdio.h> |
| |
| /* These values are straight from RFC 8878. */ |
| |
| static int16_t lit[36] = |
| { |
| 4, 3, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 1, 1, 1, |
| 2, 2, 2, 2, 2, 2, 2, 2, 2, 3, 2, 1, 1, 1, 1, 1, |
| -1,-1,-1,-1 |
| }; |
| |
| static int16_t match[53] = |
| { |
| 1, 4, 3, 2, 2, 2, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,-1,-1, |
| -1,-1,-1,-1,-1 |
| }; |
| |
| static int16_t offset[29] = |
| { |
| 1, 1, 1, 1, 1, 1, 2, 2, 2, 1, 1, 1, 1, 1, 1, 1, |
| 1, 1, 1, 1, 1, 1, 1, 1,-1,-1,-1,-1,-1 |
| }; |
| |
| static uint16_t next[256]; |
| |
| static void |
| print_table (const struct elf_zstd_fse_baseline_entry *table, size_t size) |
| { |
| size_t i; |
| |
| printf ("{\n"); |
| for (i = 0; i < size; i += 3) |
| { |
| int j; |
| |
| printf (" "); |
| for (j = 0; j < 3 && i + j < size; ++j) |
| printf (" { %u, %d, %d, %d },", table[i + j].baseline, |
| table[i + j].basebits, table[i + j].bits, |
| table[i + j].base); |
| printf ("\n"); |
| } |
| printf ("};\n"); |
| } |
| |
| int |
| main () |
| { |
| struct elf_zstd_fse_entry lit_table[64]; |
| struct elf_zstd_fse_baseline_entry lit_baseline[64]; |
| struct elf_zstd_fse_entry match_table[64]; |
| struct elf_zstd_fse_baseline_entry match_baseline[64]; |
| struct elf_zstd_fse_entry offset_table[32]; |
| struct elf_zstd_fse_baseline_entry offset_baseline[32]; |
| |
| if (!elf_zstd_build_fse (lit, sizeof lit / sizeof lit[0], next, |
| 6, lit_table)) |
| { |
| fprintf (stderr, "elf_zstd_build_fse failed\n"); |
| exit (EXIT_FAILURE); |
| } |
| |
| if (!elf_zstd_make_literal_baseline_fse (lit_table, 6, lit_baseline)) |
| { |
| fprintf (stderr, "elf_zstd_make_literal_baseline_fse failed\n"); |
| exit (EXIT_FAILURE); |
| } |
| |
| printf ("static const struct elf_zstd_fse_baseline_entry " |
| "elf_zstd_lit_table[64] =\n"); |
| print_table (lit_baseline, |
| sizeof lit_baseline / sizeof lit_baseline[0]); |
| printf ("\n"); |
| |
| if (!elf_zstd_build_fse (match, sizeof match / sizeof match[0], next, |
| 6, match_table)) |
| { |
| fprintf (stderr, "elf_zstd_build_fse failed\n"); |
| exit (EXIT_FAILURE); |
| } |
| |
| if (!elf_zstd_make_match_baseline_fse (match_table, 6, match_baseline)) |
| { |
| fprintf (stderr, "elf_zstd_make_match_baseline_fse failed\n"); |
| exit (EXIT_FAILURE); |
| } |
| |
| printf ("static const struct elf_zstd_fse_baseline_entry " |
| "elf_zstd_match_table[64] =\n"); |
| print_table (match_baseline, |
| sizeof match_baseline / sizeof match_baseline[0]); |
| printf ("\n"); |
| |
| if (!elf_zstd_build_fse (offset, sizeof offset / sizeof offset[0], next, |
| 5, offset_table)) |
| { |
| fprintf (stderr, "elf_zstd_build_fse failed\n"); |
| exit (EXIT_FAILURE); |
| } |
| |
| if (!elf_zstd_make_offset_baseline_fse (offset_table, 5, offset_baseline)) |
| { |
| fprintf (stderr, "elf_zstd_make_offset_baseline_fse failed\n"); |
| exit (EXIT_FAILURE); |
| } |
| |
| printf ("static const struct elf_zstd_fse_baseline_entry " |
| "elf_zstd_offset_table[32] =\n"); |
| print_table (offset_baseline, |
| sizeof offset_baseline / sizeof offset_baseline[0]); |
| printf ("\n"); |
| |
| return 0; |
| } |
| |
| #endif |
| |
| /* The fixed tables generated by the #ifdef'ed out main function |
| above. */ |
| |
| static const struct elf_zstd_fse_baseline_entry elf_zstd_lit_table[64] = |
| { |
| { 0, 0, 4, 0 }, { 0, 0, 4, 16 }, { 1, 0, 5, 32 }, |
| { 3, 0, 5, 0 }, { 4, 0, 5, 0 }, { 6, 0, 5, 0 }, |
| { 7, 0, 5, 0 }, { 9, 0, 5, 0 }, { 10, 0, 5, 0 }, |
| { 12, 0, 5, 0 }, { 14, 0, 6, 0 }, { 16, 1, 5, 0 }, |
| { 20, 1, 5, 0 }, { 22, 1, 5, 0 }, { 28, 2, 5, 0 }, |
| { 32, 3, 5, 0 }, { 48, 4, 5, 0 }, { 64, 6, 5, 32 }, |
| { 128, 7, 5, 0 }, { 256, 8, 6, 0 }, { 1024, 10, 6, 0 }, |
| { 4096, 12, 6, 0 }, { 0, 0, 4, 32 }, { 1, 0, 4, 0 }, |
| { 2, 0, 5, 0 }, { 4, 0, 5, 32 }, { 5, 0, 5, 0 }, |
| { 7, 0, 5, 32 }, { 8, 0, 5, 0 }, { 10, 0, 5, 32 }, |
| { 11, 0, 5, 0 }, { 13, 0, 6, 0 }, { 16, 1, 5, 32 }, |
| { 18, 1, 5, 0 }, { 22, 1, 5, 32 }, { 24, 2, 5, 0 }, |
| { 32, 3, 5, 32 }, { 40, 3, 5, 0 }, { 64, 6, 4, 0 }, |
| { 64, 6, 4, 16 }, { 128, 7, 5, 32 }, { 512, 9, 6, 0 }, |
| { 2048, 11, 6, 0 }, { 0, 0, 4, 48 }, { 1, 0, 4, 16 }, |
| { 2, 0, 5, 32 }, { 3, 0, 5, 32 }, { 5, 0, 5, 32 }, |
| { 6, 0, 5, 32 }, { 8, 0, 5, 32 }, { 9, 0, 5, 32 }, |
| { 11, 0, 5, 32 }, { 12, 0, 5, 32 }, { 15, 0, 6, 0 }, |
| { 18, 1, 5, 32 }, { 20, 1, 5, 32 }, { 24, 2, 5, 32 }, |
| { 28, 2, 5, 32 }, { 40, 3, 5, 32 }, { 48, 4, 5, 32 }, |
| { 65536, 16, 6, 0 }, { 32768, 15, 6, 0 }, { 16384, 14, 6, 0 }, |
| { 8192, 13, 6, 0 }, |
| }; |
| |
| static const struct elf_zstd_fse_baseline_entry elf_zstd_match_table[64] = |
| { |
| { 3, 0, 6, 0 }, { 4, 0, 4, 0 }, { 5, 0, 5, 32 }, |
| { 6, 0, 5, 0 }, { 8, 0, 5, 0 }, { 9, 0, 5, 0 }, |
| { 11, 0, 5, 0 }, { 13, 0, 6, 0 }, { 16, 0, 6, 0 }, |
| { 19, 0, 6, 0 }, { 22, 0, 6, 0 }, { 25, 0, 6, 0 }, |
| { 28, 0, 6, 0 }, { 31, 0, 6, 0 }, { 34, 0, 6, 0 }, |
| { 37, 1, 6, 0 }, { 41, 1, 6, 0 }, { 47, 2, 6, 0 }, |
| { 59, 3, 6, 0 }, { 83, 4, 6, 0 }, { 131, 7, 6, 0 }, |
| { 515, 9, 6, 0 }, { 4, 0, 4, 16 }, { 5, 0, 4, 0 }, |
| { 6, 0, 5, 32 }, { 7, 0, 5, 0 }, { 9, 0, 5, 32 }, |
| { 10, 0, 5, 0 }, { 12, 0, 6, 0 }, { 15, 0, 6, 0 }, |
| { 18, 0, 6, 0 }, { 21, 0, 6, 0 }, { 24, 0, 6, 0 }, |
| { 27, 0, 6, 0 }, { 30, 0, 6, 0 }, { 33, 0, 6, 0 }, |
| { 35, 1, 6, 0 }, { 39, 1, 6, 0 }, { 43, 2, 6, 0 }, |
| { 51, 3, 6, 0 }, { 67, 4, 6, 0 }, { 99, 5, 6, 0 }, |
| { 259, 8, 6, 0 }, { 4, 0, 4, 32 }, { 4, 0, 4, 48 }, |
| { 5, 0, 4, 16 }, { 7, 0, 5, 32 }, { 8, 0, 5, 32 }, |
| { 10, 0, 5, 32 }, { 11, 0, 5, 32 }, { 14, 0, 6, 0 }, |
| { 17, 0, 6, 0 }, { 20, 0, 6, 0 }, { 23, 0, 6, 0 }, |
| { 26, 0, 6, 0 }, { 29, 0, 6, 0 }, { 32, 0, 6, 0 }, |
| { 65539, 16, 6, 0 }, { 32771, 15, 6, 0 }, { 16387, 14, 6, 0 }, |
| { 8195, 13, 6, 0 }, { 4099, 12, 6, 0 }, { 2051, 11, 6, 0 }, |
| { 1027, 10, 6, 0 }, |
| }; |
| |
| static const struct elf_zstd_fse_baseline_entry elf_zstd_offset_table[32] = |
| { |
| { 1, 0, 5, 0 }, { 61, 6, 4, 0 }, { 509, 9, 5, 0 }, |
| { 32765, 15, 5, 0 }, { 2097149, 21, 5, 0 }, { 5, 3, 5, 0 }, |
| { 125, 7, 4, 0 }, { 4093, 12, 5, 0 }, { 262141, 18, 5, 0 }, |
| { 8388605, 23, 5, 0 }, { 29, 5, 5, 0 }, { 253, 8, 4, 0 }, |
| { 16381, 14, 5, 0 }, { 1048573, 20, 5, 0 }, { 1, 2, 5, 0 }, |
| { 125, 7, 4, 16 }, { 2045, 11, 5, 0 }, { 131069, 17, 5, 0 }, |
| { 4194301, 22, 5, 0 }, { 13, 4, 5, 0 }, { 253, 8, 4, 16 }, |
| { 8189, 13, 5, 0 }, { 524285, 19, 5, 0 }, { 2, 1, 5, 0 }, |
| { 61, 6, 4, 16 }, { 1021, 10, 5, 0 }, { 65533, 16, 5, 0 }, |
| { 268435453, 28, 5, 0 }, { 134217725, 27, 5, 0 }, { 67108861, 26, 5, 0 }, |
| { 33554429, 25, 5, 0 }, { 16777213, 24, 5, 0 }, |
| }; |
| |
| /* Read a zstd Huffman table and build the decoding table in *TABLE, reading |
| and updating *PPIN. This sets *PTABLE_BITS to the number of bits of the |
| table, such that the table length is 1 << *TABLE_BITS. ZDEBUG_TABLE is |
| scratch space; it must be enough for 512 uint16_t values + 256 32-bit values |
| (2048 bytes). Returns 1 on success, 0 on error. */ |
| |
| static int |
| elf_zstd_read_huff (const unsigned char **ppin, const unsigned char *pinend, |
| uint16_t *zdebug_table, uint16_t *table, int *ptable_bits) |
| { |
| const unsigned char *pin; |
| unsigned char hdr; |
| unsigned char *weights; |
| size_t count; |
| uint32_t *weight_mark; |
| size_t i; |
| uint32_t weight_mask; |
| size_t table_bits; |
| |
| pin = *ppin; |
| if (unlikely (pin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| hdr = *pin; |
| ++pin; |
| |
| weights = (unsigned char *) zdebug_table; |
| |
| if (hdr < 128) |
| { |
| /* Table is compressed using FSE. */ |
| |
| struct elf_zstd_fse_entry *fse_table; |
| int fse_table_bits; |
| uint16_t *scratch; |
| const unsigned char *pfse; |
| const unsigned char *pback; |
| uint64_t val; |
| unsigned int bits; |
| unsigned int state1, state2; |
| |
| /* SCRATCH is used temporarily by elf_zstd_read_fse. It overlaps |
| WEIGHTS. */ |
| scratch = zdebug_table; |
| fse_table = (struct elf_zstd_fse_entry *) (scratch + 512); |
| fse_table_bits = 6; |
| |
| pfse = pin; |
| if (!elf_zstd_read_fse (&pfse, pinend, scratch, 255, fse_table, |
| &fse_table_bits)) |
| return 0; |
| |
| if (unlikely (pin + hdr > pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* We no longer need SCRATCH. Start recording weights. We need up to |
| 256 bytes of weights and 64 bytes of rank counts, so it won't overlap |
| FSE_TABLE. */ |
| |
| pback = pin + hdr - 1; |
| |
| if (!elf_fetch_backward_init (&pback, pfse, &val, &bits)) |
| return 0; |
| |
| bits -= fse_table_bits; |
| state1 = (val >> bits) & ((1U << fse_table_bits) - 1); |
| bits -= fse_table_bits; |
| state2 = (val >> bits) & ((1U << fse_table_bits) - 1); |
| |
| /* There are two independent FSE streams, tracked by STATE1 and STATE2. |
| We decode them alternately. */ |
| |
| count = 0; |
| while (1) |
| { |
| struct elf_zstd_fse_entry *pt; |
| uint64_t v; |
| |
| pt = &fse_table[state1]; |
| |
| if (unlikely (pin < pinend) && bits < pt->bits) |
| { |
| if (unlikely (count >= 254)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| weights[count] = (unsigned char) pt->symbol; |
| weights[count + 1] = (unsigned char) fse_table[state2].symbol; |
| count += 2; |
| break; |
| } |
| |
| if (unlikely (pt->bits == 0)) |
| v = 0; |
| else |
| { |
| if (!elf_fetch_bits_backward (&pback, pfse, &val, &bits)) |
| return 0; |
| |
| bits -= pt->bits; |
| v = (val >> bits) & (((uint64_t)1 << pt->bits) - 1); |
| } |
| |
| state1 = pt->base + v; |
| |
| if (unlikely (count >= 255)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| weights[count] = pt->symbol; |
| ++count; |
| |
| pt = &fse_table[state2]; |
| |
| if (unlikely (pin < pinend && bits < pt->bits)) |
| { |
| if (unlikely (count >= 254)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| weights[count] = (unsigned char) pt->symbol; |
| weights[count + 1] = (unsigned char) fse_table[state1].symbol; |
| count += 2; |
| break; |
| } |
| |
| if (unlikely (pt->bits == 0)) |
| v = 0; |
| else |
| { |
| if (!elf_fetch_bits_backward (&pback, pfse, &val, &bits)) |
| return 0; |
| |
| bits -= pt->bits; |
| v = (val >> bits) & (((uint64_t)1 << pt->bits) - 1); |
| } |
| |
| state2 = pt->base + v; |
| |
| if (unlikely (count >= 255)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| weights[count] = pt->symbol; |
| ++count; |
| } |
| |
| pin += hdr; |
| } |
| else |
| { |
| /* Table is not compressed. Each weight is 4 bits. */ |
| |
| count = hdr - 127; |
| if (unlikely (pin + ((count + 1) / 2) >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| for (i = 0; i < count; i += 2) |
| { |
| unsigned char b; |
| |
| b = *pin; |
| ++pin; |
| weights[i] = b >> 4; |
| weights[i + 1] = b & 0xf; |
| } |
| } |
| |
| weight_mark = (uint32_t *) (weights + 256); |
| memset (weight_mark, 0, 13 * sizeof (uint32_t)); |
| weight_mask = 0; |
| for (i = 0; i < count; ++i) |
| { |
| unsigned char w; |
| |
| w = weights[i]; |
| if (unlikely (w > 12)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| ++weight_mark[w]; |
| if (w > 0) |
| weight_mask += 1U << (w - 1); |
| } |
| if (unlikely (weight_mask == 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| table_bits = 32 - __builtin_clz (weight_mask); |
| if (unlikely (table_bits > 11)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Work out the last weight value, which is omitted because the weights must |
| sum to a power of two. */ |
| { |
| uint32_t left; |
| uint32_t high_bit; |
| |
| left = ((uint32_t)1 << table_bits) - weight_mask; |
| if (left == 0) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| high_bit = 31 - __builtin_clz (left); |
| if (((uint32_t)1 << high_bit) != left) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely (count >= 256)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| weights[count] = high_bit + 1; |
| ++count; |
| ++weight_mark[high_bit + 1]; |
| } |
| |
| if (weight_mark[1] < 2 || (weight_mark[1] & 1) != 0) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Change WEIGHT_MARK from a count of weights to the index of the first |
| symbol for that weight. We shift the indexes to also store how many we |
| have seen so far, below. */ |
| { |
| uint32_t next; |
| |
| next = 0; |
| for (i = 0; i < table_bits; ++i) |
| { |
| uint32_t cur; |
| |
| cur = next; |
| next += weight_mark[i + 1] << i; |
| weight_mark[i + 1] = cur; |
| } |
| } |
| |
| for (i = 0; i < count; ++i) |
| { |
| unsigned char weight; |
| uint32_t length; |
| uint16_t tval; |
| size_t start; |
| uint32_t j; |
| |
| weight = weights[i]; |
| if (weight == 0) |
| continue; |
| |
| length = 1U << (weight - 1); |
| tval = (i << 8) | (table_bits + 1 - weight); |
| start = weight_mark[weight]; |
| for (j = 0; j < length; ++j) |
| table[start + j] = tval; |
| weight_mark[weight] += length; |
| } |
| |
| *ppin = pin; |
| *ptable_bits = (int)table_bits; |
| |
| return 1; |
| } |
| |
| /* Read and decompress the literals and store them ending at POUTEND. This |
| works because we are going to use all the literals in the output, so they |
| must fit into the output buffer. HUFFMAN_TABLE, and PHUFFMAN_TABLE_BITS |
| store the Huffman table across calls. SCRATCH is used to read a Huffman |
| table. Store the start of the decompressed literals in *PPLIT. Update |
| *PPIN. Return 1 on success, 0 on error. */ |
| |
| static int |
| elf_zstd_read_literals (const unsigned char **ppin, |
| const unsigned char *pinend, |
| unsigned char *pout, |
| unsigned char *poutend, |
| uint16_t *scratch, |
| uint16_t *huffman_table, |
| int *phuffman_table_bits, |
| unsigned char **pplit) |
| { |
| const unsigned char *pin; |
| unsigned char *plit; |
| unsigned char hdr; |
| uint32_t regenerated_size; |
| uint32_t compressed_size; |
| int streams; |
| uint32_t total_streams_size; |
| unsigned int huffman_table_bits; |
| uint64_t huffman_mask; |
| |
| pin = *ppin; |
| if (unlikely (pin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| hdr = *pin; |
| ++pin; |
| |
| if ((hdr & 3) == 0 || (hdr & 3) == 1) |
| { |
| int raw; |
| |
| /* Raw_Literals_Block or RLE_Literals_Block */ |
| |
| raw = (hdr & 3) == 0; |
| |
| switch ((hdr >> 2) & 3) |
| { |
| case 0: case 2: |
| regenerated_size = hdr >> 3; |
| break; |
| case 1: |
| if (unlikely (pin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| regenerated_size = (hdr >> 4) + ((uint32_t)(*pin) << 4); |
| ++pin; |
| break; |
| case 3: |
| if (unlikely (pin + 1 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| regenerated_size = ((hdr >> 4) |
| + ((uint32_t)*pin << 4) |
| + ((uint32_t)pin[1] << 12)); |
| pin += 2; |
| break; |
| default: |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely ((size_t)(poutend - pout) < regenerated_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| plit = poutend - regenerated_size; |
| |
| if (raw) |
| { |
| if (unlikely (pin + regenerated_size >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| memcpy (plit, pin, regenerated_size); |
| pin += regenerated_size; |
| } |
| else |
| { |
| if (pin >= pinend) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| memset (plit, *pin, regenerated_size); |
| ++pin; |
| } |
| |
| *ppin = pin; |
| *pplit = plit; |
| |
| return 1; |
| } |
| |
| /* Compressed_Literals_Block or Treeless_Literals_Block */ |
| |
| switch ((hdr >> 2) & 3) |
| { |
| case 0: case 1: |
| if (unlikely (pin + 1 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| regenerated_size = (hdr >> 4) | ((uint32_t)(*pin & 0x3f) << 4); |
| compressed_size = (uint32_t)*pin >> 6 | ((uint32_t)pin[1] << 2); |
| pin += 2; |
| streams = ((hdr >> 2) & 3) == 0 ? 1 : 4; |
| break; |
| case 2: |
| if (unlikely (pin + 2 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| regenerated_size = (((uint32_t)hdr >> 4) |
| | ((uint32_t)*pin << 4) |
| | (((uint32_t)pin[1] & 3) << 12)); |
| compressed_size = (((uint32_t)pin[1] >> 2) |
| | ((uint32_t)pin[2] << 6)); |
| pin += 3; |
| streams = 4; |
| break; |
| case 3: |
| if (unlikely (pin + 3 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| regenerated_size = (((uint32_t)hdr >> 4) |
| | ((uint32_t)*pin << 4) |
| | (((uint32_t)pin[1] & 0x3f) << 12)); |
| compressed_size = (((uint32_t)pin[1] >> 6) |
| | ((uint32_t)pin[2] << 2) |
| | ((uint32_t)pin[3] << 10)); |
| pin += 4; |
| streams = 4; |
| break; |
| default: |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely (pin + compressed_size > pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| pinend = pin + compressed_size; |
| *ppin = pinend; |
| |
| if (unlikely ((size_t)(poutend - pout) < regenerated_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| plit = poutend - regenerated_size; |
| |
| *pplit = plit; |
| |
| total_streams_size = compressed_size; |
| if ((hdr & 3) == 2) |
| { |
| const unsigned char *ptable; |
| |
| /* Compressed_Literals_Block. Read Huffman tree. */ |
| |
| ptable = pin; |
| if (!elf_zstd_read_huff (&ptable, pinend, scratch, huffman_table, |
| phuffman_table_bits)) |
| return 0; |
| |
| if (unlikely (total_streams_size < (size_t)(ptable - pin))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| total_streams_size -= ptable - pin; |
| pin = ptable; |
| } |
| else |
| { |
| /* Treeless_Literals_Block. Reuse previous Huffman tree. */ |
| if (unlikely (*phuffman_table_bits == 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| |
| /* Decompress COMPRESSED_SIZE bytes of data at PIN using the huffman table, |
| storing REGENERATED_SIZE bytes of decompressed data at PLIT. */ |
| |
| huffman_table_bits = (unsigned int)*phuffman_table_bits; |
| huffman_mask = ((uint64_t)1 << huffman_table_bits) - 1; |
| |
| if (streams == 1) |
| { |
| const unsigned char *pback; |
| const unsigned char *pbackend; |
| uint64_t val; |
| unsigned int bits; |
| uint32_t i; |
| |
| pback = pin + total_streams_size - 1; |
| pbackend = pin; |
| if (!elf_fetch_backward_init (&pback, pbackend, &val, &bits)) |
| return 0; |
| |
| /* This is one of the inner loops of the decompression algorithm, so we |
| put some effort into optimization. We can't get more than 64 bytes |
| from a single call to elf_fetch_bits_backward, and we can't subtract |
| more than 11 bits at a time. */ |
| |
| if (regenerated_size >= 64) |
| { |
| unsigned char *plitstart; |
| unsigned char *plitstop; |
| |
| plitstart = plit; |
| plitstop = plit + regenerated_size - 64; |
| while (plit < plitstop) |
| { |
| uint16_t t; |
| |
| if (!elf_fetch_bits_backward (&pback, pbackend, &val, &bits)) |
| return 0; |
| |
| if (bits < 16) |
| break; |
| |
| while (bits >= 33) |
| { |
| t = huffman_table[(val >> (bits - huffman_table_bits)) |
| & huffman_mask]; |
| *plit = t >> 8; |
| ++plit; |
| bits -= t & 0xff; |
| |
| t = huffman_table[(val >> (bits - huffman_table_bits)) |
| & huffman_mask]; |
| *plit = t >> 8; |
| ++plit; |
| bits -= t & 0xff; |
| |
| t = huffman_table[(val >> (bits - huffman_table_bits)) |
| & huffman_mask]; |
| *plit = t >> 8; |
| ++plit; |
| bits -= t & 0xff; |
| } |
| |
| while (bits > 11) |
| { |
| t = huffman_table[(val >> (bits - huffman_table_bits)) |
| & huffman_mask]; |
| *plit = t >> 8; |
| ++plit; |
| bits -= t & 0xff; |
| } |
| } |
| |
| regenerated_size -= plit - plitstart; |
| } |
| |
| for (i = 0; i < regenerated_size; ++i) |
| { |
| uint16_t t; |
| |
| if (!elf_fetch_bits_backward (&pback, pbackend, &val, &bits)) |
| return 0; |
| |
| if (unlikely (bits < huffman_table_bits)) |
| { |
| t = huffman_table[(val << (huffman_table_bits - bits)) |
| & huffman_mask]; |
| if (unlikely (bits < (t & 0xff))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| else |
| t = huffman_table[(val >> (bits - huffman_table_bits)) |
| & huffman_mask]; |
| |
| *plit = t >> 8; |
| ++plit; |
| bits -= t & 0xff; |
| } |
| |
| return 1; |
| } |
| |
| { |
| uint32_t stream_size1, stream_size2, stream_size3, stream_size4; |
| uint32_t tot; |
| const unsigned char *pback1, *pback2, *pback3, *pback4; |
| const unsigned char *pbackend1, *pbackend2, *pbackend3, *pbackend4; |
| uint64_t val1, val2, val3, val4; |
| unsigned int bits1, bits2, bits3, bits4; |
| unsigned char *plit1, *plit2, *plit3, *plit4; |
| uint32_t regenerated_stream_size; |
| uint32_t regenerated_stream_size4; |
| uint16_t t1, t2, t3, t4; |
| uint32_t i; |
| uint32_t limit; |
| |
| /* Read jump table. */ |
| if (unlikely (pin + 5 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| stream_size1 = (uint32_t)*pin | ((uint32_t)pin[1] << 8); |
| pin += 2; |
| stream_size2 = (uint32_t)*pin | ((uint32_t)pin[1] << 8); |
| pin += 2; |
| stream_size3 = (uint32_t)*pin | ((uint32_t)pin[1] << 8); |
| pin += 2; |
| tot = stream_size1 + stream_size2 + stream_size3; |
| if (unlikely (tot > total_streams_size - 6)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| stream_size4 = total_streams_size - 6 - tot; |
| |
| pback1 = pin + stream_size1 - 1; |
| pbackend1 = pin; |
| |
| pback2 = pback1 + stream_size2; |
| pbackend2 = pback1 + 1; |
| |
| pback3 = pback2 + stream_size3; |
| pbackend3 = pback2 + 1; |
| |
| pback4 = pback3 + stream_size4; |
| pbackend4 = pback3 + 1; |
| |
| if (!elf_fetch_backward_init (&pback1, pbackend1, &val1, &bits1)) |
| return 0; |
| if (!elf_fetch_backward_init (&pback2, pbackend2, &val2, &bits2)) |
| return 0; |
| if (!elf_fetch_backward_init (&pback3, pbackend3, &val3, &bits3)) |
| return 0; |
| if (!elf_fetch_backward_init (&pback4, pbackend4, &val4, &bits4)) |
| return 0; |
| |
| regenerated_stream_size = (regenerated_size + 3) / 4; |
| |
| plit1 = plit; |
| plit2 = plit1 + regenerated_stream_size; |
| plit3 = plit2 + regenerated_stream_size; |
| plit4 = plit3 + regenerated_stream_size; |
| |
| regenerated_stream_size4 = regenerated_size - regenerated_stream_size * 3; |
| |
| /* We can't get more than 64 literal bytes from a single call to |
| elf_fetch_bits_backward. The fourth stream can be up to 3 bytes less, |
| so use as the limit. */ |
| |
| limit = regenerated_stream_size4 <= 64 ? 0 : regenerated_stream_size4 - 64; |
| i = 0; |
| while (i < limit) |
| { |
| if (!elf_fetch_bits_backward (&pback1, pbackend1, &val1, &bits1)) |
| return 0; |
| if (!elf_fetch_bits_backward (&pback2, pbackend2, &val2, &bits2)) |
| return 0; |
| if (!elf_fetch_bits_backward (&pback3, pbackend3, &val3, &bits3)) |
| return 0; |
| if (!elf_fetch_bits_backward (&pback4, pbackend4, &val4, &bits4)) |
| return 0; |
| |
| /* We can't subtract more than 11 bits at a time. */ |
| |
| do |
| { |
| t1 = huffman_table[(val1 >> (bits1 - huffman_table_bits)) |
| & huffman_mask]; |
| t2 = huffman_table[(val2 >> (bits2 - huffman_table_bits)) |
| & huffman_mask]; |
| t3 = huffman_table[(val3 >> (bits3 - huffman_table_bits)) |
| & huffman_mask]; |
| t4 = huffman_table[(val4 >> (bits4 - huffman_table_bits)) |
| & huffman_mask]; |
| |
| *plit1 = t1 >> 8; |
| ++plit1; |
| bits1 -= t1 & 0xff; |
| |
| *plit2 = t2 >> 8; |
| ++plit2; |
| bits2 -= t2 & 0xff; |
| |
| *plit3 = t3 >> 8; |
| ++plit3; |
| bits3 -= t3 & 0xff; |
| |
| *plit4 = t4 >> 8; |
| ++plit4; |
| bits4 -= t4 & 0xff; |
| |
| ++i; |
| } |
| while (bits1 > 11 && bits2 > 11 && bits3 > 11 && bits4 > 11); |
| } |
| |
| while (i < regenerated_stream_size) |
| { |
| int use4; |
| |
| use4 = i < regenerated_stream_size4; |
| |
| if (!elf_fetch_bits_backward (&pback1, pbackend1, &val1, &bits1)) |
| return 0; |
| if (!elf_fetch_bits_backward (&pback2, pbackend2, &val2, &bits2)) |
| return 0; |
| if (!elf_fetch_bits_backward (&pback3, pbackend3, &val3, &bits3)) |
| return 0; |
| if (use4) |
| { |
| if (!elf_fetch_bits_backward (&pback4, pbackend4, &val4, &bits4)) |
| return 0; |
| } |
| |
| if (unlikely (bits1 < huffman_table_bits)) |
| { |
| t1 = huffman_table[(val1 << (huffman_table_bits - bits1)) |
| & huffman_mask]; |
| if (unlikely (bits1 < (t1 & 0xff))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| else |
| t1 = huffman_table[(val1 >> (bits1 - huffman_table_bits)) |
| & huffman_mask]; |
| |
| if (unlikely (bits2 < huffman_table_bits)) |
| { |
| t2 = huffman_table[(val2 << (huffman_table_bits - bits2)) |
| & huffman_mask]; |
| if (unlikely (bits2 < (t2 & 0xff))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| else |
| t2 = huffman_table[(val2 >> (bits2 - huffman_table_bits)) |
| & huffman_mask]; |
| |
| if (unlikely (bits3 < huffman_table_bits)) |
| { |
| t3 = huffman_table[(val3 << (huffman_table_bits - bits3)) |
| & huffman_mask]; |
| if (unlikely (bits3 < (t3 & 0xff))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| else |
| t3 = huffman_table[(val3 >> (bits3 - huffman_table_bits)) |
| & huffman_mask]; |
| |
| if (use4) |
| { |
| if (unlikely (bits4 < huffman_table_bits)) |
| { |
| t4 = huffman_table[(val4 << (huffman_table_bits - bits4)) |
| & huffman_mask]; |
| if (unlikely (bits4 < (t4 & 0xff))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| else |
| t4 = huffman_table[(val4 >> (bits4 - huffman_table_bits)) |
| & huffman_mask]; |
| |
| *plit4 = t4 >> 8; |
| ++plit4; |
| bits4 -= t4 & 0xff; |
| } |
| |
| *plit1 = t1 >> 8; |
| ++plit1; |
| bits1 -= t1 & 0xff; |
| |
| *plit2 = t2 >> 8; |
| ++plit2; |
| bits2 -= t2 & 0xff; |
| |
| *plit3 = t3 >> 8; |
| ++plit3; |
| bits3 -= t3 & 0xff; |
| |
| ++i; |
| } |
| } |
| |
| return 1; |
| } |
| |
| /* The information used to decompress a sequence code, which can be a literal |
| length, an offset, or a match length. */ |
| |
| struct elf_zstd_seq_decode |
| { |
| const struct elf_zstd_fse_baseline_entry *table; |
| int table_bits; |
| }; |
| |
| /* Unpack a sequence code compression mode. */ |
| |
| static int |
| elf_zstd_unpack_seq_decode (int mode, |
| const unsigned char **ppin, |
| const unsigned char *pinend, |
| const struct elf_zstd_fse_baseline_entry *predef, |
| int predef_bits, |
| uint16_t *scratch, |
| int maxidx, |
| struct elf_zstd_fse_baseline_entry *table, |
| int table_bits, |
| int (*conv)(const struct elf_zstd_fse_entry *, |
| int, |
| struct elf_zstd_fse_baseline_entry *), |
| struct elf_zstd_seq_decode *decode) |
| { |
| switch (mode) |
| { |
| case 0: |
| decode->table = predef; |
| decode->table_bits = predef_bits; |
| break; |
| |
| case 1: |
| { |
| struct elf_zstd_fse_entry entry; |
| |
| if (unlikely (*ppin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| entry.symbol = **ppin; |
| ++*ppin; |
| entry.bits = 0; |
| entry.base = 0; |
| decode->table_bits = 0; |
| if (!conv (&entry, 0, table)) |
| return 0; |
| } |
| break; |
| |
| case 2: |
| { |
| struct elf_zstd_fse_entry *fse_table; |
| |
| /* We use the same space for the simple FSE table and the baseline |
| table. */ |
| fse_table = (struct elf_zstd_fse_entry *)table; |
| decode->table_bits = table_bits; |
| if (!elf_zstd_read_fse (ppin, pinend, scratch, maxidx, fse_table, |
| &decode->table_bits)) |
| return 0; |
| if (!conv (fse_table, decode->table_bits, table)) |
| return 0; |
| decode->table = table; |
| } |
| break; |
| |
| case 3: |
| if (unlikely (decode->table_bits == -1)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| break; |
| |
| default: |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| return 1; |
| } |
| |
| /* Decompress a zstd stream from PIN/SIN to POUT/SOUT. Code based on RFC 8878. |
| Return 1 on success, 0 on error. */ |
| |
| static int |
| elf_zstd_decompress (const unsigned char *pin, size_t sin, |
| unsigned char *zdebug_table, unsigned char *pout, |
| size_t sout) |
| { |
| const unsigned char *pinend; |
| unsigned char *poutstart; |
| unsigned char *poutend; |
| struct elf_zstd_seq_decode literal_decode; |
| struct elf_zstd_fse_baseline_entry *literal_fse_table; |
| struct elf_zstd_seq_decode match_decode; |
| struct elf_zstd_fse_baseline_entry *match_fse_table; |
| struct elf_zstd_seq_decode offset_decode; |
| struct elf_zstd_fse_baseline_entry *offset_fse_table; |
| uint16_t *huffman_table; |
| int huffman_table_bits; |
| uint32_t repeated_offset1; |
| uint32_t repeated_offset2; |
| uint32_t repeated_offset3; |
| uint16_t *scratch; |
| unsigned char hdr; |
| int has_checksum; |
| uint64_t content_size; |
| int last_block; |
| |
| pinend = pin + sin; |
| poutstart = pout; |
| poutend = pout + sout; |
| |
| literal_decode.table = NULL; |
| literal_decode.table_bits = -1; |
| literal_fse_table = ((struct elf_zstd_fse_baseline_entry *) |
| (zdebug_table + ZSTD_TABLE_LITERAL_FSE_OFFSET)); |
| |
| match_decode.table = NULL; |
| match_decode.table_bits = -1; |
| match_fse_table = ((struct elf_zstd_fse_baseline_entry *) |
| (zdebug_table + ZSTD_TABLE_MATCH_FSE_OFFSET)); |
| |
| offset_decode.table = NULL; |
| offset_decode.table_bits = -1; |
| offset_fse_table = ((struct elf_zstd_fse_baseline_entry *) |
| (zdebug_table + ZSTD_TABLE_OFFSET_FSE_OFFSET)); |
| huffman_table = ((uint16_t *) |
| (zdebug_table + ZSTD_TABLE_HUFFMAN_OFFSET)); |
| huffman_table_bits = 0; |
| scratch = ((uint16_t *) |
| (zdebug_table + ZSTD_TABLE_WORK_OFFSET)); |
| |
| repeated_offset1 = 1; |
| repeated_offset2 = 4; |
| repeated_offset3 = 8; |
| |
| if (unlikely (sin < 4)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* These values are the zstd magic number. */ |
| if (unlikely (pin[0] != 0x28 |
| || pin[1] != 0xb5 |
| || pin[2] != 0x2f |
| || pin[3] != 0xfd)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| pin += 4; |
| |
| if (unlikely (pin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| hdr = *pin++; |
| |
| /* We expect a single frame. */ |
| if (unlikely ((hdr & (1 << 5)) == 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| /* Reserved bit must be zero. */ |
| if (unlikely ((hdr & (1 << 3)) != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| /* We do not expect a dictionary. */ |
| if (unlikely ((hdr & 3) != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| has_checksum = (hdr & (1 << 2)) != 0; |
| switch (hdr >> 6) |
| { |
| case 0: |
| if (unlikely (pin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| content_size = (uint64_t) *pin++; |
| break; |
| case 1: |
| if (unlikely (pin + 1 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| content_size = (((uint64_t) pin[0]) | (((uint64_t) pin[1]) << 8)) + 256; |
| pin += 2; |
| break; |
| case 2: |
| if (unlikely (pin + 3 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| content_size = ((uint64_t) pin[0] |
| | (((uint64_t) pin[1]) << 8) |
| | (((uint64_t) pin[2]) << 16) |
| | (((uint64_t) pin[3]) << 24)); |
| pin += 4; |
| break; |
| case 3: |
| if (unlikely (pin + 7 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| content_size = ((uint64_t) pin[0] |
| | (((uint64_t) pin[1]) << 8) |
| | (((uint64_t) pin[2]) << 16) |
| | (((uint64_t) pin[3]) << 24) |
| | (((uint64_t) pin[4]) << 32) |
| | (((uint64_t) pin[5]) << 40) |
| | (((uint64_t) pin[6]) << 48) |
| | (((uint64_t) pin[7]) << 56)); |
| pin += 8; |
| break; |
| default: |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely (content_size != (size_t) content_size |
| || (size_t) content_size != sout)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| last_block = 0; |
| while (!last_block) |
| { |
| uint32_t block_hdr; |
| int block_type; |
| uint32_t block_size; |
| |
| if (unlikely (pin + 2 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| block_hdr = ((uint32_t) pin[0] |
| | (((uint32_t) pin[1]) << 8) |
| | (((uint32_t) pin[2]) << 16)); |
| pin += 3; |
| |
| last_block = block_hdr & 1; |
| block_type = (block_hdr >> 1) & 3; |
| block_size = block_hdr >> 3; |
| |
| switch (block_type) |
| { |
| case 0: |
| /* Raw_Block */ |
| if (unlikely ((size_t) block_size > (size_t) (pinend - pin))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| if (unlikely ((size_t) block_size > (size_t) (poutend - pout))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| memcpy (pout, pin, block_size); |
| pout += block_size; |
| pin += block_size; |
| break; |
| |
| case 1: |
| /* RLE_Block */ |
| if (unlikely (pin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| if (unlikely ((size_t) block_size > (size_t) (poutend - pout))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| memset (pout, *pin, block_size); |
| pout += block_size; |
| pin++; |
| break; |
| |
| case 2: |
| { |
| const unsigned char *pblockend; |
| unsigned char *plitstack; |
| unsigned char *plit; |
| uint32_t literal_count; |
| unsigned char seq_hdr; |
| size_t seq_count; |
| size_t seq; |
| const unsigned char *pback; |
| uint64_t val; |
| unsigned int bits; |
| unsigned int literal_state; |
| unsigned int offset_state; |
| unsigned int match_state; |
| |
| /* Compressed_Block */ |
| if (unlikely ((size_t) block_size > (size_t) (pinend - pin))) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| pblockend = pin + block_size; |
| |
| /* Read the literals into the end of the output space, and leave |
| PLIT pointing at them. */ |
| |
| if (!elf_zstd_read_literals (&pin, pblockend, pout, poutend, |
| scratch, huffman_table, |
| &huffman_table_bits, |
| &plitstack)) |
| return 0; |
| plit = plitstack; |
| literal_count = poutend - plit; |
| |
| seq_hdr = *pin; |
| pin++; |
| if (seq_hdr < 128) |
| seq_count = seq_hdr; |
| else if (seq_hdr < 255) |
| { |
| if (unlikely (pin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| seq_count = ((seq_hdr - 128) << 8) + *pin; |
| pin++; |
| } |
| else |
| { |
| if (unlikely (pin + 1 >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| seq_count = *pin + (pin[1] << 8) + 0x7f00; |
| pin += 2; |
| } |
| |
| if (seq_count > 0) |
| { |
| int (*pfn)(const struct elf_zstd_fse_entry *, |
| int, struct elf_zstd_fse_baseline_entry *); |
| |
| if (unlikely (pin >= pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| seq_hdr = *pin; |
| ++pin; |
| |
| pfn = elf_zstd_make_literal_baseline_fse; |
| if (!elf_zstd_unpack_seq_decode ((seq_hdr >> 6) & 3, |
| &pin, pinend, |
| &elf_zstd_lit_table[0], 6, |
| scratch, 35, |
| literal_fse_table, 9, pfn, |
| &literal_decode)) |
| return 0; |
| |
| pfn = elf_zstd_make_offset_baseline_fse; |
| if (!elf_zstd_unpack_seq_decode ((seq_hdr >> 4) & 3, |
| &pin, pinend, |
| &elf_zstd_offset_table[0], 5, |
| scratch, 31, |
| offset_fse_table, 8, pfn, |
| &offset_decode)) |
| return 0; |
| |
| pfn = elf_zstd_make_match_baseline_fse; |
| if (!elf_zstd_unpack_seq_decode ((seq_hdr >> 2) & 3, |
| &pin, pinend, |
| &elf_zstd_match_table[0], 6, |
| scratch, 52, |
| match_fse_table, 9, pfn, |
| &match_decode)) |
| return 0; |
| } |
| |
| pback = pblockend - 1; |
| if (!elf_fetch_backward_init (&pback, pin, &val, &bits)) |
| return 0; |
| |
| bits -= literal_decode.table_bits; |
| literal_state = ((val >> bits) |
| & ((1U << literal_decode.table_bits) - 1)); |
| |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| bits -= offset_decode.table_bits; |
| offset_state = ((val >> bits) |
| & ((1U << offset_decode.table_bits) - 1)); |
| |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| bits -= match_decode.table_bits; |
| match_state = ((val >> bits) |
| & ((1U << match_decode.table_bits) - 1)); |
| |
| seq = 0; |
| while (1) |
| { |
| const struct elf_zstd_fse_baseline_entry *pt; |
| uint32_t offset_basebits; |
| uint32_t offset_baseline; |
| uint32_t offset_bits; |
| uint32_t offset_base; |
| uint32_t offset; |
| uint32_t match_baseline; |
| uint32_t match_bits; |
| uint32_t match_base; |
| uint32_t match; |
| uint32_t literal_baseline; |
| uint32_t literal_bits; |
| uint32_t literal_base; |
| uint32_t literal; |
| uint32_t need; |
| uint32_t add; |
| |
| pt = &offset_decode.table[offset_state]; |
| offset_basebits = pt->basebits; |
| offset_baseline = pt->baseline; |
| offset_bits = pt->bits; |
| offset_base = pt->base; |
| |
| /* This case can be more than 16 bits, which is all that |
| elf_fetch_bits_backward promises. */ |
| need = offset_basebits; |
| add = 0; |
| if (unlikely (need > 16)) |
| { |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| bits -= 16; |
| add = (val >> bits) & ((1U << 16) - 1); |
| need -= 16; |
| add <<= need; |
| } |
| if (need > 0) |
| { |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| bits -= need; |
| add += (val >> bits) & ((1U << need) - 1); |
| } |
| |
| offset = offset_baseline + add; |
| |
| pt = &match_decode.table[match_state]; |
| need = pt->basebits; |
| match_baseline = pt->baseline; |
| match_bits = pt->bits; |
| match_base = pt->base; |
| |
| add = 0; |
| if (need > 0) |
| { |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| bits -= need; |
| add = (val >> bits) & ((1U << need) - 1); |
| } |
| |
| match = match_baseline + add; |
| |
| pt = &literal_decode.table[literal_state]; |
| need = pt->basebits; |
| literal_baseline = pt->baseline; |
| literal_bits = pt->bits; |
| literal_base = pt->base; |
| |
| add = 0; |
| if (need > 0) |
| { |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| bits -= need; |
| add = (val >> bits) & ((1U << need) - 1); |
| } |
| |
| literal = literal_baseline + add; |
| |
| /* See the comment in elf_zstd_make_offset_baseline_fse. */ |
| if (offset_basebits > 1) |
| { |
| repeated_offset3 = repeated_offset2; |
| repeated_offset2 = repeated_offset1; |
| repeated_offset1 = offset; |
| } |
| else |
| { |
| if (unlikely (literal == 0)) |
| ++offset; |
| switch (offset) |
| { |
| case 1: |
| offset = repeated_offset1; |
| break; |
| case 2: |
| offset = repeated_offset2; |
| repeated_offset2 = repeated_offset1; |
| repeated_offset1 = offset; |
| break; |
| case 3: |
| offset = repeated_offset3; |
| repeated_offset3 = repeated_offset2; |
| repeated_offset2 = repeated_offset1; |
| repeated_offset1 = offset; |
| break; |
| case 4: |
| offset = repeated_offset1 - 1; |
| repeated_offset3 = repeated_offset2; |
| repeated_offset2 = repeated_offset1; |
| repeated_offset1 = offset; |
| break; |
| } |
| } |
| |
| ++seq; |
| if (seq < seq_count) |
| { |
| uint32_t v; |
| |
| /* Update the three states. */ |
| |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| |
| need = literal_bits; |
| bits -= need; |
| v = (val >> bits) & (((uint32_t)1 << need) - 1); |
| |
| literal_state = literal_base + v; |
| |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| |
| need = match_bits; |
| bits -= need; |
| v = (val >> bits) & (((uint32_t)1 << need) - 1); |
| |
| match_state = match_base + v; |
| |
| if (!elf_fetch_bits_backward (&pback, pin, &val, &bits)) |
| return 0; |
| |
| need = offset_bits; |
| bits -= need; |
| v = (val >> bits) & (((uint32_t)1 << need) - 1); |
| |
| offset_state = offset_base + v; |
| } |
| |
| /* The next sequence is now in LITERAL, OFFSET, MATCH. */ |
| |
| /* Copy LITERAL bytes from the literals. */ |
| |
| if (unlikely ((size_t)(poutend - pout) < literal)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely (literal_count < literal)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| literal_count -= literal; |
| |
| /* Often LITERAL is small, so handle small cases quickly. */ |
| switch (literal) |
| { |
| case 8: |
| *pout++ = *plit++; |
| /* FALLTHROUGH */ |
| case 7: |
| *pout++ = *plit++; |
| /* FALLTHROUGH */ |
| case 6: |
| *pout++ = *plit++; |
| /* FALLTHROUGH */ |
| case 5: |
| *pout++ = *plit++; |
| /* FALLTHROUGH */ |
| case 4: |
| *pout++ = *plit++; |
| /* FALLTHROUGH */ |
| case 3: |
| *pout++ = *plit++; |
| /* FALLTHROUGH */ |
| case 2: |
| *pout++ = *plit++; |
| /* FALLTHROUGH */ |
| case 1: |
| *pout++ = *plit++; |
| break; |
| |
| case 0: |
| break; |
| |
| default: |
| if (unlikely ((size_t)(plit - pout) < literal)) |
| { |
| uint32_t move; |
| |
| move = plit - pout; |
| while (literal > move) |
| { |
| memcpy (pout, plit, move); |
| pout += move; |
| plit += move; |
| literal -= move; |
| } |
| } |
| |
| memcpy (pout, plit, literal); |
| pout += literal; |
| plit += literal; |
| } |
| |
| if (match > 0) |
| { |
| /* Copy MATCH bytes from the decoded output at OFFSET. */ |
| |
| if (unlikely ((size_t)(poutend - pout) < match)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely ((size_t)(pout - poutstart) < offset)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (offset >= match) |
| { |
| memcpy (pout, pout - offset, match); |
| pout += match; |
| } |
| else |
| { |
| while (match > 0) |
| { |
| uint32_t copy; |
| |
| copy = match < offset ? match : offset; |
| memcpy (pout, pout - offset, copy); |
| match -= copy; |
| pout += copy; |
| } |
| } |
| } |
| |
| if (unlikely (seq >= seq_count)) |
| { |
| /* Copy remaining literals. */ |
| if (literal_count > 0 && plit != pout) |
| { |
| if (unlikely ((size_t)(poutend - pout) |
| < literal_count)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if ((size_t)(plit - pout) < literal_count) |
| { |
| uint32_t move; |
| |
| move = plit - pout; |
| while (literal_count > move) |
| { |
| memcpy (pout, plit, move); |
| pout += move; |
| plit += move; |
| literal_count -= move; |
| } |
| } |
| |
| memcpy (pout, plit, literal_count); |
| } |
| |
| pout += literal_count; |
| |
| break; |
| } |
| } |
| |
| pin = pblockend; |
| } |
| break; |
| |
| case 3: |
| default: |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| |
| if (has_checksum) |
| { |
| if (unlikely (pin + 4 > pinend)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* We don't currently verify the checksum. Currently running GNU ld with |
| --compress-debug-sections=zstd does not seem to generate a |
| checksum. */ |
| |
| pin += 4; |
| } |
| |
| if (pin != pinend) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| return 1; |
| } |
| |
| #define ZDEBUG_TABLE_SIZE \ |
| (ZLIB_TABLE_SIZE > ZSTD_TABLE_SIZE ? ZLIB_TABLE_SIZE : ZSTD_TABLE_SIZE) |
| |
| /* Uncompress the old compressed debug format, the one emitted by |
| --compress-debug-sections=zlib-gnu. The compressed data is in |
| COMPRESSED / COMPRESSED_SIZE, and the function writes to |
| *UNCOMPRESSED / *UNCOMPRESSED_SIZE. ZDEBUG_TABLE is work space to |
| hold Huffman tables. Returns 0 on error, 1 on successful |
| decompression or if something goes wrong. In general we try to |
| carry on, by returning 1, even if we can't decompress. */ |
| |
| static int |
| elf_uncompress_zdebug (struct backtrace_state *state, |
| const unsigned char *compressed, size_t compressed_size, |
| uint16_t *zdebug_table, |
| backtrace_error_callback error_callback, void *data, |
| unsigned char **uncompressed, size_t *uncompressed_size) |
| { |
| size_t sz; |
| size_t i; |
| unsigned char *po; |
| |
| *uncompressed = NULL; |
| *uncompressed_size = 0; |
| |
| /* The format starts with the four bytes ZLIB, followed by the 8 |
| byte length of the uncompressed data in big-endian order, |
| followed by a zlib stream. */ |
| |
| if (compressed_size < 12 || memcmp (compressed, "ZLIB", 4) != 0) |
| return 1; |
| |
| sz = 0; |
| for (i = 0; i < 8; i++) |
| sz = (sz << 8) | compressed[i + 4]; |
| |
| if (*uncompressed != NULL && *uncompressed_size >= sz) |
| po = *uncompressed; |
| else |
| { |
| po = (unsigned char *) backtrace_alloc (state, sz, error_callback, data); |
| if (po == NULL) |
| return 0; |
| } |
| |
| if (!elf_zlib_inflate_and_verify (compressed + 12, compressed_size - 12, |
| zdebug_table, po, sz)) |
| return 1; |
| |
| *uncompressed = po; |
| *uncompressed_size = sz; |
| |
| return 1; |
| } |
| |
| /* Uncompress the new compressed debug format, the official standard |
| ELF approach emitted by --compress-debug-sections=zlib-gabi. The |
| compressed data is in COMPRESSED / COMPRESSED_SIZE, and the |
| function writes to *UNCOMPRESSED / *UNCOMPRESSED_SIZE. |
| ZDEBUG_TABLE is work space as for elf_uncompress_zdebug. Returns 0 |
| on error, 1 on successful decompression or if something goes wrong. |
| In general we try to carry on, by returning 1, even if we can't |
| decompress. */ |
| |
| static int |
| elf_uncompress_chdr (struct backtrace_state *state, |
| const unsigned char *compressed, size_t compressed_size, |
| uint16_t *zdebug_table, |
| backtrace_error_callback error_callback, void *data, |
| unsigned char **uncompressed, size_t *uncompressed_size) |
| { |
| b_elf_chdr chdr; |
| char *alc; |
| size_t alc_len; |
| unsigned char *po; |
| |
| *uncompressed = NULL; |
| *uncompressed_size = 0; |
| |
| /* The format starts with an ELF compression header. */ |
| if (compressed_size < sizeof (b_elf_chdr)) |
| return 1; |
| |
| /* The lld linker can misalign a compressed section, so we can't safely read |
| the fields directly as we can for other ELF sections. See |
| https://github.com/ianlancetaylor/libbacktrace/pull/120. */ |
| memcpy (&chdr, compressed, sizeof (b_elf_chdr)); |
| |
| alc = NULL; |
| alc_len = 0; |
| if (*uncompressed != NULL && *uncompressed_size >= chdr.ch_size) |
| po = *uncompressed; |
| else |
| { |
| alc_len = chdr.ch_size; |
| alc = backtrace_alloc (state, alc_len, error_callback, data); |
| if (alc == NULL) |
| return 0; |
| po = (unsigned char *) alc; |
| } |
| |
| switch (chdr.ch_type) |
| { |
| case ELFCOMPRESS_ZLIB: |
| if (!elf_zlib_inflate_and_verify (compressed + sizeof (b_elf_chdr), |
| compressed_size - sizeof (b_elf_chdr), |
| zdebug_table, po, chdr.ch_size)) |
| goto skip; |
| break; |
| |
| case ELFCOMPRESS_ZSTD: |
| if (!elf_zstd_decompress (compressed + sizeof (b_elf_chdr), |
| compressed_size - sizeof (b_elf_chdr), |
| (unsigned char *)zdebug_table, po, |
| chdr.ch_size)) |
| goto skip; |
| break; |
| |
| default: |
| /* Unsupported compression algorithm. */ |
| goto skip; |
| } |
| |
| *uncompressed = po; |
| *uncompressed_size = chdr.ch_size; |
| |
| return 1; |
| |
| skip: |
| if (alc != NULL && alc_len > 0) |
| backtrace_free (state, alc, alc_len, error_callback, data); |
| return 1; |
| } |
| |
| /* This function is a hook for testing the zlib support. It is only |
| used by tests. */ |
| |
| int |
| backtrace_uncompress_zdebug (struct backtrace_state *state, |
| const unsigned char *compressed, |
| size_t compressed_size, |
| backtrace_error_callback error_callback, |
| void *data, unsigned char **uncompressed, |
| size_t *uncompressed_size) |
| { |
| uint16_t *zdebug_table; |
| int ret; |
| |
| zdebug_table = ((uint16_t *) backtrace_alloc (state, ZDEBUG_TABLE_SIZE, |
| error_callback, data)); |
| if (zdebug_table == NULL) |
| return 0; |
| ret = elf_uncompress_zdebug (state, compressed, compressed_size, |
| zdebug_table, error_callback, data, |
| uncompressed, uncompressed_size); |
| backtrace_free (state, zdebug_table, ZDEBUG_TABLE_SIZE, |
| error_callback, data); |
| return ret; |
| } |
| |
| /* This function is a hook for testing the zstd support. It is only used by |
| tests. */ |
| |
| int |
| backtrace_uncompress_zstd (struct backtrace_state *state, |
| const unsigned char *compressed, |
| size_t compressed_size, |
| backtrace_error_callback error_callback, |
| void *data, unsigned char *uncompressed, |
| size_t uncompressed_size) |
| { |
| unsigned char *zdebug_table; |
| int ret; |
| |
| zdebug_table = ((unsigned char *) backtrace_alloc (state, ZDEBUG_TABLE_SIZE, |
| error_callback, data)); |
| if (zdebug_table == NULL) |
| return 0; |
| ret = elf_zstd_decompress (compressed, compressed_size, |
| zdebug_table, uncompressed, uncompressed_size); |
| backtrace_free (state, zdebug_table, ZDEBUG_TABLE_SIZE, |
| error_callback, data); |
| return ret; |
| } |
| |
| /* Number of LZMA states. */ |
| #define LZMA_STATES (12) |
| |
| /* Number of LZMA position states. The pb value of the property byte |
| is the number of bits to include in these states, and the maximum |
| value of pb is 4. */ |
| #define LZMA_POS_STATES (16) |
| |
| /* Number of LZMA distance states. These are used match distances |
| with a short match length: up to 4 bytes. */ |
| #define LZMA_DIST_STATES (4) |
| |
| /* Number of LZMA distance slots. LZMA uses six bits to encode larger |
| match lengths, so 1 << 6 possible probabilities. */ |
| #define LZMA_DIST_SLOTS (64) |
| |
| /* LZMA distances 0 to 3 are encoded directly, larger values use a |
| probability model. */ |
| #define LZMA_DIST_MODEL_START (4) |
| |
| /* The LZMA probability model ends at 14. */ |
| #define LZMA_DIST_MODEL_END (14) |
| |
| /* LZMA distance slots for distances less than 127. */ |
| #define LZMA_FULL_DISTANCES (128) |
| |
| /* LZMA uses four alignment bits. */ |
| #define LZMA_ALIGN_SIZE (16) |
| |
| /* LZMA match length is encoded with 4, 5, or 10 bits, some of which |
| are already known. */ |
| #define LZMA_LEN_LOW_SYMBOLS (8) |
| #define LZMA_LEN_MID_SYMBOLS (8) |
| #define LZMA_LEN_HIGH_SYMBOLS (256) |
| |
| /* LZMA literal encoding. */ |
| #define LZMA_LITERAL_CODERS_MAX (16) |
| #define LZMA_LITERAL_CODER_SIZE (0x300) |
| |
| /* LZMA is based on a large set of probabilities, each managed |
| independently. Each probability is an 11 bit number that we store |
| in a uint16_t. We use a single large array of probabilities. */ |
| |
| /* Lengths of entries in the LZMA probabilities array. The names used |
| here are copied from the Linux kernel implementation. */ |
| |
| #define LZMA_PROB_IS_MATCH_LEN (LZMA_STATES * LZMA_POS_STATES) |
| #define LZMA_PROB_IS_REP_LEN LZMA_STATES |
| #define LZMA_PROB_IS_REP0_LEN LZMA_STATES |
| #define LZMA_PROB_IS_REP1_LEN LZMA_STATES |
| #define LZMA_PROB_IS_REP2_LEN LZMA_STATES |
| #define LZMA_PROB_IS_REP0_LONG_LEN (LZMA_STATES * LZMA_POS_STATES) |
| #define LZMA_PROB_DIST_SLOT_LEN (LZMA_DIST_STATES * LZMA_DIST_SLOTS) |
| #define LZMA_PROB_DIST_SPECIAL_LEN (LZMA_FULL_DISTANCES - LZMA_DIST_MODEL_END) |
| #define LZMA_PROB_DIST_ALIGN_LEN LZMA_ALIGN_SIZE |
| #define LZMA_PROB_MATCH_LEN_CHOICE_LEN 1 |
| #define LZMA_PROB_MATCH_LEN_CHOICE2_LEN 1 |
| #define LZMA_PROB_MATCH_LEN_LOW_LEN (LZMA_POS_STATES * LZMA_LEN_LOW_SYMBOLS) |
| #define LZMA_PROB_MATCH_LEN_MID_LEN (LZMA_POS_STATES * LZMA_LEN_MID_SYMBOLS) |
| #define LZMA_PROB_MATCH_LEN_HIGH_LEN LZMA_LEN_HIGH_SYMBOLS |
| #define LZMA_PROB_REP_LEN_CHOICE_LEN 1 |
| #define LZMA_PROB_REP_LEN_CHOICE2_LEN 1 |
| #define LZMA_PROB_REP_LEN_LOW_LEN (LZMA_POS_STATES * LZMA_LEN_LOW_SYMBOLS) |
| #define LZMA_PROB_REP_LEN_MID_LEN (LZMA_POS_STATES * LZMA_LEN_MID_SYMBOLS) |
| #define LZMA_PROB_REP_LEN_HIGH_LEN LZMA_LEN_HIGH_SYMBOLS |
| #define LZMA_PROB_LITERAL_LEN \ |
| (LZMA_LITERAL_CODERS_MAX * LZMA_LITERAL_CODER_SIZE) |
| |
| /* Offsets into the LZMA probabilities array. This is mechanically |
| generated from the above lengths. */ |
| |
| #define LZMA_PROB_IS_MATCH_OFFSET 0 |
| #define LZMA_PROB_IS_REP_OFFSET \ |
| (LZMA_PROB_IS_MATCH_OFFSET + LZMA_PROB_IS_MATCH_LEN) |
| #define LZMA_PROB_IS_REP0_OFFSET \ |
| (LZMA_PROB_IS_REP_OFFSET + LZMA_PROB_IS_REP_LEN) |
| #define LZMA_PROB_IS_REP1_OFFSET \ |
| (LZMA_PROB_IS_REP0_OFFSET + LZMA_PROB_IS_REP0_LEN) |
| #define LZMA_PROB_IS_REP2_OFFSET \ |
| (LZMA_PROB_IS_REP1_OFFSET + LZMA_PROB_IS_REP1_LEN) |
| #define LZMA_PROB_IS_REP0_LONG_OFFSET \ |
| (LZMA_PROB_IS_REP2_OFFSET + LZMA_PROB_IS_REP2_LEN) |
| #define LZMA_PROB_DIST_SLOT_OFFSET \ |
| (LZMA_PROB_IS_REP0_LONG_OFFSET + LZMA_PROB_IS_REP0_LONG_LEN) |
| #define LZMA_PROB_DIST_SPECIAL_OFFSET \ |
| (LZMA_PROB_DIST_SLOT_OFFSET + LZMA_PROB_DIST_SLOT_LEN) |
| #define LZMA_PROB_DIST_ALIGN_OFFSET \ |
| (LZMA_PROB_DIST_SPECIAL_OFFSET + LZMA_PROB_DIST_SPECIAL_LEN) |
| #define LZMA_PROB_MATCH_LEN_CHOICE_OFFSET \ |
| (LZMA_PROB_DIST_ALIGN_OFFSET + LZMA_PROB_DIST_ALIGN_LEN) |
| #define LZMA_PROB_MATCH_LEN_CHOICE2_OFFSET \ |
| (LZMA_PROB_MATCH_LEN_CHOICE_OFFSET + LZMA_PROB_MATCH_LEN_CHOICE_LEN) |
| #define LZMA_PROB_MATCH_LEN_LOW_OFFSET \ |
| (LZMA_PROB_MATCH_LEN_CHOICE2_OFFSET + LZMA_PROB_MATCH_LEN_CHOICE2_LEN) |
| #define LZMA_PROB_MATCH_LEN_MID_OFFSET \ |
| (LZMA_PROB_MATCH_LEN_LOW_OFFSET + LZMA_PROB_MATCH_LEN_LOW_LEN) |
| #define LZMA_PROB_MATCH_LEN_HIGH_OFFSET \ |
| (LZMA_PROB_MATCH_LEN_MID_OFFSET + LZMA_PROB_MATCH_LEN_MID_LEN) |
| #define LZMA_PROB_REP_LEN_CHOICE_OFFSET \ |
| (LZMA_PROB_MATCH_LEN_HIGH_OFFSET + LZMA_PROB_MATCH_LEN_HIGH_LEN) |
| #define LZMA_PROB_REP_LEN_CHOICE2_OFFSET \ |
| (LZMA_PROB_REP_LEN_CHOICE_OFFSET + LZMA_PROB_REP_LEN_CHOICE_LEN) |
| #define LZMA_PROB_REP_LEN_LOW_OFFSET \ |
| (LZMA_PROB_REP_LEN_CHOICE2_OFFSET + LZMA_PROB_REP_LEN_CHOICE2_LEN) |
| #define LZMA_PROB_REP_LEN_MID_OFFSET \ |
| (LZMA_PROB_REP_LEN_LOW_OFFSET + LZMA_PROB_REP_LEN_LOW_LEN) |
| #define LZMA_PROB_REP_LEN_HIGH_OFFSET \ |
| (LZMA_PROB_REP_LEN_MID_OFFSET + LZMA_PROB_REP_LEN_MID_LEN) |
| #define LZMA_PROB_LITERAL_OFFSET \ |
| (LZMA_PROB_REP_LEN_HIGH_OFFSET + LZMA_PROB_REP_LEN_HIGH_LEN) |
| |
| #define LZMA_PROB_TOTAL_COUNT \ |
| (LZMA_PROB_LITERAL_OFFSET + LZMA_PROB_LITERAL_LEN) |
| |
| /* Check that the number of LZMA probabilities is the same as the |
| Linux kernel implementation. */ |
| |
| #if LZMA_PROB_TOTAL_COUNT != 1846 + (1 << 4) * 0x300 |
| #error Wrong number of LZMA probabilities |
| #endif |
| |
| /* Expressions for the offset in the LZMA probabilities array of a |
| specific probability. */ |
| |
| #define LZMA_IS_MATCH(state, pos) \ |
| (LZMA_PROB_IS_MATCH_OFFSET + (state) * LZMA_POS_STATES + (pos)) |
| #define LZMA_IS_REP(state) \ |
| (LZMA_PROB_IS_REP_OFFSET + (state)) |
| #define LZMA_IS_REP0(state) \ |
| (LZMA_PROB_IS_REP0_OFFSET + (state)) |
| #define LZMA_IS_REP1(state) \ |
| (LZMA_PROB_IS_REP1_OFFSET + (state)) |
| #define LZMA_IS_REP2(state) \ |
| (LZMA_PROB_IS_REP2_OFFSET + (state)) |
| #define LZMA_IS_REP0_LONG(state, pos) \ |
| (LZMA_PROB_IS_REP0_LONG_OFFSET + (state) * LZMA_POS_STATES + (pos)) |
| #define LZMA_DIST_SLOT(dist, slot) \ |
| (LZMA_PROB_DIST_SLOT_OFFSET + (dist) * LZMA_DIST_SLOTS + (slot)) |
| #define LZMA_DIST_SPECIAL(dist) \ |
| (LZMA_PROB_DIST_SPECIAL_OFFSET + (dist)) |
| #define LZMA_DIST_ALIGN(dist) \ |
| (LZMA_PROB_DIST_ALIGN_OFFSET + (dist)) |
| #define LZMA_MATCH_LEN_CHOICE \ |
| LZMA_PROB_MATCH_LEN_CHOICE_OFFSET |
| #define LZMA_MATCH_LEN_CHOICE2 \ |
| LZMA_PROB_MATCH_LEN_CHOICE2_OFFSET |
| #define LZMA_MATCH_LEN_LOW(pos, sym) \ |
| (LZMA_PROB_MATCH_LEN_LOW_OFFSET + (pos) * LZMA_LEN_LOW_SYMBOLS + (sym)) |
| #define LZMA_MATCH_LEN_MID(pos, sym) \ |
| (LZMA_PROB_MATCH_LEN_MID_OFFSET + (pos) * LZMA_LEN_MID_SYMBOLS + (sym)) |
| #define LZMA_MATCH_LEN_HIGH(sym) \ |
| (LZMA_PROB_MATCH_LEN_HIGH_OFFSET + (sym)) |
| #define LZMA_REP_LEN_CHOICE \ |
| LZMA_PROB_REP_LEN_CHOICE_OFFSET |
| #define LZMA_REP_LEN_CHOICE2 \ |
| LZMA_PROB_REP_LEN_CHOICE2_OFFSET |
| #define LZMA_REP_LEN_LOW(pos, sym) \ |
| (LZMA_PROB_REP_LEN_LOW_OFFSET + (pos) * LZMA_LEN_LOW_SYMBOLS + (sym)) |
| #define LZMA_REP_LEN_MID(pos, sym) \ |
| (LZMA_PROB_REP_LEN_MID_OFFSET + (pos) * LZMA_LEN_MID_SYMBOLS + (sym)) |
| #define LZMA_REP_LEN_HIGH(sym) \ |
| (LZMA_PROB_REP_LEN_HIGH_OFFSET + (sym)) |
| #define LZMA_LITERAL(code, size) \ |
| (LZMA_PROB_LITERAL_OFFSET + (code) * LZMA_LITERAL_CODER_SIZE + (size)) |
| |
| /* Read an LZMA varint from BUF, reading and updating *POFFSET, |
| setting *VAL. Returns 0 on error, 1 on success. */ |
| |
| static int |
| elf_lzma_varint (const unsigned char *compressed, size_t compressed_size, |
| size_t *poffset, uint64_t *val) |
| { |
| size_t off; |
| int i; |
| uint64_t v; |
| unsigned char b; |
| |
| off = *poffset; |
| i = 0; |
| v = 0; |
| while (1) |
| { |
| if (unlikely (off >= compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| b = compressed[off]; |
| v |= (b & 0x7f) << (i * 7); |
| ++off; |
| if ((b & 0x80) == 0) |
| { |
| *poffset = off; |
| *val = v; |
| return 1; |
| } |
| ++i; |
| if (unlikely (i >= 9)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| } |
| |
| /* Normalize the LZMA range decoder, pulling in an extra input byte if |
| needed. */ |
| |
| static void |
| elf_lzma_range_normalize (const unsigned char *compressed, |
| size_t compressed_size, size_t *poffset, |
| uint32_t *prange, uint32_t *pcode) |
| { |
| if (*prange < (1U << 24)) |
| { |
| if (unlikely (*poffset >= compressed_size)) |
| { |
| /* We assume this will be caught elsewhere. */ |
| elf_uncompress_failed (); |
| return; |
| } |
| *prange <<= 8; |
| *pcode <<= 8; |
| *pcode += compressed[*poffset]; |
| ++*poffset; |
| } |
| } |
| |
| /* Read and return a single bit from the LZMA stream, reading and |
| updating *PROB. Each bit comes from the range coder. */ |
| |
| static int |
| elf_lzma_bit (const unsigned char *compressed, size_t compressed_size, |
| uint16_t *prob, size_t *poffset, uint32_t *prange, |
| uint32_t *pcode) |
| { |
| uint32_t bound; |
| |
| elf_lzma_range_normalize (compressed, compressed_size, poffset, |
| prange, pcode); |
| bound = (*prange >> 11) * (uint32_t) *prob; |
| if (*pcode < bound) |
| { |
| *prange = bound; |
| *prob += ((1U << 11) - *prob) >> 5; |
| return 0; |
| } |
| else |
| { |
| *prange -= bound; |
| *pcode -= bound; |
| *prob -= *prob >> 5; |
| return 1; |
| } |
| } |
| |
| /* Read an integer of size BITS from the LZMA stream, most significant |
| bit first. The bits are predicted using PROBS. */ |
| |
| static uint32_t |
| elf_lzma_integer (const unsigned char *compressed, size_t compressed_size, |
| uint16_t *probs, uint32_t bits, size_t *poffset, |
| uint32_t *prange, uint32_t *pcode) |
| { |
| uint32_t sym; |
| uint32_t i; |
| |
| sym = 1; |
| for (i = 0; i < bits; i++) |
| { |
| int bit; |
| |
| bit = elf_lzma_bit (compressed, compressed_size, probs + sym, poffset, |
| prange, pcode); |
| sym <<= 1; |
| sym += bit; |
| } |
| return sym - (1 << bits); |
| } |
| |
| /* Read an integer of size BITS from the LZMA stream, least |
| significant bit first. The bits are predicted using PROBS. */ |
| |
| static uint32_t |
| elf_lzma_reverse_integer (const unsigned char *compressed, |
| size_t compressed_size, uint16_t *probs, |
| uint32_t bits, size_t *poffset, uint32_t *prange, |
| uint32_t *pcode) |
| { |
| uint32_t sym; |
| uint32_t val; |
| uint32_t i; |
| |
| sym = 1; |
| val = 0; |
| for (i = 0; i < bits; i++) |
| { |
| int bit; |
| |
| bit = elf_lzma_bit (compressed, compressed_size, probs + sym, poffset, |
| prange, pcode); |
| sym <<= 1; |
| sym += bit; |
| val += bit << i; |
| } |
| return val; |
| } |
| |
| /* Read a length from the LZMA stream. IS_REP picks either LZMA_MATCH |
| or LZMA_REP probabilities. */ |
| |
| static uint32_t |
| elf_lzma_len (const unsigned char *compressed, size_t compressed_size, |
| uint16_t *probs, int is_rep, unsigned int pos_state, |
| size_t *poffset, uint32_t *prange, uint32_t *pcode) |
| { |
| uint16_t *probs_choice; |
| uint16_t *probs_sym; |
| uint32_t bits; |
| uint32_t len; |
| |
| probs_choice = probs + (is_rep |
| ? LZMA_REP_LEN_CHOICE |
| : LZMA_MATCH_LEN_CHOICE); |
| if (elf_lzma_bit (compressed, compressed_size, probs_choice, poffset, |
| prange, pcode)) |
| { |
| probs_choice = probs + (is_rep |
| ? LZMA_REP_LEN_CHOICE2 |
| : LZMA_MATCH_LEN_CHOICE2); |
| if (elf_lzma_bit (compressed, compressed_size, probs_choice, |
| poffset, prange, pcode)) |
| { |
| probs_sym = probs + (is_rep |
| ? LZMA_REP_LEN_HIGH (0) |
| : LZMA_MATCH_LEN_HIGH (0)); |
| bits = 8; |
| len = 2 + 8 + 8; |
| } |
| else |
| { |
| probs_sym = probs + (is_rep |
| ? LZMA_REP_LEN_MID (pos_state, 0) |
| : LZMA_MATCH_LEN_MID (pos_state, 0)); |
| bits = 3; |
| len = 2 + 8; |
| } |
| } |
| else |
| { |
| probs_sym = probs + (is_rep |
| ? LZMA_REP_LEN_LOW (pos_state, 0) |
| : LZMA_MATCH_LEN_LOW (pos_state, 0)); |
| bits = 3; |
| len = 2; |
| } |
| |
| len += elf_lzma_integer (compressed, compressed_size, probs_sym, bits, |
| poffset, prange, pcode); |
| return len; |
| } |
| |
| /* Uncompress one LZMA block from a minidebug file. The compressed |
| data is at COMPRESSED + *POFFSET. Update *POFFSET. Store the data |
| into the memory at UNCOMPRESSED, size UNCOMPRESSED_SIZE. CHECK is |
| the stream flag from the xz header. Return 1 on successful |
| decompression. */ |
| |
| static int |
| elf_uncompress_lzma_block (const unsigned char *compressed, |
| size_t compressed_size, unsigned char check, |
| uint16_t *probs, unsigned char *uncompressed, |
| size_t uncompressed_size, size_t *poffset) |
| { |
| size_t off; |
| size_t block_header_offset; |
| size_t block_header_size; |
| unsigned char block_flags; |
| uint64_t header_compressed_size; |
| uint64_t header_uncompressed_size; |
| unsigned char lzma2_properties; |
| size_t crc_offset; |
| uint32_t computed_crc; |
| uint32_t stream_crc; |
| size_t uncompressed_offset; |
| size_t dict_start_offset; |
| unsigned int lc; |
| unsigned int lp; |
| unsigned int pb; |
| uint32_t range; |
| uint32_t code; |
| uint32_t lstate; |
| uint32_t dist[4]; |
| |
| off = *poffset; |
| block_header_offset = off; |
| |
| /* Block header size is a single byte. */ |
| if (unlikely (off >= compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| block_header_size = (compressed[off] + 1) * 4; |
| if (unlikely (off + block_header_size > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Block flags. */ |
| block_flags = compressed[off + 1]; |
| if (unlikely ((block_flags & 0x3c) != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| off += 2; |
| |
| /* Optional compressed size. */ |
| header_compressed_size = 0; |
| if ((block_flags & 0x40) != 0) |
| { |
| *poffset = off; |
| if (!elf_lzma_varint (compressed, compressed_size, poffset, |
| &header_compressed_size)) |
| return 0; |
| off = *poffset; |
| } |
| |
| /* Optional uncompressed size. */ |
| header_uncompressed_size = 0; |
| if ((block_flags & 0x80) != 0) |
| { |
| *poffset = off; |
| if (!elf_lzma_varint (compressed, compressed_size, poffset, |
| &header_uncompressed_size)) |
| return 0; |
| off = *poffset; |
| } |
| |
| /* The recipe for creating a minidebug file is to run the xz program |
| with no arguments, so we expect exactly one filter: lzma2. */ |
| |
| if (unlikely ((block_flags & 0x3) != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (unlikely (off + 2 >= block_header_offset + block_header_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* The filter ID for LZMA2 is 0x21. */ |
| if (unlikely (compressed[off] != 0x21)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| ++off; |
| |
| /* The size of the filter properties for LZMA2 is 1. */ |
| if (unlikely (compressed[off] != 1)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| ++off; |
| |
| lzma2_properties = compressed[off]; |
| ++off; |
| |
| if (unlikely (lzma2_properties > 40)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* The properties describe the dictionary size, but we don't care |
| what that is. */ |
| |
| /* Skip to just before CRC, verifying zero bytes in between. */ |
| crc_offset = block_header_offset + block_header_size - 4; |
| if (unlikely (crc_offset + 4 > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| for (; off < crc_offset; off++) |
| { |
| if (compressed[off] != 0) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| |
| /* Block header CRC. */ |
| computed_crc = elf_crc32 (0, compressed + block_header_offset, |
| block_header_size - 4); |
| stream_crc = ((uint32_t)compressed[off] |
| | ((uint32_t)compressed[off + 1] << 8) |
| | ((uint32_t)compressed[off + 2] << 16) |
| | ((uint32_t)compressed[off + 3] << 24)); |
| if (unlikely (computed_crc != stream_crc)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| off += 4; |
| |
| /* Read a sequence of LZMA2 packets. */ |
| |
| uncompressed_offset = 0; |
| dict_start_offset = 0; |
| lc = 0; |
| lp = 0; |
| pb = 0; |
| lstate = 0; |
| while (off < compressed_size) |
| { |
| unsigned char control; |
| |
| range = 0xffffffff; |
| code = 0; |
| |
| control = compressed[off]; |
| ++off; |
| if (unlikely (control == 0)) |
| { |
| /* End of packets. */ |
| break; |
| } |
| |
| if (control == 1 || control >= 0xe0) |
| { |
| /* Reset dictionary to empty. */ |
| dict_start_offset = uncompressed_offset; |
| } |
| |
| if (control < 0x80) |
| { |
| size_t chunk_size; |
| |
| /* The only valid values here are 1 or 2. A 1 means to |
| reset the dictionary (done above). Then we see an |
| uncompressed chunk. */ |
| |
| if (unlikely (control > 2)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* An uncompressed chunk is a two byte size followed by |
| data. */ |
| |
| if (unlikely (off + 2 > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| chunk_size = compressed[off] << 8; |
| chunk_size += compressed[off + 1]; |
| ++chunk_size; |
| |
| off += 2; |
| |
| if (unlikely (off + chunk_size > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| if (unlikely (uncompressed_offset + chunk_size > uncompressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| memcpy (uncompressed + uncompressed_offset, compressed + off, |
| chunk_size); |
| uncompressed_offset += chunk_size; |
| off += chunk_size; |
| } |
| else |
| { |
| size_t uncompressed_chunk_start; |
| size_t uncompressed_chunk_size; |
| size_t compressed_chunk_size; |
| size_t limit; |
| |
| /* An LZMA chunk. This starts with an uncompressed size and |
| a compressed size. */ |
| |
| if (unlikely (off + 4 >= compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| uncompressed_chunk_start = uncompressed_offset; |
| |
| uncompressed_chunk_size = (control & 0x1f) << 16; |
| uncompressed_chunk_size += compressed[off] << 8; |
| uncompressed_chunk_size += compressed[off + 1]; |
| ++uncompressed_chunk_size; |
| |
| compressed_chunk_size = compressed[off + 2] << 8; |
| compressed_chunk_size += compressed[off + 3]; |
| ++compressed_chunk_size; |
| |
| off += 4; |
| |
| /* Bit 7 (0x80) is set. |
| Bits 6 and 5 (0x40 and 0x20) are as follows: |
| 0: don't reset anything |
| 1: reset state |
| 2: reset state, read properties |
| 3: reset state, read properties, reset dictionary (done above) */ |
| |
| if (control >= 0xc0) |
| { |
| unsigned char props; |
| |
| /* Bit 6 is set, read properties. */ |
| |
| if (unlikely (off >= compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| props = compressed[off]; |
| ++off; |
| if (unlikely (props > (4 * 5 + 4) * 9 + 8)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| pb = 0; |
| while (props >= 9 * 5) |
| { |
| props -= 9 * 5; |
| ++pb; |
| } |
| lp = 0; |
| while (props > 9) |
| { |
| props -= 9; |
| ++lp; |
| } |
| lc = props; |
| if (unlikely (lc + lp > 4)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| } |
| |
| if (control >= 0xa0) |
| { |
| size_t i; |
| |
| /* Bit 5 or 6 is set, reset LZMA state. */ |
| |
| lstate = 0; |
| memset (&dist, 0, sizeof dist); |
| for (i = 0; i < LZMA_PROB_TOTAL_COUNT; i++) |
| probs[i] = 1 << 10; |
| range = 0xffffffff; |
| code = 0; |
| } |
| |
| /* Read the range code. */ |
| |
| if (unlikely (off + 5 > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* The byte at compressed[off] is ignored for some |
| reason. */ |
| |
| code = ((compressed[off + 1] << 24) |
| + (compressed[off + 2] << 16) |
| + (compressed[off + 3] << 8) |
| + compressed[off + 4]); |
| off += 5; |
| |
| /* This is the main LZMA decode loop. */ |
| |
| limit = off + compressed_chunk_size; |
| *poffset = off; |
| while (*poffset < limit) |
| { |
| unsigned int pos_state; |
| |
| if (unlikely (uncompressed_offset |
| == (uncompressed_chunk_start |
| + uncompressed_chunk_size))) |
| { |
| /* We've decompressed all the expected bytes. */ |
| break; |
| } |
| |
| pos_state = ((uncompressed_offset - dict_start_offset) |
| & ((1 << pb) - 1)); |
| |
| if (elf_lzma_bit (compressed, compressed_size, |
| probs + LZMA_IS_MATCH (lstate, pos_state), |
| poffset, &range, &code)) |
| { |
| uint32_t len; |
| |
| if (elf_lzma_bit (compressed, compressed_size, |
| probs + LZMA_IS_REP (lstate), |
| poffset, &range, &code)) |
| { |
| int short_rep; |
| uint32_t next_dist; |
| |
| /* Repeated match. */ |
| |
| short_rep = 0; |
| if (elf_lzma_bit (compressed, compressed_size, |
| probs + LZMA_IS_REP0 (lstate), |
| poffset, &range, &code)) |
| { |
| if (elf_lzma_bit (compressed, compressed_size, |
| probs + LZMA_IS_REP1 (lstate), |
| poffset, &range, &code)) |
| { |
| if (elf_lzma_bit (compressed, compressed_size, |
| probs + LZMA_IS_REP2 (lstate), |
| poffset, &range, &code)) |
| { |
| next_dist = dist[3]; |
| dist[3] = dist[2]; |
| } |
| else |
| { |
| next_dist = dist[2]; |
| } |
| dist[2] = dist[1]; |
| } |
| else |
| { |
| next_dist = dist[1]; |
| } |
| |
| dist[1] = dist[0]; |
| dist[0] = next_dist; |
| } |
| else |
| { |
| if (!elf_lzma_bit (compressed, compressed_size, |
| (probs |
| + LZMA_IS_REP0_LONG (lstate, |
| pos_state)), |
| poffset, &range, &code)) |
| short_rep = 1; |
| } |
| |
| if (lstate < 7) |
| lstate = short_rep ? 9 : 8; |
| else |
| lstate = 11; |
| |
| if (short_rep) |
| len = 1; |
| else |
| len = elf_lzma_len (compressed, compressed_size, |
| probs, 1, pos_state, poffset, |
| &range, &code); |
| } |
| else |
| { |
| uint32_t dist_state; |
| uint32_t dist_slot; |
| uint16_t *probs_dist; |
| |
| /* Match. */ |
| |
| if (lstate < 7) |
| lstate = 7; |
| else |
| lstate = 10; |
| dist[3] = dist[2]; |
| dist[2] = dist[1]; |
| dist[1] = dist[0]; |
| len = elf_lzma_len (compressed, compressed_size, |
| probs, 0, pos_state, poffset, |
| &range, &code); |
| |
| if (len < 4 + 2) |
| dist_state = len - 2; |
| else |
| dist_state = 3; |
| probs_dist = probs + LZMA_DIST_SLOT (dist_state, 0); |
| dist_slot = elf_lzma_integer (compressed, |
| compressed_size, |
| probs_dist, 6, |
| poffset, &range, |
| &code); |
| if (dist_slot < LZMA_DIST_MODEL_START) |
| dist[0] = dist_slot; |
| else |
| { |
| uint32_t limit; |
| |
| limit = (dist_slot >> 1) - 1; |
| dist[0] = 2 + (dist_slot & 1); |
| if (dist_slot < LZMA_DIST_MODEL_END) |
| { |
| dist[0] <<= limit; |
| probs_dist = (probs |
| + LZMA_DIST_SPECIAL(dist[0] |
| - dist_slot |
| - 1)); |
| dist[0] += |
| elf_lzma_reverse_integer (compressed, |
| compressed_size, |
| probs_dist, |
| limit, poffset, |
| &range, &code); |
| } |
| else |
| { |
| uint32_t dist0; |
| uint32_t i; |
| |
| dist0 = dist[0]; |
| for (i = 0; i < limit - 4; i++) |
| { |
| uint32_t mask; |
| |
| elf_lzma_range_normalize (compressed, |
| compressed_size, |
| poffset, |
| &range, &code); |
| range >>= 1; |
| code -= range; |
| mask = -(code >> 31); |
| code += range & mask; |
| dist0 <<= 1; |
| dist0 += mask + 1; |
| } |
| dist0 <<= 4; |
| probs_dist = probs + LZMA_DIST_ALIGN (0); |
| dist0 += |
| elf_lzma_reverse_integer (compressed, |
| compressed_size, |
| probs_dist, 4, |
| poffset, |
| &range, &code); |
| dist[0] = dist0; |
| } |
| } |
| } |
| |
| if (unlikely (uncompressed_offset |
| - dict_start_offset < dist[0] + 1)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| if (unlikely (uncompressed_offset + len > uncompressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| if (dist[0] == 0) |
| { |
| /* A common case, meaning repeat the last |
| character LEN times. */ |
| memset (uncompressed + uncompressed_offset, |
| uncompressed[uncompressed_offset - 1], |
| len); |
| uncompressed_offset += len; |
| } |
| else if (dist[0] + 1 >= len) |
| { |
| memcpy (uncompressed + uncompressed_offset, |
| uncompressed + uncompressed_offset - dist[0] - 1, |
| len); |
| uncompressed_offset += len; |
| } |
| else |
| { |
| while (len > 0) |
| { |
| uint32_t copy; |
| |
| copy = len < dist[0] + 1 ? len : dist[0] + 1; |
| memcpy (uncompressed + uncompressed_offset, |
| (uncompressed + uncompressed_offset |
| - dist[0] - 1), |
| copy); |
| len -= copy; |
| uncompressed_offset += copy; |
| } |
| } |
| } |
| else |
| { |
| unsigned char prev; |
| unsigned char low; |
| size_t high; |
| uint16_t *lit_probs; |
| unsigned int sym; |
| |
| /* Literal value. */ |
| |
| if (uncompressed_offset > 0) |
| prev = uncompressed[uncompressed_offset - 1]; |
| else |
| prev = 0; |
| low = prev >> (8 - lc); |
| high = (((uncompressed_offset - dict_start_offset) |
| & ((1 << lp) - 1)) |
| << lc); |
| lit_probs = probs + LZMA_LITERAL (low + high, 0); |
| if (lstate < 7) |
| sym = elf_lzma_integer (compressed, compressed_size, |
| lit_probs, 8, poffset, &range, |
| &code); |
| else |
| { |
| unsigned int match; |
| unsigned int bit; |
| unsigned int match_bit; |
| unsigned int idx; |
| |
| sym = 1; |
| if (uncompressed_offset >= dist[0] + 1) |
| match = uncompressed[uncompressed_offset - dist[0] - 1]; |
| else |
| match = 0; |
| match <<= 1; |
| bit = 0x100; |
| do |
| { |
| match_bit = match & bit; |
| match <<= 1; |
| idx = bit + match_bit + sym; |
| sym <<= 1; |
| if (elf_lzma_bit (compressed, compressed_size, |
| lit_probs + idx, poffset, |
| &range, &code)) |
| { |
| ++sym; |
| bit &= match_bit; |
| } |
| else |
| { |
| bit &= ~ match_bit; |
| } |
| } |
| while (sym < 0x100); |
| } |
| |
| if (unlikely (uncompressed_offset >= uncompressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| uncompressed[uncompressed_offset] = (unsigned char) sym; |
| ++uncompressed_offset; |
| if (lstate <= 3) |
| lstate = 0; |
| else if (lstate <= 9) |
| lstate -= 3; |
| else |
| lstate -= 6; |
| } |
| } |
| |
| elf_lzma_range_normalize (compressed, compressed_size, poffset, |
| &range, &code); |
| |
| off = *poffset; |
| } |
| } |
| |
| /* We have reached the end of the block. Pad to four byte |
| boundary. */ |
| off = (off + 3) &~ (size_t) 3; |
| if (unlikely (off > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| switch (check) |
| { |
| case 0: |
| /* No check. */ |
| break; |
| |
| case 1: |
| /* CRC32 */ |
| if (unlikely (off + 4 > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| computed_crc = elf_crc32 (0, uncompressed, uncompressed_offset); |
| stream_crc = (compressed[off] |
| | (compressed[off + 1] << 8) |
| | (compressed[off + 2] << 16) |
| | (compressed[off + 3] << 24)); |
| if (computed_crc != stream_crc) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| off += 4; |
| break; |
| |
| case 4: |
| /* CRC64. We don't bother computing a CRC64 checksum. */ |
| if (unlikely (off + 8 > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| off += 8; |
| break; |
| |
| case 10: |
| /* SHA. We don't bother computing a SHA checksum. */ |
| if (unlikely (off + 32 > compressed_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| off += 32; |
| break; |
| |
| default: |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| *poffset = off; |
| |
| return 1; |
| } |
| |
| /* Uncompress LZMA data found in a minidebug file. The minidebug |
| format is described at |
| https://sourceware.org/gdb/current/onlinedocs/gdb/MiniDebugInfo.html. |
| Returns 0 on error, 1 on successful decompression. For this |
| function we return 0 on failure to decompress, as the calling code |
| will carry on in that case. */ |
| |
| static int |
| elf_uncompress_lzma (struct backtrace_state *state, |
| const unsigned char *compressed, size_t compressed_size, |
| backtrace_error_callback error_callback, void *data, |
| unsigned char **uncompressed, size_t *uncompressed_size) |
| { |
| size_t header_size; |
| size_t footer_size; |
| unsigned char check; |
| uint32_t computed_crc; |
| uint32_t stream_crc; |
| size_t offset; |
| size_t index_size; |
| size_t footer_offset; |
| size_t index_offset; |
| uint64_t index_compressed_size; |
| uint64_t index_uncompressed_size; |
| unsigned char *mem; |
| uint16_t *probs; |
| size_t compressed_block_size; |
| |
| /* The format starts with a stream header and ends with a stream |
| footer. */ |
| header_size = 12; |
| footer_size = 12; |
| if (unlikely (compressed_size < header_size + footer_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* The stream header starts with a magic string. */ |
| if (unlikely (memcmp (compressed, "\375" "7zXZ\0", 6) != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Next come stream flags. The first byte is zero, the second byte |
| is the check. */ |
| if (unlikely (compressed[6] != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| check = compressed[7]; |
| if (unlikely ((check & 0xf8) != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Next comes a CRC of the stream flags. */ |
| computed_crc = elf_crc32 (0, compressed + 6, 2); |
| stream_crc = ((uint32_t)compressed[8] |
| | ((uint32_t)compressed[9] << 8) |
| | ((uint32_t)compressed[10] << 16) |
| | ((uint32_t)compressed[11] << 24)); |
| if (unlikely (computed_crc != stream_crc)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Now that we've parsed the header, parse the footer, so that we |
| can get the uncompressed size. */ |
| |
| /* The footer ends with two magic bytes. */ |
| |
| offset = compressed_size; |
| if (unlikely (memcmp (compressed + offset - 2, "YZ", 2) != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| offset -= 2; |
| |
| /* Before that are the stream flags, which should be the same as the |
| flags in the header. */ |
| if (unlikely (compressed[offset - 2] != 0 |
| || compressed[offset - 1] != check)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| offset -= 2; |
| |
| /* Before that is the size of the index field, which precedes the |
| footer. */ |
| index_size = (compressed[offset - 4] |
| | (compressed[offset - 3] << 8) |
| | (compressed[offset - 2] << 16) |
| | (compressed[offset - 1] << 24)); |
| index_size = (index_size + 1) * 4; |
| offset -= 4; |
| |
| /* Before that is a footer CRC. */ |
| computed_crc = elf_crc32 (0, compressed + offset, 6); |
| stream_crc = ((uint32_t)compressed[offset - 4] |
| | ((uint32_t)compressed[offset - 3] << 8) |
| | ((uint32_t)compressed[offset - 2] << 16) |
| | ((uint32_t)compressed[offset - 1] << 24)); |
| if (unlikely (computed_crc != stream_crc)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| offset -= 4; |
| |
| /* The index comes just before the footer. */ |
| if (unlikely (offset < index_size + header_size)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| footer_offset = offset; |
| offset -= index_size; |
| index_offset = offset; |
| |
| /* The index starts with a zero byte. */ |
| if (unlikely (compressed[offset] != 0)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| ++offset; |
| |
| /* Next is the number of blocks. We expect zero blocks for an empty |
| stream, and otherwise a single block. */ |
| if (unlikely (compressed[offset] == 0)) |
| { |
| *uncompressed = NULL; |
| *uncompressed_size = 0; |
| return 1; |
| } |
| if (unlikely (compressed[offset] != 1)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| ++offset; |
| |
| /* Next is the compressed size and the uncompressed size. */ |
| if (!elf_lzma_varint (compressed, compressed_size, &offset, |
| &index_compressed_size)) |
| return 0; |
| if (!elf_lzma_varint (compressed, compressed_size, &offset, |
| &index_uncompressed_size)) |
| return 0; |
| |
| /* Pad to a four byte boundary. */ |
| offset = (offset + 3) &~ (size_t) 3; |
| |
| /* Next is a CRC of the index. */ |
| computed_crc = elf_crc32 (0, compressed + index_offset, |
| offset - index_offset); |
| stream_crc = ((uint32_t)compressed[offset] |
| | ((uint32_t)compressed[offset + 1] << 8) |
| | ((uint32_t)compressed[offset + 2] << 16) |
| | ((uint32_t)compressed[offset + 3] << 24)); |
| if (unlikely (computed_crc != stream_crc)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| offset += 4; |
| |
| /* We should now be back at the footer. */ |
| if (unlikely (offset != footer_offset)) |
| { |
| elf_uncompress_failed (); |
| return 0; |
| } |
| |
| /* Allocate space to hold the uncompressed data. If we succeed in |
| uncompressing the LZMA data, we never free this memory. */ |
| mem = (unsigned char *) backtrace_alloc (state, index_uncompressed_size, |
| error_callback, data); |
| if (unlikely (mem == NULL)) |
| return 0; |
| *uncompressed = mem; |
| *uncompressed_size = index_uncompressed_size; |
| |
| /* Allocate space for probabilities. */ |
| probs = ((uint16_t *) |
| backtrace_alloc (state, |
| LZMA_PROB_TOTAL_COUNT * sizeof (uint16_t), |
| error_callback, data)); |
| if (unlikely (probs == NULL)) |
| { |
| backtrace_free (state, mem, index_uncompressed_size, error_callback, |
| data); |
| return 0; |
| } |
| |
| /* Uncompress the block, which follows the header. */ |
| offset = 12; |
| if (!elf_uncompress_lzma_block (compressed, compressed_size, check, probs, |
| mem, index_uncompressed_size, &offset)) |
| { |
| backtrace_free (state, mem, index_uncompressed_size, error_callback, |
| data); |
| return 0; |
| } |
| |
| compressed_block_size = offset - 12; |
| if (unlikely (compressed_block_size |
| != ((index_compressed_size + 3) &~ (size_t) 3))) |
| { |
| elf_uncompress_failed (); |
| backtrace_free (state, mem, index_uncompressed_size, error_callback, |
| data); |
| return 0; |
| } |
| |
| offset = (offset + 3) &~ (size_t) 3; |
| if (unlikely (offset != index_offset)) |
| { |
| elf_uncompress_failed (); |
| backtrace_free (state, mem, index_uncompressed_size, error_callback, |
| data); |
| return 0; |
| } |
| |
| return 1; |
| } |
| |
| /* This function is a hook for testing the LZMA support. It is only |
| used by tests. */ |
| |
| int |
| backtrace_uncompress_lzma (struct backtrace_state *state, |
| const unsigned char *compressed, |
| size_t compressed_size, |
| backtrace_error_callback error_callback, |
| void *data, unsigned char **uncompressed, |
| size_t *uncompressed_size) |
| { |
| return elf_uncompress_lzma (state, compressed, compressed_size, |
| error_callback, data, uncompressed, |
| uncompressed_size); |
| } |
| |
| /* Add the backtrace data for one ELF file. Returns 1 on success, |
| 0 on failure (in both cases descriptor is closed) or -1 if exe |
| is non-zero and the ELF file is ET_DYN, which tells the caller that |
| elf_add will need to be called on the descriptor again after |
| base_address is determined. */ |
| |
| static int |
| elf_add (struct backtrace_state *state, const char *filename, int descriptor, |
| const unsigned char *memory, size_t memory_size, |
| uintptr_t base_address, struct elf_ppc64_opd_data *caller_opd, |
| backtrace_error_callback error_callback, void *data, |
| fileline *fileline_fn, int *found_sym, int *found_dwarf, |
| struct dwarf_data **fileline_entry, int exe, int debuginfo, |
| const char *with_buildid_data, uint32_t with_buildid_size) |
| { |
| struct elf_view ehdr_view; |
| b_elf_ehdr ehdr; |
| off_t shoff; |
| unsigned int shnum; |
| unsigned int shstrndx; |
| struct elf_view shdrs_view; |
| int shdrs_view_valid; |
| const b_elf_shdr *shdrs; |
| const b_elf_shdr *shstrhdr; |
| size_t shstr_size; |
| off_t shstr_off; |
| struct elf_view names_view; |
| int names_view_valid; |
| const char *names; |
| unsigned int symtab_shndx; |
| unsigned int dynsym_shndx; |
| unsigned int i; |
| struct debug_section_info sections[DEBUG_MAX]; |
| struct debug_section_info zsections[DEBUG_MAX]; |
| struct elf_view symtab_view; |
| int symtab_view_valid; |
| struct elf_view strtab_view; |
| int strtab_view_valid; |
| struct elf_view buildid_view; |
| int buildid_view_valid; |
| const char *buildid_data; |
| uint32_t buildid_size; |
| struct elf_view debuglink_view; |
| int debuglink_view_valid; |
| const char *debuglink_name; |
| uint32_t debuglink_crc; |
| struct elf_view debugaltlink_view; |
| int debugaltlink_view_valid; |
| const char *debugaltlink_name; |
| const char *debugaltlink_buildid_data; |
| uint32_t debugaltlink_buildid_size; |
| struct elf_view gnu_debugdata_view; |
| int gnu_debugdata_view_valid; |
| size_t gnu_debugdata_size; |
| unsigned char *gnu_debugdata_uncompressed; |
| size_t gnu_debugdata_uncompressed_size; |
| off_t min_offset; |
| off_t max_offset; |
| off_t debug_size; |
| struct elf_view debug_view; |
| int debug_view_valid; |
| unsigned int using_debug_view; |
| uint16_t *zdebug_table; |
| struct elf_view split_debug_view[DEBUG_MAX]; |
| unsigned char split_debug_view_valid[DEBUG_MAX]; |
| struct elf_ppc64_opd_data opd_data, *opd; |
| int opd_view_valid; |
| struct dwarf_sections dwarf_sections; |
| |
| if (!debuginfo) |
| { |
| *found_sym = 0; |
| *found_dwarf = 0; |
| } |
| |
| shdrs_view_valid = 0; |
| names_view_valid = 0; |
| symtab_view_valid = 0; |
| strtab_view_valid = 0; |
| buildid_view_valid = 0; |
| buildid_data = NULL; |
| buildid_size = 0; |
| debuglink_view_valid = 0; |
| debuglink_name = NULL; |
| debuglink_crc = 0; |
| debugaltlink_view_valid = 0; |
| debugaltlink_name = NULL; |
| debugaltlink_buildid_data = NULL; |
| debugaltlink_buildid_size = 0; |
| gnu_debugdata_view_valid = 0; |
| gnu_debugdata_size = 0; |
| debug_view_valid = 0; |
| memset (&split_debug_view_valid[0], 0, sizeof split_debug_view_valid); |
| opd = NULL; |
| opd_view_valid = 0; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, 0, sizeof ehdr, |
| error_callback, data, &ehdr_view)) |
| goto fail; |
| |
| memcpy (&ehdr, ehdr_view.view.data, sizeof ehdr); |
| |
| elf_release_view (state, &ehdr_view, error_callback, data); |
| |
| if (ehdr.e_ident[EI_MAG0] != ELFMAG0 |
| || ehdr.e_ident[EI_MAG1] != ELFMAG1 |
| || ehdr.e_ident[EI_MAG2] != ELFMAG2 |
| || ehdr.e_ident[EI_MAG3] != ELFMAG3) |
| { |
| error_callback (data, "executable file is not ELF", 0); |
| goto fail; |
| } |
| if (ehdr.e_ident[EI_VERSION] != EV_CURRENT) |
| { |
| error_callback (data, "executable file is unrecognized ELF version", 0); |
| goto fail; |
| } |
| |
| #if BACKTRACE_ELF_SIZE == 32 |
| #define BACKTRACE_ELFCLASS ELFCLASS32 |
| #else |
| #define BACKTRACE_ELFCLASS ELFCLASS64 |
| #endif |
| |
| if (ehdr.e_ident[EI_CLASS] != BACKTRACE_ELFCLASS) |
| { |
| error_callback (data, "executable file is unexpected ELF class", 0); |
| goto fail; |
| } |
| |
| if (ehdr.e_ident[EI_DATA] != ELFDATA2LSB |
| && ehdr.e_ident[EI_DATA] != ELFDATA2MSB) |
| { |
| error_callback (data, "executable file has unknown endianness", 0); |
| goto fail; |
| } |
| |
| /* If the executable is ET_DYN, it is either a PIE, or we are running |
| directly a shared library with .interp. We need to wait for |
| dl_iterate_phdr in that case to determine the actual base_address. */ |
| if (exe && ehdr.e_type == ET_DYN) |
| return -1; |
| |
| shoff = ehdr.e_shoff; |
| shnum = ehdr.e_shnum; |
| shstrndx = ehdr.e_shstrndx; |
| |
| if ((shnum == 0 || shstrndx == SHN_XINDEX) |
| && shoff != 0) |
| { |
| struct elf_view shdr_view; |
| const b_elf_shdr *shdr; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, shoff, |
| sizeof shdr, error_callback, data, &shdr_view)) |
| goto fail; |
| |
| shdr = (const b_elf_shdr *) shdr_view.view.data; |
| |
| if (shnum == 0) |
| shnum = shdr->sh_size; |
| |
| if (shstrndx == SHN_XINDEX) |
| { |
| shstrndx = shdr->sh_link; |
| |
| /* Versions of the GNU binutils between 2.12 and 2.18 did |
| not handle objects with more than SHN_LORESERVE sections |
| correctly. All large section indexes were offset by |
| 0x100. There is more information at |
| http://sourceware.org/bugzilla/show_bug.cgi?id-5900 . |
| Fortunately these object files are easy to detect, as the |
| GNU binutils always put the section header string table |
| near the end of the list of sections. Thus if the |
| section header string table index is larger than the |
| number of sections, then we know we have to subtract |
| 0x100 to get the real section index. */ |
| if (shstrndx >= shnum && shstrndx >= SHN_LORESERVE + 0x100) |
| shstrndx -= 0x100; |
| } |
| |
| elf_release_view (state, &shdr_view, error_callback, data); |
| } |
| |
| if (shnum == 0 || shstrndx == 0) |
| goto fail; |
| |
| /* To translate PC to file/line when using DWARF, we need to find |
| the .debug_info and .debug_line sections. */ |
| |
| /* Read the section headers, skipping the first one. */ |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| shoff + sizeof (b_elf_shdr), |
| (shnum - 1) * sizeof (b_elf_shdr), |
| error_callback, data, &shdrs_view)) |
| goto fail; |
| shdrs_view_valid = 1; |
| shdrs = (const b_elf_shdr *) shdrs_view.view.data; |
| |
| /* Read the section names. */ |
| |
| shstrhdr = &shdrs[shstrndx - 1]; |
| shstr_size = shstrhdr->sh_size; |
| shstr_off = shstrhdr->sh_offset; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, shstr_off, |
| shstrhdr->sh_size, error_callback, data, &names_view)) |
| goto fail; |
| names_view_valid = 1; |
| names = (const char *) names_view.view.data; |
| |
| symtab_shndx = 0; |
| dynsym_shndx = 0; |
| |
| memset (sections, 0, sizeof sections); |
| memset (zsections, 0, sizeof zsections); |
| |
| /* Look for the symbol table. */ |
| for (i = 1; i < shnum; ++i) |
| { |
| const b_elf_shdr *shdr; |
| unsigned int sh_name; |
| const char *name; |
| int j; |
| |
| shdr = &shdrs[i - 1]; |
| |
| if (shdr->sh_type == SHT_SYMTAB) |
| symtab_shndx = i; |
| else if (shdr->sh_type == SHT_DYNSYM) |
| dynsym_shndx = i; |
| |
| sh_name = shdr->sh_name; |
| if (sh_name >= shstr_size) |
| { |
| error_callback (data, "ELF section name out of range", 0); |
| goto fail; |
| } |
| |
| name = names + sh_name; |
| |
| for (j = 0; j < (int) DEBUG_MAX; ++j) |
| { |
| if (strcmp (name, dwarf_section_names[j]) == 0) |
| { |
| sections[j].offset = shdr->sh_offset; |
| sections[j].size = shdr->sh_size; |
| sections[j].compressed = (shdr->sh_flags & SHF_COMPRESSED) != 0; |
| break; |
| } |
| } |
| |
| if (name[0] == '.' && name[1] == 'z') |
| { |
| for (j = 0; j < (int) DEBUG_MAX; ++j) |
| { |
| if (strcmp (name + 2, dwarf_section_names[j] + 1) == 0) |
| { |
| zsections[j].offset = shdr->sh_offset; |
| zsections[j].size = shdr->sh_size; |
| break; |
| } |
| } |
| } |
| |
| /* Read the build ID if present. This could check for any |
| SHT_NOTE section with the right note name and type, but gdb |
| looks for a specific section name. */ |
| if ((!debuginfo || with_buildid_data != NULL) |
| && !buildid_view_valid |
| && strcmp (name, ".note.gnu.build-id") == 0) |
| { |
| const b_elf_note *note; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| shdr->sh_offset, shdr->sh_size, error_callback, |
| data, &buildid_view)) |
| goto fail; |
| |
| buildid_view_valid = 1; |
| note = (const b_elf_note *) buildid_view.view.data; |
| if (note->type == NT_GNU_BUILD_ID |
| && note->namesz == 4 |
| && strncmp (note->name, "GNU", 4) == 0 |
| && shdr->sh_size <= 12 + ((note->namesz + 3) & ~ 3) + note->descsz) |
| { |
| buildid_data = ¬e->name[0] + ((note->namesz + 3) & ~ 3); |
| buildid_size = note->descsz; |
| } |
| |
| if (with_buildid_size != 0) |
| { |
| if (buildid_size != with_buildid_size) |
| goto fail; |
| |
| if (memcmp (buildid_data, with_buildid_data, buildid_size) != 0) |
| goto fail; |
| } |
| } |
| |
| /* Read the debuglink file if present. */ |
| if (!debuginfo |
| && !debuglink_view_valid |
| && strcmp (name, ".gnu_debuglink") == 0) |
| { |
| const char *debuglink_data; |
| size_t crc_offset; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| shdr->sh_offset, shdr->sh_size, error_callback, |
| data, &debuglink_view)) |
| goto fail; |
| |
| debuglink_view_valid = 1; |
| debuglink_data = (const char *) debuglink_view.view.data; |
| crc_offset = strnlen (debuglink_data, shdr->sh_size); |
| crc_offset = (crc_offset + 3) & ~3; |
| if (crc_offset + 4 <= shdr->sh_size) |
| { |
| debuglink_name = debuglink_data; |
| debuglink_crc = *(const uint32_t*)(debuglink_data + crc_offset); |
| } |
| } |
| |
| if (!debugaltlink_view_valid |
| && strcmp (name, ".gnu_debugaltlink") == 0) |
| { |
| const char *debugaltlink_data; |
| size_t debugaltlink_name_len; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| shdr->sh_offset, shdr->sh_size, error_callback, |
| data, &debugaltlink_view)) |
| goto fail; |
| |
| debugaltlink_view_valid = 1; |
| debugaltlink_data = (const char *) debugaltlink_view.view.data; |
| debugaltlink_name = debugaltlink_data; |
| debugaltlink_name_len = strnlen (debugaltlink_data, shdr->sh_size); |
| if (debugaltlink_name_len < shdr->sh_size) |
| { |
| /* Include terminating zero. */ |
| debugaltlink_name_len += 1; |
| |
| debugaltlink_buildid_data |
| = debugaltlink_data + debugaltlink_name_len; |
| debugaltlink_buildid_size = shdr->sh_size - debugaltlink_name_len; |
| } |
| } |
| |
| if (!gnu_debugdata_view_valid |
| && strcmp (name, ".gnu_debugdata") == 0) |
| { |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| shdr->sh_offset, shdr->sh_size, error_callback, |
| data, &gnu_debugdata_view)) |
| goto fail; |
| |
| gnu_debugdata_size = shdr->sh_size; |
| gnu_debugdata_view_valid = 1; |
| } |
| |
| /* Read the .opd section on PowerPC64 ELFv1. */ |
| if (ehdr.e_machine == EM_PPC64 |
| && (ehdr.e_flags & EF_PPC64_ABI) < 2 |
| && shdr->sh_type == SHT_PROGBITS |
| && strcmp (name, ".opd") == 0) |
| { |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| shdr->sh_offset, shdr->sh_size, error_callback, |
| data, &opd_data.view)) |
| goto fail; |
| |
| opd = &opd_data; |
| opd->addr = shdr->sh_addr; |
| opd->data = (const char *) opd_data.view.view.data; |
| opd->size = shdr->sh_size; |
| opd_view_valid = 1; |
| } |
| } |
| |
| /* A debuginfo file may not have a useful .opd section, but we can use the |
| one from the original executable. */ |
| if (opd == NULL) |
| opd = caller_opd; |
| |
| if (symtab_shndx == 0) |
| symtab_shndx = dynsym_shndx; |
| if (symtab_shndx != 0) |
| { |
| const b_elf_shdr *symtab_shdr; |
| unsigned int strtab_shndx; |
| const b_elf_shdr *strtab_shdr; |
| struct elf_syminfo_data *sdata; |
| |
| symtab_shdr = &shdrs[symtab_shndx - 1]; |
| strtab_shndx = symtab_shdr->sh_link; |
| if (strtab_shndx >= shnum) |
| { |
| error_callback (data, |
| "ELF symbol table strtab link out of range", 0); |
| goto fail; |
| } |
| strtab_shdr = &shdrs[strtab_shndx - 1]; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| symtab_shdr->sh_offset, symtab_shdr->sh_size, |
| error_callback, data, &symtab_view)) |
| goto fail; |
| symtab_view_valid = 1; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| strtab_shdr->sh_offset, strtab_shdr->sh_size, |
| error_callback, data, &strtab_view)) |
| goto fail; |
| strtab_view_valid = 1; |
| |
| sdata = ((struct elf_syminfo_data *) |
| backtrace_alloc (state, sizeof *sdata, error_callback, data)); |
| if (sdata == NULL) |
| goto fail; |
| |
| if (!elf_initialize_syminfo (state, base_address, |
| symtab_view.view.data, symtab_shdr->sh_size, |
| strtab_view.view.data, strtab_shdr->sh_size, |
| error_callback, data, sdata, opd)) |
| { |
| backtrace_free (state, sdata, sizeof *sdata, error_callback, data); |
| goto fail; |
| } |
| |
| /* We no longer need the symbol table, but we hold on to the |
| string table permanently. */ |
| elf_release_view (state, &symtab_view, error_callback, data); |
| symtab_view_valid = 0; |
| strtab_view_valid = 0; |
| |
| *found_sym = 1; |
| |
| elf_add_syminfo_data (state, sdata); |
| } |
| |
| elf_release_view (state, &shdrs_view, error_callback, data); |
| shdrs_view_valid = 0; |
| elf_release_view (state, &names_view, error_callback, data); |
| names_view_valid = 0; |
| |
| /* If the debug info is in a separate file, read that one instead. */ |
| |
| if (buildid_data != NULL) |
| { |
| int d; |
| |
| d = elf_open_debugfile_by_buildid (state, buildid_data, buildid_size, |
| error_callback, data); |
| if (d >= 0) |
| { |
| int ret; |
| |
| elf_release_view (state, &buildid_view, error_callback, data); |
| if (debuglink_view_valid) |
| elf_release_view (state, &debuglink_view, error_callback, data); |
| if (debugaltlink_view_valid) |
| elf_release_view (state, &debugaltlink_view, error_callback, data); |
| ret = elf_add (state, "", d, NULL, 0, base_address, opd, |
| error_callback, data, fileline_fn, found_sym, |
| found_dwarf, NULL, 0, 1, NULL, 0); |
| if (ret < 0) |
| backtrace_close (d, error_callback, data); |
| else if (descriptor >= 0) |
| backtrace_close (descriptor, error_callback, data); |
| return ret; |
| } |
| } |
| |
| if (buildid_view_valid) |
| { |
| elf_release_view (state, &buildid_view, error_callback, data); |
| buildid_view_valid = 0; |
| } |
| |
| if (debuglink_name != NULL) |
| { |
| int d; |
| |
| d = elf_open_debugfile_by_debuglink (state, filename, debuglink_name, |
| debuglink_crc, error_callback, |
| data); |
| if (d >= 0) |
| { |
| int ret; |
| |
| elf_release_view (state, &debuglink_view, error_callback, data); |
| if (debugaltlink_view_valid) |
| elf_release_view (state, &debugaltlink_view, error_callback, data); |
| ret = elf_add (state, "", d, NULL, 0, base_address, opd, |
| error_callback, data, fileline_fn, found_sym, |
| found_dwarf, NULL, 0, 1, NULL, 0); |
| if (ret < 0) |
| backtrace_close (d, error_callback, data); |
| else if (descriptor >= 0) |
| backtrace_close(descriptor, error_callback, data); |
| return ret; |
| } |
| } |
| |
| if (debuglink_view_valid) |
| { |
| elf_release_view (state, &debuglink_view, error_callback, data); |
| debuglink_view_valid = 0; |
| } |
| |
| struct dwarf_data *fileline_altlink = NULL; |
| if (debugaltlink_name != NULL) |
| { |
| int d; |
| |
| d = elf_open_debugfile_by_debuglink (state, filename, debugaltlink_name, |
| 0, error_callback, data); |
| if (d >= 0) |
| { |
| int ret; |
| |
| ret = elf_add (state, filename, d, NULL, 0, base_address, opd, |
| error_callback, data, fileline_fn, found_sym, |
| found_dwarf, &fileline_altlink, 0, 1, |
| debugaltlink_buildid_data, debugaltlink_buildid_size); |
| elf_release_view (state, &debugaltlink_view, error_callback, data); |
| debugaltlink_view_valid = 0; |
| if (ret < 0) |
| { |
| backtrace_close (d, error_callback, data); |
| return ret; |
| } |
| } |
| } |
| |
| if (debugaltlink_view_valid) |
| { |
| elf_release_view (state, &debugaltlink_view, error_callback, data); |
| debugaltlink_view_valid = 0; |
| } |
| |
| if (gnu_debugdata_view_valid) |
| { |
| int ret; |
| |
| ret = elf_uncompress_lzma (state, |
| ((const unsigned char *) |
| gnu_debugdata_view.view.data), |
| gnu_debugdata_size, error_callback, data, |
| &gnu_debugdata_uncompressed, |
| &gnu_debugdata_uncompressed_size); |
| |
| elf_release_view (state, &gnu_debugdata_view, error_callback, data); |
| gnu_debugdata_view_valid = 0; |
| |
| if (ret) |
| { |
| ret = elf_add (state, filename, -1, gnu_debugdata_uncompressed, |
| gnu_debugdata_uncompressed_size, base_address, opd, |
| error_callback, data, fileline_fn, found_sym, |
| found_dwarf, NULL, 0, 0, NULL, 0); |
| if (ret >= 0 && descriptor >= 0) |
| backtrace_close(descriptor, error_callback, data); |
| return ret; |
| } |
| } |
| |
| if (opd_view_valid) |
| { |
| elf_release_view (state, &opd->view, error_callback, data); |
| opd_view_valid = 0; |
| opd = NULL; |
| } |
| |
| /* Read all the debug sections in a single view, since they are |
| probably adjacent in the file. If any of sections are |
| uncompressed, we never release this view. */ |
| |
| min_offset = 0; |
| max_offset = 0; |
| debug_size = 0; |
| for (i = 0; i < (int) DEBUG_MAX; ++i) |
| { |
| off_t end; |
| |
| if (sections[i].size != 0) |
| { |
| if (min_offset == 0 || sections[i].offset < min_offset) |
| min_offset = sections[i].offset; |
| end = sections[i].offset + sections[i].size; |
| if (end > max_offset) |
| max_offset = end; |
| debug_size += sections[i].size; |
| } |
| if (zsections[i].size != 0) |
| { |
| if (min_offset == 0 || zsections[i].offset < min_offset) |
| min_offset = zsections[i].offset; |
| end = zsections[i].offset + zsections[i].size; |
| if (end > max_offset) |
| max_offset = end; |
| debug_size += zsections[i].size; |
| } |
| } |
| if (min_offset == 0 || max_offset == 0) |
| { |
| if (descriptor >= 0) |
| { |
| if (!backtrace_close (descriptor, error_callback, data)) |
| goto fail; |
| } |
| return 1; |
| } |
| |
| /* If the total debug section size is large, assume that there are |
| gaps between the sections, and read them individually. */ |
| |
| if (max_offset - min_offset < 0x20000000 |
| || max_offset - min_offset < debug_size + 0x10000) |
| { |
| if (!elf_get_view (state, descriptor, memory, memory_size, min_offset, |
| max_offset - min_offset, error_callback, data, |
| &debug_view)) |
| goto fail; |
| debug_view_valid = 1; |
| } |
| else |
| { |
| memset (&split_debug_view[0], 0, sizeof split_debug_view); |
| for (i = 0; i < (int) DEBUG_MAX; ++i) |
| { |
| struct debug_section_info *dsec; |
| |
| if (sections[i].size != 0) |
| dsec = §ions[i]; |
| else if (zsections[i].size != 0) |
| dsec = &zsections[i]; |
| else |
| continue; |
| |
| if (!elf_get_view (state, descriptor, memory, memory_size, |
| dsec->offset, dsec->size, error_callback, data, |
| &split_debug_view[i])) |
| goto fail; |
| split_debug_view_valid[i] = 1; |
| |
| if (sections[i].size != 0) |
| sections[i].data = ((const unsigned char *) |
| split_debug_view[i].view.data); |
| else |
| zsections[i].data = ((const unsigned char *) |
| split_debug_view[i].view.data); |
| } |
| } |
| |
| /* We've read all we need from the executable. */ |
| if (descriptor >= 0) |
| { |
| if (!backtrace_close (descriptor, error_callback, data)) |
| goto fail; |
| descriptor = -1; |
| } |
| |
| using_debug_view = 0; |
| if (debug_view_valid) |
| { |
| for (i = 0; i < (int) DEBUG_MAX; ++i) |
| { |
| if (sections[i].size == 0) |
| sections[i].data = NULL; |
| else |
| { |
| sections[i].data = ((const unsigned char *) debug_view.view.data |
| + (sections[i].offset - min_offset)); |
| ++using_debug_view; |
| } |
| |
| if (zsections[i].size == 0) |
| zsections[i].data = NULL; |
| else |
| zsections[i].data = ((const unsigned char *) debug_view.view.data |
| + (zsections[i].offset - min_offset)); |
| } |
| } |
| |
| /* Uncompress the old format (--compress-debug-sections=zlib-gnu). */ |
| |
| zdebug_table = NULL; |
| for (i = 0; i < (int) DEBUG_MAX; ++i) |
| { |
| if (sections[i].size == 0 && zsections[i].size > 0) |
| { |
| unsigned char *uncompressed_data; |
| size_t uncompressed_size; |
| |
| if (zdebug_table == NULL) |
| { |
| zdebug_table = ((uint16_t *) |
| backtrace_alloc (state, ZLIB_TABLE_SIZE, |
| error_callback, data)); |
| if (zdebug_table == NULL) |
| goto fail; |
| } |
| |
| uncompressed_data = NULL; |
| uncompressed_size = 0; |
| if (!elf_uncompress_zdebug (state, zsections[i].data, |
| zsections[i].size, zdebug_table, |
| error_callback, data, |
| &uncompressed_data, &uncompressed_size)) |
| goto fail; |
| sections[i].data = uncompressed_data; |
| sections[i].size = uncompressed_size; |
| sections[i].compressed = 0; |
| |
| if (split_debug_view_valid[i]) |
| { |
| elf_release_view (state, &split_debug_view[i], |
| error_callback, data); |
| split_debug_view_valid[i] = 0; |
| } |
| } |
| } |
| |
| if (zdebug_table != NULL) |
| { |
| backtrace_free (state, zdebug_table, ZLIB_TABLE_SIZE, |
| error_callback, data); |
| zdebug_table = NULL; |
| } |
| |
| /* Uncompress the official ELF format |
| (--compress-debug-sections=zlib-gabi, --compress-debug-sections=zstd). */ |
| for (i = 0; i < (int) DEBUG_MAX; ++i) |
| { |
| unsigned char *uncompressed_data; |
| size_t uncompressed_size; |
| |
| if (sections[i].size == 0 || !sections[i].compressed) |
| continue; |
| |
| if (zdebug_table == NULL) |
| { |
| zdebug_table = ((uint16_t *) |
| backtrace_alloc (state, ZDEBUG_TABLE_SIZE, |
| error_callback, data)); |
| if (zdebug_table == NULL) |
| goto fail; |
| } |
| |
| uncompressed_data = NULL; |
| uncompressed_size = 0; |
| if (!elf_uncompress_chdr (state, sections[i].data, sections[i].size, |
| zdebug_table, error_callback, data, |
| &uncompressed_data, &uncompressed_size)) |
| goto fail; |
| sections[i].data = uncompressed_data; |
| sections[i].size = uncompressed_size; |
| sections[i].compressed = 0; |
| |
| if (debug_view_valid) |
| --using_debug_view; |
| else if (split_debug_view_valid[i]) |
| { |
| elf_release_view (state, &split_debug_view[i], error_callback, data); |
| split_debug_view_valid[i] = 0; |
| } |
| } |
| |
| if (zdebug_table != NULL) |
| backtrace_free (state, zdebug_table, ZDEBUG_TABLE_SIZE, |
| error_callback, data); |
| |
| if (debug_view_valid && using_debug_view == 0) |
| { |
| elf_release_view (state, &debug_view, error_callback, data); |
| debug_view_valid = 0; |
| } |
| |
| for (i = 0; i < (int) DEBUG_MAX; ++i) |
| { |
| dwarf_sections.data[i] = sections[i].data; |
| dwarf_sections.size[i] = sections[i].size; |
| } |
| |
| if (!backtrace_dwarf_add (state, base_address, &dwarf_sections, |
| ehdr.e_ident[EI_DATA] == ELFDATA2MSB, |
| fileline_altlink, |
| error_callback, data, fileline_fn, |
| fileline_entry)) |
| goto fail; |
| |
| *found_dwarf = 1; |
| |
| return 1; |
| |
| fail: |
| if (shdrs_view_valid) |
| elf_release_view (state, &shdrs_view, error_callback, data); |
| if (names_view_valid) |
| elf_release_view (state, &names_view, error_callback, data); |
| if (symtab_view_valid) |
| elf_release_view (state, &symtab_view, error_callback, data); |
| if (strtab_view_valid) |
| elf_release_view (state, &strtab_view, error_callback, data); |
| if (debuglink_view_valid) |
| elf_release_view (state, &debuglink_view, error_callback, data); |
| if (debugaltlink_view_valid) |
| elf_release_view (state, &debugaltlink_view, error_callback, data); |
| if (gnu_debugdata_view_valid) |
| elf_release_view (state, &gnu_debugdata_view, error_callback, data); |
| if (buildid_view_valid) |
| elf_release_view (state, &buildid_view, error_callback, data); |
| if (debug_view_valid) |
| elf_release_view (state, &debug_view, error_callback, data); |
| for (i = 0; i < (int) DEBUG_MAX; ++i) |
| { |
| if (split_debug_view_valid[i]) |
| elf_release_view (state, &split_debug_view[i], error_callback, data); |
| } |
| if (opd_view_valid) |
| elf_release_view (state, &opd->view, error_callback, data); |
| if (descriptor >= 0) |
| backtrace_close (descriptor, error_callback, data); |
| return 0; |
| } |
| |
| /* Data passed to phdr_callback. */ |
| |
| struct phdr_data |
| { |
| struct backtrace_state *state; |
| backtrace_error_callback error_callback; |
| void *data; |
| fileline *fileline_fn; |
| int *found_sym; |
| int *found_dwarf; |
| const char *exe_filename; |
| int exe_descriptor; |
| }; |
| |
| /* Callback passed to dl_iterate_phdr. Load debug info from shared |
| libraries. */ |
| |
| static int |
| #ifdef __i386__ |
| __attribute__ ((__force_align_arg_pointer__)) |
| #endif |
| phdr_callback (struct dl_phdr_info *info, size_t size ATTRIBUTE_UNUSED, |
| void *pdata) |
| { |
| struct phdr_data *pd = (struct phdr_data *) pdata; |
| const char *filename; |
| int descriptor; |
| int does_not_exist; |
| fileline elf_fileline_fn; |
| int found_dwarf; |
| |
| /* There is not much we can do if we don't have the module name, |
| unless executable is ET_DYN, where we expect the very first |
| phdr_callback to be for the PIE. */ |
| if (info->dlpi_name == NULL || info->dlpi_name[0] == '\0') |
| { |
| if (pd->exe_descriptor == -1) |
| return 0; |
| filename = pd->exe_filename; |
| descriptor = pd->exe_descriptor; |
| pd->exe_descriptor = -1; |
| } |
| else |
| { |
| if (pd->exe_descriptor != -1) |
| { |
| backtrace_close (pd->exe_descriptor, pd->error_callback, pd->data); |
| pd->exe_descriptor = -1; |
| } |
| |
| filename = info->dlpi_name; |
| descriptor = backtrace_open (info->dlpi_name, pd->error_callback, |
| pd->data, &does_not_exist); |
| if (descriptor < 0) |
| return 0; |
| } |
| |
| if (elf_add (pd->state, filename, descriptor, NULL, 0, info->dlpi_addr, NULL, |
| pd->error_callback, pd->data, &elf_fileline_fn, pd->found_sym, |
| &found_dwarf, NULL, 0, 0, NULL, 0)) |
| { |
| if (found_dwarf) |
| { |
| *pd->found_dwarf = 1; |
| *pd->fileline_fn = elf_fileline_fn; |
| } |
| } |
| |
| return 0; |
| } |
| |
| /* Initialize the backtrace data we need from an ELF executable. At |
| the ELF level, all we need to do is find the debug info |
| sections. */ |
| |
| int |
| backtrace_initialize (struct backtrace_state *state, const char *filename, |
| int descriptor, backtrace_error_callback error_callback, |
| void *data, fileline *fileline_fn) |
| { |
| int ret; |
| int found_sym; |
| int found_dwarf; |
| fileline elf_fileline_fn = elf_nodebug; |
| struct phdr_data pd; |
| |
| ret = elf_add (state, filename, descriptor, NULL, 0, 0, NULL, error_callback, |
| data, &elf_fileline_fn, &found_sym, &found_dwarf, NULL, 1, 0, |
| NULL, 0); |
| if (!ret) |
| return 0; |
| |
| pd.state = state; |
| pd.error_callback = error_callback; |
| pd.data = data; |
| pd.fileline_fn = &elf_fileline_fn; |
| pd.found_sym = &found_sym; |
| pd.found_dwarf = &found_dwarf; |
| pd.exe_filename = filename; |
| pd.exe_descriptor = ret < 0 ? descriptor : -1; |
| |
| dl_iterate_phdr (phdr_callback, (void *) &pd); |
| |
| if (!state->threaded) |
| { |
| if (found_sym) |
| state->syminfo_fn = elf_syminfo; |
| else if (state->syminfo_fn == NULL) |
| state->syminfo_fn = elf_nosyms; |
| } |
| else |
| { |
| if (found_sym) |
| backtrace_atomic_store_pointer (&state->syminfo_fn, elf_syminfo); |
| else |
| (void) __sync_bool_compare_and_swap (&state->syminfo_fn, NULL, |
| elf_nosyms); |
| } |
| |
| if (!state->threaded) |
| *fileline_fn = state->fileline_fn; |
| else |
| *fileline_fn = backtrace_atomic_load_pointer (&state->fileline_fn); |
| |
| if (*fileline_fn == NULL || *fileline_fn == elf_nodebug) |
| *fileline_fn = elf_fileline_fn; |
| |
| return 1; |
| } |