1 /*
   2  * Copyright (c) 2003, 2019, Oracle and/or its affiliates. All rights reserved.
   3  * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
   4  *
   5  * This code is free software; you can redistribute it and/or modify it
   6  * under the terms of the GNU General Public License version 2 only, as
   7  * published by the Free Software Foundation.
   8  *
   9  * This code is distributed in the hope that it will be useful, but WITHOUT
  10  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
  11  * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License
  12  * version 2 for more details (a copy is included in the LICENSE file that
  13  * accompanied this code).
  14  *
  15  * You should have received a copy of the GNU General Public License version
  16  * 2 along with this work; if not, write to the Free Software Foundation,
  17  * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA.
  18  *
  19  * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA
  20  * or visit www.oracle.com if you need additional information or have any
  21  * questions.
  22  *
  23  */
  24 
  25 #include <jni.h>
  26 #include <unistd.h>
  27 #include <fcntl.h>
  28 #include <string.h>
  29 #include <stdlib.h>
  30 #include <stddef.h>
  31 #include <elf.h>
  32 #include <link.h>
  33 #include "libproc_impl.h"
  34 #include "proc_service.h"
  35 #include "salibelf.h"
  36 #include "cds.h"
  37 
  38 // This file has the libproc implementation to read core files.
  39 // For live processes, refer to ps_proc.c. Portions of this is adapted
  40 // /modelled after Solaris libproc.so (in particular Pcore.c)
  41 
  42 //----------------------------------------------------------------------
  43 // ps_prochandle cleanup helper functions
  44 
  45 // close all file descriptors
  46 static void close_files(struct ps_prochandle* ph) {
  47   lib_info* lib = NULL;
  48 
  49   // close core file descriptor
  50   if (ph->core->core_fd >= 0)
  51     close(ph->core->core_fd);
  52 
  53   // close exec file descriptor
  54   if (ph->core->exec_fd >= 0)
  55     close(ph->core->exec_fd);
  56 
  57   // close interp file descriptor
  58   if (ph->core->interp_fd >= 0)
  59     close(ph->core->interp_fd);
  60 
  61   // close class share archive file
  62   if (ph->core->classes_jsa_fd >= 0)
  63     close(ph->core->classes_jsa_fd);
  64 
  65   // close all library file descriptors
  66   lib = ph->libs;
  67   while (lib) {
  68     int fd = lib->fd;
  69     if (fd >= 0 && fd != ph->core->exec_fd) {
  70       close(fd);
  71     }
  72     lib = lib->next;
  73   }
  74 }
  75 
  76 // clean all map_info stuff
  77 static void destroy_map_info(struct ps_prochandle* ph) {
  78   map_info* map = ph->core->maps;
  79   while (map) {
  80     map_info* next = map->next;
  81     free(map);
  82     map = next;
  83   }
  84 
  85   if (ph->core->map_array) {
  86     free(ph->core->map_array);
  87   }
  88 
  89   // Part of the class sharing workaround
  90   map = ph->core->class_share_maps;
  91   while (map) {
  92     map_info* next = map->next;
  93     free(map);
  94     map = next;
  95   }
  96 }
  97 
  98 // ps_prochandle operations
  99 static void core_release(struct ps_prochandle* ph) {
 100   if (ph->core) {
 101     close_files(ph);
 102     destroy_map_info(ph);
 103     free(ph->core);
 104   }
 105 }
 106 
 107 static map_info* allocate_init_map(int fd, off_t offset, uintptr_t vaddr, size_t memsz) {
 108   map_info* map;
 109   if ( (map = (map_info*) calloc(1, sizeof(map_info))) == NULL) {
 110     print_debug("can't allocate memory for map_info\n");
 111     return NULL;
 112   }
 113 
 114   // initialize map
 115   map->fd     = fd;
 116   map->offset = offset;
 117   map->vaddr  = vaddr;
 118   map->memsz  = memsz;
 119   return map;
 120 }
 121 
 122 // add map info with given fd, offset, vaddr and memsz
 123 static map_info* add_map_info(struct ps_prochandle* ph, int fd, off_t offset,
 124                              uintptr_t vaddr, size_t memsz) {
 125   map_info* map;
 126   if ((map = allocate_init_map(fd, offset, vaddr, memsz)) == NULL) {
 127     return NULL;
 128   }
 129 
 130   // add this to map list
 131   map->next  = ph->core->maps;
 132   ph->core->maps   = map;
 133   ph->core->num_maps++;
 134 
 135   return map;
 136 }
 137 
 138 // Part of the class sharing workaround
 139 static map_info* add_class_share_map_info(struct ps_prochandle* ph, off_t offset,
 140                              uintptr_t vaddr, size_t memsz) {
 141   map_info* map;
 142   if ((map = allocate_init_map(ph->core->classes_jsa_fd,
 143                                offset, vaddr, memsz)) == NULL) {
 144     return NULL;
 145   }
 146 
 147   map->next = ph->core->class_share_maps;
 148   ph->core->class_share_maps = map;
 149   return map;
 150 }
 151 
 152 // Return the map_info for the given virtual address.  We keep a sorted
 153 // array of pointers in ph->map_array, so we can binary search.
 154 static map_info* core_lookup(struct ps_prochandle *ph, uintptr_t addr) {
 155   int mid, lo = 0, hi = ph->core->num_maps - 1;
 156   map_info *mp;
 157 
 158   while (hi - lo > 1) {
 159     mid = (lo + hi) / 2;
 160     if (addr >= ph->core->map_array[mid]->vaddr) {
 161       lo = mid;
 162     } else {
 163       hi = mid;
 164     }
 165   }
 166 
 167   if (addr < ph->core->map_array[hi]->vaddr) {
 168     mp = ph->core->map_array[lo];
 169   } else {
 170     mp = ph->core->map_array[hi];
 171   }
 172 
 173   if (addr >= mp->vaddr && addr < mp->vaddr + mp->memsz) {
 174     return (mp);
 175   }
 176 
 177 
 178   // Part of the class sharing workaround
 179   // Unfortunately, we have no way of detecting -Xshare state.
 180   // Check out the share maps atlast, if we don't find anywhere.
 181   // This is done this way so to avoid reading share pages
 182   // ahead of other normal maps. For eg. with -Xshare:off we don't
 183   // want to prefer class sharing data to data from core.
 184   mp = ph->core->class_share_maps;
 185   if (mp) {
 186     print_debug("can't locate map_info at 0x%lx, trying class share maps\n", addr);
 187   }
 188   while (mp) {
 189     if (addr >= mp->vaddr && addr < mp->vaddr + mp->memsz) {
 190       print_debug("located map_info at 0x%lx from class share maps\n", addr);
 191       return (mp);
 192     }
 193     mp = mp->next;
 194   }
 195 
 196   print_debug("can't locate map_info at 0x%lx\n", addr);
 197   return (NULL);
 198 }
 199 
 200 //---------------------------------------------------------------
 201 // Part of the class sharing workaround:
 202 //
 203 // With class sharing, pages are mapped from classes.jsa file.
 204 // The read-only class sharing pages are mapped as MAP_SHARED,
 205 // PROT_READ pages. These pages are not dumped into core dump.
 206 // With this workaround, these pages are read from classes.jsa.
 207 
 208 static bool read_jboolean(struct ps_prochandle* ph, uintptr_t addr, jboolean* pvalue) {
 209   jboolean i;
 210   if (ps_pdread(ph, (psaddr_t) addr, &i, sizeof(i)) == PS_OK) {
 211     *pvalue = i;
 212     return true;
 213   } else {
 214     return false;
 215   }
 216 }
 217 
 218 static bool read_pointer(struct ps_prochandle* ph, uintptr_t addr, uintptr_t* pvalue) {
 219   uintptr_t uip;
 220   if (ps_pdread(ph, (psaddr_t) addr, (char *)&uip, sizeof(uip)) == PS_OK) {
 221     *pvalue = uip;
 222     return true;
 223   } else {
 224     return false;
 225   }
 226 }
 227 
 228 // used to read strings from debuggee
 229 static bool read_string(struct ps_prochandle* ph, uintptr_t addr, char* buf, size_t size) {
 230   size_t i = 0;
 231   char  c = ' ';
 232 
 233   while (c != '\0') {
 234     if (ps_pdread(ph, (psaddr_t) addr, &c, sizeof(char)) != PS_OK) {
 235       return false;
 236     }
 237     if (i < size - 1) {
 238       buf[i] = c;
 239     } else {
 240       // smaller buffer
 241       return false;
 242     }
 243     i++; addr++;
 244   }
 245 
 246   buf[i] = '\0';
 247   return true;
 248 }
 249 
 250 #define USE_SHARED_SPACES_SYM "UseSharedSpaces"
 251 // mangled name of Arguments::SharedArchivePath
 252 #define SHARED_ARCHIVE_PATH_SYM "_ZN9Arguments17SharedArchivePathE"
 253 #define LIBJVM_NAME "/libjvm.so"
 254 
 255 static bool init_classsharing_workaround(struct ps_prochandle* ph) {
 256   lib_info* lib = ph->libs;
 257   while (lib != NULL) {
 258     // we are iterating over shared objects from the core dump. look for
 259     // libjvm.so.
 260     const char *jvm_name = 0;
 261     if ((jvm_name = strstr(lib->name, LIBJVM_NAME)) != 0) {
 262       char classes_jsa[PATH_MAX];
 263       CDSFileMapHeaderBase header;
 264       int fd = -1;
 265       int m = 0;
 266       size_t n = 0;
 267       uintptr_t base = 0, useSharedSpacesAddr = 0;
 268       uintptr_t sharedArchivePathAddrAddr = 0, sharedArchivePathAddr = 0;
 269       jboolean useSharedSpaces = 0;
 270       map_info* mi = 0;
 271 
 272       memset(classes_jsa, 0, sizeof(classes_jsa));
 273       jvm_name = lib->name;
 274       useSharedSpacesAddr = lookup_symbol(ph, jvm_name, USE_SHARED_SPACES_SYM);
 275       if (useSharedSpacesAddr == 0) {
 276         print_debug("can't lookup 'UseSharedSpaces' flag\n");
 277         return false;
 278       }
 279 
 280       // Hotspot vm types are not exported to build this library. So
 281       // using equivalent type jboolean to read the value of
 282       // UseSharedSpaces which is same as hotspot type "bool".
 283       if (read_jboolean(ph, useSharedSpacesAddr, &useSharedSpaces) != true) {
 284         print_debug("can't read the value of 'UseSharedSpaces' flag\n");
 285         return false;
 286       }
 287 
 288       if ((int)useSharedSpaces == 0) {
 289         print_debug("UseSharedSpaces is false, assuming -Xshare:off!\n");
 290         return true;
 291       }
 292 
 293       sharedArchivePathAddrAddr = lookup_symbol(ph, jvm_name, SHARED_ARCHIVE_PATH_SYM);
 294       if (sharedArchivePathAddrAddr == 0) {
 295         print_debug("can't lookup shared archive path symbol\n");
 296         return false;
 297       }
 298 
 299       if (read_pointer(ph, sharedArchivePathAddrAddr, &sharedArchivePathAddr) != true) {
 300         print_debug("can't read shared archive path pointer\n");
 301         return false;
 302       }
 303 
 304       if (read_string(ph, sharedArchivePathAddr, classes_jsa, sizeof(classes_jsa)) != true) {
 305         print_debug("can't read shared archive path value\n");
 306         return false;
 307       }
 308 
 309       print_debug("looking for %s\n", classes_jsa);
 310       // open the class sharing archive file
 311       fd = pathmap_open(classes_jsa);
 312       if (fd < 0) {
 313         print_debug("can't open %s!\n", classes_jsa);
 314         ph->core->classes_jsa_fd = -1;
 315         return false;
 316       } else {
 317         print_debug("opened %s\n", classes_jsa);
 318       }
 319 
 320       // read CDSFileMapHeaderBase from the file
 321       memset(&header, 0, sizeof(CDSFileMapHeaderBase));
 322       if ((n = read(fd, &header, sizeof(CDSFileMapHeaderBase)))
 323            != sizeof(CDSFileMapHeaderBase)) {
 324         print_debug("can't read shared archive file map header from %s\n", classes_jsa);
 325         close(fd);
 326         return false;
 327       }
 328 
 329       // check file magic
 330       if (header._magic != CDS_ARCHIVE_MAGIC) {
 331         print_debug("%s has bad shared archive file magic number 0x%x, expecting 0x%x\n",
 332                     classes_jsa, header._magic, CDS_ARCHIVE_MAGIC);
 333         close(fd);
 334         return false;
 335       }
 336 
 337       // check version
 338       if (header._version != CURRENT_CDS_ARCHIVE_VERSION) {
 339         print_debug("%s has wrong shared archive file version %d, expecting %d\n",
 340                      classes_jsa, header._version, CURRENT_CDS_ARCHIVE_VERSION);
 341         close(fd);
 342         return false;
 343       }
 344 
 345       ph->core->classes_jsa_fd = fd;
 346       // add read-only maps from classes.jsa to the list of maps
 347       for (m = 0; m < NUM_CDS_REGIONS; m++) {
 348         if (header._space[m]._read_only) {
 349           base = (uintptr_t) header._space[m]._addr._base;
 350           // no need to worry about the fractional pages at-the-end.
 351           // possible fractional pages are handled by core_read_data.
 352           add_class_share_map_info(ph, (off_t) header._space[m]._file_offset,
 353                                    base, (size_t) header._space[m]._used);
 354           print_debug("added a share archive map at 0x%lx\n", base);
 355         }
 356       }
 357       return true;
 358    }
 359    lib = lib->next;
 360   }
 361   return true;
 362 }
 363 
 364 
 365 //---------------------------------------------------------------------------
 366 // functions to handle map_info
 367 
 368 // Order mappings based on virtual address.  We use this function as the
 369 // callback for sorting the array of map_info pointers.
 370 static int core_cmp_mapping(const void *lhsp, const void *rhsp)
 371 {
 372   const map_info *lhs = *((const map_info **)lhsp);
 373   const map_info *rhs = *((const map_info **)rhsp);
 374 
 375   if (lhs->vaddr == rhs->vaddr) {
 376     return (0);
 377   }
 378 
 379   return (lhs->vaddr < rhs->vaddr ? -1 : 1);
 380 }
 381 
 382 // we sort map_info by starting virtual address so that we can do
 383 // binary search to read from an address.
 384 static bool sort_map_array(struct ps_prochandle* ph) {
 385   size_t num_maps = ph->core->num_maps;
 386   map_info* map = ph->core->maps;
 387   int i = 0;
 388 
 389   // allocate map_array
 390   map_info** array;
 391   if ( (array = (map_info**) malloc(sizeof(map_info*) * num_maps)) == NULL) {
 392     print_debug("can't allocate memory for map array\n");
 393     return false;
 394   }
 395 
 396   // add maps to array
 397   while (map) {
 398     array[i] = map;
 399     i++;
 400     map = map->next;
 401   }
 402 
 403   // sort is called twice. If this is second time, clear map array
 404   if (ph->core->map_array) {
 405     free(ph->core->map_array);
 406   }
 407 
 408   ph->core->map_array = array;
 409   // sort the map_info array by base virtual address.
 410   qsort(ph->core->map_array, ph->core->num_maps, sizeof (map_info*),
 411         core_cmp_mapping);
 412 
 413   // print map
 414   if (is_debug()) {
 415     int j = 0;
 416     print_debug("---- sorted virtual address map ----\n");
 417     for (j = 0; j < ph->core->num_maps; j++) {
 418       print_debug("base = 0x%lx\tsize = %zu\n", ph->core->map_array[j]->vaddr,
 419                   ph->core->map_array[j]->memsz);
 420     }
 421   }
 422 
 423   return true;
 424 }
 425 
 426 #ifndef MIN
 427 #define MIN(x, y) (((x) < (y))? (x): (y))
 428 #endif
 429 
 430 static bool core_read_data(struct ps_prochandle* ph, uintptr_t addr, char *buf, size_t size) {
 431    ssize_t resid = size;
 432    int page_size=sysconf(_SC_PAGE_SIZE);
 433    while (resid != 0) {
 434       map_info *mp = core_lookup(ph, addr);
 435       uintptr_t mapoff;
 436       ssize_t len, rem;
 437       off_t off;
 438       int fd;
 439 
 440       if (mp == NULL) {
 441          break;  /* No mapping for this address */
 442       }
 443 
 444       fd = mp->fd;
 445       mapoff = addr - mp->vaddr;
 446       len = MIN(resid, mp->memsz - mapoff);
 447       off = mp->offset + mapoff;
 448 
 449       if ((len = pread(fd, buf, len, off)) <= 0) {
 450          break;
 451       }
 452 
 453       resid -= len;
 454       addr += len;
 455       buf = (char *)buf + len;
 456 
 457       // mappings always start at page boundary. But, may end in fractional
 458       // page. fill zeros for possible fractional page at the end of a mapping.
 459       rem = mp->memsz % page_size;
 460       if (rem > 0) {
 461          rem = page_size - rem;
 462          len = MIN(resid, rem);
 463          resid -= len;
 464          addr += len;
 465          // we are not assuming 'buf' to be zero initialized.
 466          memset(buf, 0, len);
 467          buf += len;
 468       }
 469    }
 470 
 471    if (resid) {
 472       print_debug("core read failed for %d byte(s) @ 0x%lx (%d more bytes)\n",
 473               size, addr, resid);
 474       return false;
 475    } else {
 476       return true;
 477    }
 478 }
 479 
 480 // null implementation for write
 481 static bool core_write_data(struct ps_prochandle* ph,
 482                              uintptr_t addr, const char *buf , size_t size) {
 483    return false;
 484 }
 485 
 486 static bool core_get_lwp_regs(struct ps_prochandle* ph, lwpid_t lwp_id,
 487                           struct user_regs_struct* regs) {
 488    // for core we have cached the lwp regs from NOTE section
 489    thread_info* thr = ph->threads;
 490    while (thr) {
 491      if (thr->lwp_id == lwp_id) {
 492        memcpy(regs, &thr->regs, sizeof(struct user_regs_struct));
 493        return true;
 494      }
 495      thr = thr->next;
 496    }
 497    return false;
 498 }
 499 
 500 static ps_prochandle_ops core_ops = {
 501    .release=  core_release,
 502    .p_pread=  core_read_data,
 503    .p_pwrite= core_write_data,
 504    .get_lwp_regs= core_get_lwp_regs
 505 };
 506 
 507 // read regs and create thread from NT_PRSTATUS entries from core file
 508 static bool core_handle_prstatus(struct ps_prochandle* ph, const char* buf, size_t nbytes) {
 509    // we have to read prstatus_t from buf
 510    // assert(nbytes == sizeof(prstaus_t), "size mismatch on prstatus_t");
 511    prstatus_t* prstat = (prstatus_t*) buf;
 512    thread_info* newthr;
 513    print_debug("got integer regset for lwp %d\n", prstat->pr_pid);
 514    if((newthr = add_thread_info(ph, prstat->pr_pid)) == NULL)
 515       return false;
 516 
 517    // copy regs
 518    memcpy(&newthr->regs, prstat->pr_reg, sizeof(struct user_regs_struct));
 519 
 520    if (is_debug()) {
 521       print_debug("integer regset\n");
 522 #ifdef i386
 523       // print the regset
 524       print_debug("\teax = 0x%x\n", newthr->regs.eax);
 525       print_debug("\tebx = 0x%x\n", newthr->regs.ebx);
 526       print_debug("\tecx = 0x%x\n", newthr->regs.ecx);
 527       print_debug("\tedx = 0x%x\n", newthr->regs.edx);
 528       print_debug("\tesp = 0x%x\n", newthr->regs.esp);
 529       print_debug("\tebp = 0x%x\n", newthr->regs.ebp);
 530       print_debug("\tesi = 0x%x\n", newthr->regs.esi);
 531       print_debug("\tedi = 0x%x\n", newthr->regs.edi);
 532       print_debug("\teip = 0x%x\n", newthr->regs.eip);
 533 #endif
 534 
 535 #if defined(amd64) || defined(x86_64)
 536       // print the regset
 537       print_debug("\tr15 = 0x%lx\n", newthr->regs.r15);
 538       print_debug("\tr14 = 0x%lx\n", newthr->regs.r14);
 539       print_debug("\tr13 = 0x%lx\n", newthr->regs.r13);
 540       print_debug("\tr12 = 0x%lx\n", newthr->regs.r12);
 541       print_debug("\trbp = 0x%lx\n", newthr->regs.rbp);
 542       print_debug("\trbx = 0x%lx\n", newthr->regs.rbx);
 543       print_debug("\tr11 = 0x%lx\n", newthr->regs.r11);
 544       print_debug("\tr10 = 0x%lx\n", newthr->regs.r10);
 545       print_debug("\tr9 = 0x%lx\n", newthr->regs.r9);
 546       print_debug("\tr8 = 0x%lx\n", newthr->regs.r8);
 547       print_debug("\trax = 0x%lx\n", newthr->regs.rax);
 548       print_debug("\trcx = 0x%lx\n", newthr->regs.rcx);
 549       print_debug("\trdx = 0x%lx\n", newthr->regs.rdx);
 550       print_debug("\trsi = 0x%lx\n", newthr->regs.rsi);
 551       print_debug("\trdi = 0x%lx\n", newthr->regs.rdi);
 552       print_debug("\torig_rax = 0x%lx\n", newthr->regs.orig_rax);
 553       print_debug("\trip = 0x%lx\n", newthr->regs.rip);
 554       print_debug("\tcs = 0x%lx\n", newthr->regs.cs);
 555       print_debug("\teflags = 0x%lx\n", newthr->regs.eflags);
 556       print_debug("\trsp = 0x%lx\n", newthr->regs.rsp);
 557       print_debug("\tss = 0x%lx\n", newthr->regs.ss);
 558       print_debug("\tfs_base = 0x%lx\n", newthr->regs.fs_base);
 559       print_debug("\tgs_base = 0x%lx\n", newthr->regs.gs_base);
 560       print_debug("\tds = 0x%lx\n", newthr->regs.ds);
 561       print_debug("\tes = 0x%lx\n", newthr->regs.es);
 562       print_debug("\tfs = 0x%lx\n", newthr->regs.fs);
 563       print_debug("\tgs = 0x%lx\n", newthr->regs.gs);
 564 #endif
 565    }
 566 
 567    return true;
 568 }
 569 
 570 #define ROUNDUP(x, y)  ((((x)+((y)-1))/(y))*(y))
 571 
 572 // read NT_PRSTATUS entries from core NOTE segment
 573 static bool core_handle_note(struct ps_prochandle* ph, ELF_PHDR* note_phdr) {
 574    char* buf = NULL;
 575    char* p = NULL;
 576    size_t size = note_phdr->p_filesz;
 577 
 578    // we are interested in just prstatus entries. we will ignore the rest.
 579    // Advance the seek pointer to the start of the PT_NOTE data
 580    if (lseek(ph->core->core_fd, note_phdr->p_offset, SEEK_SET) == (off_t)-1) {
 581       print_debug("failed to lseek to PT_NOTE data\n");
 582       return false;
 583    }
 584 
 585    // Now process the PT_NOTE structures.  Each one is preceded by
 586    // an Elf{32/64}_Nhdr structure describing its type and size.
 587    if ( (buf = (char*) malloc(size)) == NULL) {
 588       print_debug("can't allocate memory for reading core notes\n");
 589       goto err;
 590    }
 591 
 592    // read notes into buffer
 593    if (read(ph->core->core_fd, buf, size) != size) {
 594       print_debug("failed to read notes, core file must have been truncated\n");
 595       goto err;
 596    }
 597 
 598    p = buf;
 599    while (p < buf + size) {
 600       ELF_NHDR* notep = (ELF_NHDR*) p;
 601       char* descdata  = p + sizeof(ELF_NHDR) + ROUNDUP(notep->n_namesz, 4);
 602       print_debug("Note header with n_type = %d and n_descsz = %u\n",
 603                                    notep->n_type, notep->n_descsz);
 604 
 605       if (notep->n_type == NT_PRSTATUS) {
 606         if (core_handle_prstatus(ph, descdata, notep->n_descsz) != true) {
 607           return false;
 608         }
 609       } else if (notep->n_type == NT_AUXV) {
 610         // Get first segment from entry point
 611         ELF_AUXV *auxv = (ELF_AUXV *)descdata;
 612         while (auxv->a_type != AT_NULL) {
 613           if (auxv->a_type == AT_ENTRY) {
 614             // Set entry point address to address of dynamic section.
 615             // We will adjust it in read_exec_segments().
 616             ph->core->dynamic_addr = auxv->a_un.a_val;
 617             break;
 618           }
 619           auxv++;
 620         }
 621       }
 622       p = descdata + ROUNDUP(notep->n_descsz, 4);
 623    }
 624 
 625    free(buf);
 626    return true;
 627 
 628 err:
 629    if (buf) free(buf);
 630    return false;
 631 }
 632 
 633 // read all segments from core file
 634 static bool read_core_segments(struct ps_prochandle* ph, ELF_EHDR* core_ehdr) {
 635    int i = 0;
 636    ELF_PHDR* phbuf = NULL;
 637    ELF_PHDR* core_php = NULL;
 638 
 639    if ((phbuf =  read_program_header_table(ph->core->core_fd, core_ehdr)) == NULL)
 640       return false;
 641 
 642    /*
 643     * Now iterate through the program headers in the core file.
 644     * We're interested in two types of Phdrs: PT_NOTE (which
 645     * contains a set of saved /proc structures), and PT_LOAD (which
 646     * represents a memory mapping from the process's address space).
 647     *
 648     * Difference b/w Solaris PT_NOTE and Linux/BSD PT_NOTE:
 649     *
 650     *     In Solaris there are two PT_NOTE segments the first PT_NOTE (if present)
 651     *     contains /proc structs in the pre-2.6 unstructured /proc format. the last
 652     *     PT_NOTE has data in new /proc format.
 653     *
 654     *     In Solaris, there is only one pstatus (process status). pstatus contains
 655     *     integer register set among other stuff. For each LWP, we have one lwpstatus
 656     *     entry that has integer regset for that LWP.
 657     *
 658     *     Linux threads are actually 'clone'd processes. To support core analysis
 659     *     of "multithreaded" process, Linux creates more than one pstatus (called
 660     *     "prstatus") entry in PT_NOTE. Each prstatus entry has integer regset for one
 661     *     "thread". Please refer to Linux kernel src file 'fs/binfmt_elf.c', in particular
 662     *     function "elf_core_dump".
 663     */
 664 
 665     for (core_php = phbuf, i = 0; i < core_ehdr->e_phnum; i++) {
 666       switch (core_php->p_type) {
 667          case PT_NOTE:
 668             if (core_handle_note(ph, core_php) != true) {
 669               goto err;
 670             }
 671             break;
 672 
 673          case PT_LOAD: {
 674             if (core_php->p_filesz != 0) {
 675                if (add_map_info(ph, ph->core->core_fd, core_php->p_offset,
 676                   core_php->p_vaddr, core_php->p_filesz) == NULL) goto err;
 677             }
 678             break;
 679          }
 680       }
 681 
 682       core_php++;
 683    }
 684 
 685    free(phbuf);
 686    return true;
 687 err:
 688    free(phbuf);
 689    return false;
 690 }
 691 
 692 // read segments of a shared object
 693 static bool read_lib_segments(struct ps_prochandle* ph, int lib_fd, ELF_EHDR* lib_ehdr, uintptr_t lib_base) {
 694   int i = 0;
 695   ELF_PHDR* phbuf;
 696   ELF_PHDR* lib_php = NULL;
 697 
 698   int page_size = sysconf(_SC_PAGE_SIZE);
 699 
 700   if ((phbuf = read_program_header_table(lib_fd, lib_ehdr)) == NULL) {
 701     return false;
 702   }
 703 
 704   // we want to process only PT_LOAD segments that are not writable.
 705   // i.e., text segments. The read/write/exec (data) segments would
 706   // have been already added from core file segments.
 707   for (lib_php = phbuf, i = 0; i < lib_ehdr->e_phnum; i++) {
 708     if ((lib_php->p_type == PT_LOAD) && !(lib_php->p_flags & PF_W) && (lib_php->p_filesz != 0)) {
 709 
 710       uintptr_t target_vaddr = lib_php->p_vaddr + lib_base;
 711       map_info *existing_map = core_lookup(ph, target_vaddr);
 712 
 713       if (existing_map == NULL){
 714         if (add_map_info(ph, lib_fd, lib_php->p_offset,
 715                           target_vaddr, lib_php->p_memsz) == NULL) {
 716           goto err;
 717         }
 718       } else {
 719         // Coredump stores value of p_memsz elf field
 720         // rounded up to page boundary.
 721 
 722         if ((existing_map->memsz != page_size) &&
 723             (existing_map->fd != lib_fd) &&
 724             (ROUNDUP(existing_map->memsz, page_size) != ROUNDUP(lib_php->p_memsz, page_size))) {
 725 
 726           print_debug("address conflict @ 0x%lx (existing map size = %ld, size = %ld, flags = %d)\n",
 727                         target_vaddr, existing_map->memsz, lib_php->p_memsz, lib_php->p_flags);
 728           goto err;
 729         }
 730 
 731         /* replace PT_LOAD segment with library segment */
 732         print_debug("overwrote with new address mapping (memsz %ld -> %ld)\n",
 733                      existing_map->memsz, ROUNDUP(lib_php->p_memsz, page_size));
 734 
 735         existing_map->fd = lib_fd;
 736         existing_map->offset = lib_php->p_offset;
 737         existing_map->memsz = ROUNDUP(lib_php->p_memsz, page_size);
 738       }
 739     }
 740 
 741     lib_php++;
 742   }
 743 
 744   free(phbuf);
 745   return true;
 746 err:
 747   free(phbuf);
 748   return false;
 749 }
 750 
 751 // process segments from interpreter (ld.so or ld-linux.so)
 752 static bool read_interp_segments(struct ps_prochandle* ph) {
 753   ELF_EHDR interp_ehdr;
 754 
 755   if (read_elf_header(ph->core->interp_fd, &interp_ehdr) != true) {
 756     print_debug("interpreter is not a valid ELF file\n");
 757     return false;
 758   }
 759 
 760   if (read_lib_segments(ph, ph->core->interp_fd, &interp_ehdr, ph->core->ld_base_addr) != true) {
 761     print_debug("can't read segments of interpreter\n");
 762     return false;
 763   }
 764 
 765   return true;
 766 }
 767 
 768 // process segments of a a.out
 769 static bool read_exec_segments(struct ps_prochandle* ph, ELF_EHDR* exec_ehdr) {
 770   int i = 0;
 771   ELF_PHDR* phbuf = NULL;
 772   ELF_PHDR* exec_php = NULL;
 773 
 774   if ((phbuf = read_program_header_table(ph->core->exec_fd, exec_ehdr)) == NULL) {
 775     return false;
 776   }
 777 
 778   for (exec_php = phbuf, i = 0; i < exec_ehdr->e_phnum; i++) {
 779     switch (exec_php->p_type) {
 780 
 781       // add mappings for PT_LOAD segments
 782     case PT_LOAD: {
 783       // add only non-writable segments of non-zero filesz
 784       if (!(exec_php->p_flags & PF_W) && exec_php->p_filesz != 0) {
 785         if (add_map_info(ph, ph->core->exec_fd, exec_php->p_offset, exec_php->p_vaddr, exec_php->p_filesz) == NULL) goto err;
 786       }
 787       break;
 788     }
 789 
 790     // read the interpreter and it's segments
 791     case PT_INTERP: {
 792       char interp_name[BUF_SIZE + 1];
 793 
 794       // BUF_SIZE is PATH_MAX + NAME_MAX + 1.
 795       if (exec_php->p_filesz > BUF_SIZE) {
 796         goto err;
 797       }
 798       if (pread(ph->core->exec_fd, interp_name,
 799                 exec_php->p_filesz, exec_php->p_offset) != exec_php->p_filesz) {
 800         print_debug("Unable to read in the ELF interpreter\n");
 801         goto err;
 802       }
 803       interp_name[exec_php->p_filesz] = '\0';
 804       print_debug("ELF interpreter %s\n", interp_name);
 805       // read interpreter segments as well
 806       if ((ph->core->interp_fd = pathmap_open(interp_name)) < 0) {
 807         print_debug("can't open runtime loader\n");
 808         goto err;
 809       }
 810       break;
 811     }
 812 
 813     // from PT_DYNAMIC we want to read address of first link_map addr
 814     case PT_DYNAMIC: {
 815       if (exec_ehdr->e_type == ET_EXEC) {
 816         ph->core->dynamic_addr = exec_php->p_vaddr;
 817       } else { // ET_DYN
 818         // dynamic_addr has entry point of executable.
 819         // Thus we should substract it.
 820         ph->core->dynamic_addr += exec_php->p_vaddr - exec_ehdr->e_entry;
 821       }
 822       print_debug("address of _DYNAMIC is 0x%lx\n", ph->core->dynamic_addr);
 823       break;
 824     }
 825 
 826     } // switch
 827     exec_php++;
 828   } // for
 829 
 830   free(phbuf);
 831   return true;
 832  err:
 833   free(phbuf);
 834   return false;
 835 }
 836 
 837 
 838 #define FIRST_LINK_MAP_OFFSET offsetof(struct r_debug,  r_map)
 839 #define LD_BASE_OFFSET        offsetof(struct r_debug,  r_ldbase)
 840 #define LINK_MAP_ADDR_OFFSET  offsetof(struct link_map, l_addr)
 841 #define LINK_MAP_NAME_OFFSET  offsetof(struct link_map, l_name)
 842 #define LINK_MAP_NEXT_OFFSET  offsetof(struct link_map, l_next)
 843 
 844 // read shared library info from runtime linker's data structures.
 845 // This work is done by librtlb_db in Solaris
 846 static bool read_shared_lib_info(struct ps_prochandle* ph) {
 847   uintptr_t addr = ph->core->dynamic_addr;
 848   uintptr_t debug_base;
 849   uintptr_t first_link_map_addr;
 850   uintptr_t ld_base_addr;
 851   uintptr_t link_map_addr;
 852   uintptr_t lib_base_diff;
 853   uintptr_t lib_base;
 854   uintptr_t lib_name_addr;
 855   char lib_name[BUF_SIZE];
 856   ELF_DYN dyn;
 857   ELF_EHDR elf_ehdr;
 858   int lib_fd;
 859 
 860   // _DYNAMIC has information of the form
 861   //         [tag] [data] [tag] [data] .....
 862   // Both tag and data are pointer sized.
 863   // We look for dynamic info with DT_DEBUG. This has shared object info.
 864   // refer to struct r_debug in link.h
 865 
 866   dyn.d_tag = DT_NULL;
 867   while (dyn.d_tag != DT_DEBUG) {
 868     if (ps_pdread(ph, (psaddr_t) addr, &dyn, sizeof(ELF_DYN)) != PS_OK) {
 869       print_debug("can't read debug info from _DYNAMIC\n");
 870       return false;
 871     }
 872     addr += sizeof(ELF_DYN);
 873   }
 874 
 875   // we have got Dyn entry with DT_DEBUG
 876   debug_base = dyn.d_un.d_ptr;
 877   // at debug_base we have struct r_debug. This has first link map in r_map field
 878   if (ps_pdread(ph, (psaddr_t) debug_base + FIRST_LINK_MAP_OFFSET,
 879                  &first_link_map_addr, sizeof(uintptr_t)) != PS_OK) {
 880     print_debug("can't read first link map address\n");
 881     return false;
 882   }
 883 
 884   // read ld_base address from struct r_debug
 885   if (ps_pdread(ph, (psaddr_t) debug_base + LD_BASE_OFFSET, &ld_base_addr,
 886                  sizeof(uintptr_t)) != PS_OK) {
 887     print_debug("can't read ld base address\n");
 888     return false;
 889   }
 890   ph->core->ld_base_addr = ld_base_addr;
 891 
 892   print_debug("interpreter base address is 0x%lx\n", ld_base_addr);
 893 
 894   // now read segments from interp (i.e ld.so or ld-linux.so or ld-elf.so)
 895   if (read_interp_segments(ph) != true) {
 896       return false;
 897   }
 898 
 899   // after adding interpreter (ld.so) mappings sort again
 900   if (sort_map_array(ph) != true) {
 901     return false;
 902   }
 903 
 904    print_debug("first link map is at 0x%lx\n", first_link_map_addr);
 905 
 906    link_map_addr = first_link_map_addr;
 907    while (link_map_addr != 0) {
 908       // read library base address of the .so. Note that even though <sys/link.h> calls
 909       // link_map->l_addr as "base address",  this is * not * really base virtual
 910       // address of the shared object. This is actually the difference b/w the virtual
 911       // address mentioned in shared object and the actual virtual base where runtime
 912       // linker loaded it. We use "base diff" in read_lib_segments call below.
 913 
 914       if (ps_pdread(ph, (psaddr_t) link_map_addr + LINK_MAP_ADDR_OFFSET,
 915                    &lib_base_diff, sizeof(uintptr_t)) != PS_OK) {
 916          print_debug("can't read shared object base address diff\n");
 917          return false;
 918       }
 919 
 920       // read address of the name
 921       if (ps_pdread(ph, (psaddr_t) link_map_addr + LINK_MAP_NAME_OFFSET,
 922                     &lib_name_addr, sizeof(uintptr_t)) != PS_OK) {
 923          print_debug("can't read address of shared object name\n");
 924          return false;
 925       }
 926 
 927       // read name of the shared object
 928       lib_name[0] = '\0';
 929       if (lib_name_addr != 0 &&
 930           read_string(ph, (uintptr_t) lib_name_addr, lib_name, sizeof(lib_name)) != true) {
 931          print_debug("can't read shared object name\n");
 932          // don't let failure to read the name stop opening the file.  If something is really wrong
 933          // it will fail later.
 934       }
 935 
 936       if (lib_name[0] != '\0') {
 937          // ignore empty lib names
 938          lib_fd = pathmap_open(lib_name);
 939 
 940          if (lib_fd < 0) {
 941             print_debug("can't open shared object %s\n", lib_name);
 942             // continue with other libraries...
 943          } else {
 944             if (read_elf_header(lib_fd, &elf_ehdr)) {
 945                lib_base = lib_base_diff + find_base_address(lib_fd, &elf_ehdr);
 946                print_debug("reading library %s @ 0x%lx [ 0x%lx ]\n",
 947                            lib_name, lib_base, lib_base_diff);
 948                // while adding library mappings we need to use "base difference".
 949                if (! read_lib_segments(ph, lib_fd, &elf_ehdr, lib_base_diff)) {
 950                   print_debug("can't read shared object's segments\n");
 951                   close(lib_fd);
 952                   return false;
 953                }
 954                add_lib_info_fd(ph, lib_name, lib_fd, lib_base);
 955                // Map info is added for the library (lib_name) so
 956                // we need to re-sort it before calling the p_pdread.
 957                if (sort_map_array(ph) != true)
 958                   return false;
 959             } else {
 960                print_debug("can't read ELF header for shared object %s\n", lib_name);
 961                close(lib_fd);
 962                // continue with other libraries...
 963             }
 964          }
 965       }
 966 
 967     // read next link_map address
 968     if (ps_pdread(ph, (psaddr_t) link_map_addr + LINK_MAP_NEXT_OFFSET,
 969                    &link_map_addr, sizeof(uintptr_t)) != PS_OK) {
 970       print_debug("can't read next link in link_map\n");
 971       return false;
 972     }
 973   }
 974 
 975   return true;
 976 }
 977 
 978 // the one and only one exposed stuff from this file
 979 JNIEXPORT struct ps_prochandle* JNICALL
 980 Pgrab_core(const char* exec_file, const char* core_file) {
 981   ELF_EHDR core_ehdr;
 982   ELF_EHDR exec_ehdr;
 983   ELF_EHDR lib_ehdr;
 984 
 985   struct ps_prochandle* ph = (struct ps_prochandle*) calloc(1, sizeof(struct ps_prochandle));
 986   if (ph == NULL) {
 987     print_debug("can't allocate ps_prochandle\n");
 988     return NULL;
 989   }
 990 
 991   if ((ph->core = (struct core_data*) calloc(1, sizeof(struct core_data))) == NULL) {
 992     free(ph);
 993     print_debug("can't allocate ps_prochandle\n");
 994     return NULL;
 995   }
 996 
 997   // initialize ph
 998   ph->ops = &core_ops;
 999   ph->core->core_fd   = -1;
1000   ph->core->exec_fd   = -1;
1001   ph->core->interp_fd = -1;
1002 
1003   // open the core file
1004   if ((ph->core->core_fd = open(core_file, O_RDONLY)) < 0) {
1005     print_debug("can't open core file\n");
1006     goto err;
1007   }
1008 
1009   // read core file ELF header
1010   if (read_elf_header(ph->core->core_fd, &core_ehdr) != true || core_ehdr.e_type != ET_CORE) {
1011     print_debug("core file is not a valid ELF ET_CORE file\n");
1012     goto err;
1013   }
1014 
1015   if ((ph->core->exec_fd = open(exec_file, O_RDONLY)) < 0) {
1016     print_debug("can't open executable file\n");
1017     goto err;
1018   }
1019 
1020   if (read_elf_header(ph->core->exec_fd, &exec_ehdr) != true ||
1021       ((exec_ehdr.e_type != ET_EXEC) && (exec_ehdr.e_type != ET_DYN))) {
1022     print_debug("executable file is not a valid ELF file\n");
1023     goto err;
1024   }
1025 
1026   // process core file segments
1027   if (read_core_segments(ph, &core_ehdr) != true) {
1028     goto err;
1029   }
1030 
1031   // process exec file segments
1032   if (read_exec_segments(ph, &exec_ehdr) != true) {
1033     goto err;
1034   }
1035 
1036   // exec file is also treated like a shared object for symbol search
1037   if (add_lib_info_fd(ph, exec_file, ph->core->exec_fd,
1038                       (uintptr_t)0 + find_base_address(ph->core->exec_fd, &exec_ehdr)) == NULL) {
1039     goto err;
1040   }
1041 
1042   // allocate and sort maps into map_array, we need to do this
1043   // here because read_shared_lib_info needs to read from debuggee
1044   // address space
1045   if (sort_map_array(ph) != true) {
1046     goto err;
1047   }
1048 
1049   if (read_shared_lib_info(ph) != true) {
1050     goto err;
1051   }
1052 
1053   // sort again because we have added more mappings from shared objects
1054   if (sort_map_array(ph) != true) {
1055     goto err;
1056   }
1057 
1058   if (init_classsharing_workaround(ph) != true) {
1059     goto err;
1060   }
1061 
1062   return ph;
1063 
1064 err:
1065   Prelease(ph);
1066   return NULL;
1067 }