#define __APPLE_API_PRIVATE 1
#define __APPLE_API_UNSTABLE 1
#include <kern/debug.h>
#include <mach/i386/vm_param.h>
#include <string.h>
#include <mach/vm_param.h>
#include <mach/vm_prot.h>
#include <mach/machine.h>
#include <mach/time_value.h>
#include <sys/kdebug.h>
#include <kern/spl.h>
#include <kern/assert.h>
#include <kern/lock_group.h>
#include <kern/misc_protos.h>
#include <kern/startup.h>
#include <kern/clock.h>
#include <kern/cpu_data.h>
#include <kern/machine.h>
#include <i386/postcode.h>
#include <i386/mp_desc.h>
#include <i386/misc_protos.h>
#include <i386/thread.h>
#include <i386/trap.h>
#include <i386/machine_routines.h>
#include <i386/mp.h>
#include <i386/cpuid.h>
#include <i386/fpu.h>
#include <i386/machine_cpu.h>
#include <i386/pmap.h>
#if CONFIG_MTRR
#include <i386/mtrr.h>
#endif
#include <i386/ucode.h>
#include <i386/pmCPU.h>
#include <i386/panic_hooks.h>
#include <architecture/i386/pio.h>
#include <pexpert/i386/boot.h>
#include <kdp/kdp_dyld.h>
#include <kdp/kdp_core.h>
#include <vm/pmap.h>
#include <vm/vm_map.h>
#include <vm/vm_kern.h>
#include <IOKit/IOBSD.h>
#include <IOKit/IOPlatformExpert.h>
#include <IOKit/IOHibernatePrivate.h>
#include <pexpert/i386/efi.h>
#include <kern/thread.h>
#include <kern/sched.h>
#include <mach-o/loader.h>
#include <mach-o/nlist.h>
#include <libkern/kernel_mach_header.h>
#include <libkern/OSKextLibPrivate.h>
#include <libkern/crc.h>
#if DEBUG || DEVELOPMENT
#define DPRINTF(x...) kprintf(x)
#else
#define DPRINTF(x...)
#endif
#ifndef ROUNDUP
#define ROUNDUP(a, b) (((a) + ((b) - 1)) & (~((b) - 1)))
#endif
#ifndef ROUNDDOWN
#define ROUNDDOWN(x, y) (((x)/(y))*(y))
#endif
static void machine_conf(void);
void panic_print_symbol_name(vm_address_t search);
void RecordPanicStackshot(void);
typedef enum paniclog_flush_type {
kPaniclogFlushBase = 1,
kPaniclogFlushStackshot = 2,
kPaniclogFlushOtherLog = 3
} paniclog_flush_type_t;
void paniclog_flush_internal(paniclog_flush_type_t variant);
extern const char version[];
extern char osversion[];
extern int max_unsafe_quanta;
extern int max_poll_quanta;
extern unsigned int panic_is_inited;
extern int proc_pid(void *p);
#define FP_ALIGNMENT_MASK ((uint32_t)(0x3))
#define FP_LR_OFFSET ((uint32_t)4)
#define FP_LR_OFFSET64 ((uint32_t)8)
#define FP_MAX_NUM_TO_EVALUATE (50)
volatile int pbtcpu = -1;
hw_lock_data_t pbtlock;
uint32_t pbtcnt = 0;
volatile int panic_double_fault_cpu = -1;
#define PRINT_ARGS_FROM_STACK_FRAME 0
typedef struct _cframe_t {
struct _cframe_t *prev;
uintptr_t caller;
#if PRINT_ARGS_FROM_STACK_FRAME
unsigned args[0];
#endif
} cframe_t;
static unsigned panic_io_port;
static unsigned commit_paniclog_to_nvram;
boolean_t coprocessor_paniclog_flush = FALSE;
struct kcdata_descriptor kc_panic_data;
static boolean_t begun_panic_stackshot = FALSE;
extern kern_return_t do_stackshot(void *);
extern void kdp_snapshot_preflight(int pid, void *tracebuf,
uint32_t tracebuf_size, uint32_t flags,
kcdata_descriptor_t data_p,
boolean_t enable_faulting);
extern int kdp_stack_snapshot_bytes_traced(void);
vm_offset_t panic_stackshot_buf = 0;
size_t panic_stackshot_buf_len = 0;
size_t panic_stackshot_len = 0;
void
print_one_backtrace(pmap_t pmap, vm_offset_t topfp, const char *cur_marker,
boolean_t is_64_bit)
{
int i = 0;
addr64_t lr;
addr64_t fp;
addr64_t fp_for_ppn;
ppnum_t ppn;
boolean_t dump_kernel_stack;
fp = topfp;
fp_for_ppn = 0;
ppn = (ppnum_t)NULL;
if (fp >= VM_MIN_KERNEL_ADDRESS) {
dump_kernel_stack = TRUE;
} else {
dump_kernel_stack = FALSE;
}
do {
if ((fp == 0) || ((fp & FP_ALIGNMENT_MASK) != 0)) {
break;
}
if (dump_kernel_stack && ((fp < VM_MIN_KERNEL_ADDRESS) || (fp > VM_MAX_KERNEL_ADDRESS))) {
break;
}
if ((!dump_kernel_stack) && (fp >= VM_MIN_KERNEL_ADDRESS)) {
break;
}
if ((((fp + FP_LR_OFFSET) ^ fp_for_ppn) >> PAGE_SHIFT) != 0x0U) {
ppn = pmap_find_phys(pmap, fp + FP_LR_OFFSET);
fp_for_ppn = fp + (is_64_bit ? FP_LR_OFFSET64 : FP_LR_OFFSET);
}
if (ppn != (ppnum_t)NULL) {
if (is_64_bit) {
lr = ml_phys_read_double_64(((((vm_offset_t)ppn) << PAGE_SHIFT)) | ((fp + FP_LR_OFFSET64) & PAGE_MASK));
} else {
lr = ml_phys_read_word(((((vm_offset_t)ppn) << PAGE_SHIFT)) | ((fp + FP_LR_OFFSET) & PAGE_MASK));
}
} else {
if (is_64_bit) {
paniclog_append_noflush("%s\t Could not read LR from frame at 0x%016llx\n", cur_marker, fp + FP_LR_OFFSET64);
} else {
paniclog_append_noflush("%s\t Could not read LR from frame at 0x%08x\n", cur_marker, (uint32_t)(fp + FP_LR_OFFSET));
}
break;
}
if (((fp ^ fp_for_ppn) >> PAGE_SHIFT) != 0x0U) {
ppn = pmap_find_phys(pmap, fp);
fp_for_ppn = fp;
}
if (ppn != (ppnum_t)NULL) {
if (is_64_bit) {
fp = ml_phys_read_double_64(((((vm_offset_t)ppn) << PAGE_SHIFT)) | (fp & PAGE_MASK));
} else {
fp = ml_phys_read_word(((((vm_offset_t)ppn) << PAGE_SHIFT)) | (fp & PAGE_MASK));
}
} else {
if (is_64_bit) {
paniclog_append_noflush("%s\t Could not read FP from frame at 0x%016llx\n", cur_marker, fp);
} else {
paniclog_append_noflush("%s\t Could not read FP from frame at 0x%08x\n", cur_marker, (uint32_t)fp);
}
break;
}
if (is_64_bit) {
paniclog_append_noflush("%s\t0x%016llx\n", cur_marker, lr);
} else {
paniclog_append_noflush("%s\t0x%08x\n", cur_marker, (uint32_t)lr);
}
} while ((++i < FP_MAX_NUM_TO_EVALUATE) && (fp != topfp));
}
void
machine_startup(void)
{
int boot_arg;
#if 0
if (PE_get_hotkey( kPEControlKey )) {
halt_in_debugger = halt_in_debugger ? 0 : 1;
}
#endif
if (!PE_parse_boot_argn("nvram_paniclog", &commit_paniclog_to_nvram, sizeof(commit_paniclog_to_nvram))) {
commit_paniclog_to_nvram = 1;
}
if (PE_parse_boot_argn("pmsafe_debug", &boot_arg, sizeof(boot_arg))) {
pmsafe_debug = boot_arg;
}
hw_lock_init(&pbtlock);
if (PE_parse_boot_argn("preempt", &boot_arg, sizeof(boot_arg))) {
default_preemption_rate = boot_arg;
}
if (PE_parse_boot_argn("unsafe", &boot_arg, sizeof(boot_arg))) {
max_unsafe_quanta = boot_arg;
}
if (PE_parse_boot_argn("poll", &boot_arg, sizeof(boot_arg))) {
max_poll_quanta = boot_arg;
}
if (PE_parse_boot_argn("yield", &boot_arg, sizeof(boot_arg))) {
sched_poll_yield_shift = boot_arg;
}
if (PE_parse_boot_argn("panic_io_port", &boot_arg, sizeof(boot_arg))) {
panic_io_port = boot_arg & 0xffff;
}
machine_conf();
panic_hooks_init();
kernel_bootstrap();
}
static void
machine_conf(void)
{
machine_info.memory_size = (typeof(machine_info.memory_size))mem_size;
}
extern void *gPEEFIRuntimeServices;
extern void *gPEEFISystemTable;
static void
efi_set_tables_64(EFI_SYSTEM_TABLE_64 * system_table)
{
EFI_RUNTIME_SERVICES_64 *runtime;
uint32_t hdr_cksum;
uint32_t cksum;
DPRINTF("Processing 64-bit EFI tables at %p\n", system_table);
do {
DPRINTF("Header:\n");
DPRINTF(" Signature: 0x%016llx\n", system_table->Hdr.Signature);
DPRINTF(" Revision: 0x%08x\n", system_table->Hdr.Revision);
DPRINTF(" HeaderSize: 0x%08x\n", system_table->Hdr.HeaderSize);
DPRINTF(" CRC32: 0x%08x\n", system_table->Hdr.CRC32);
DPRINTF("RuntimeServices: 0x%016llx\n", system_table->RuntimeServices);
if (system_table->Hdr.Signature != EFI_SYSTEM_TABLE_SIGNATURE) {
kprintf("Bad EFI system table signature\n");
break;
}
hdr_cksum = system_table->Hdr.CRC32;
system_table->Hdr.CRC32 = 0;
cksum = crc32(0L, system_table, system_table->Hdr.HeaderSize);
DPRINTF("System table calculated CRC32 = 0x%x, header = 0x%x\n", cksum, hdr_cksum);
system_table->Hdr.CRC32 = hdr_cksum;
if (cksum != hdr_cksum) {
kprintf("Bad EFI system table checksum\n");
break;
}
gPEEFISystemTable = system_table;
if (system_table->RuntimeServices == 0) {
kprintf("No runtime table present\n");
break;
}
DPRINTF("RuntimeServices table at 0x%qx\n", system_table->RuntimeServices);
runtime = (EFI_RUNTIME_SERVICES_64 *) (uintptr_t)system_table->RuntimeServices;
DPRINTF("Checking runtime services table %p\n", runtime);
if (runtime->Hdr.Signature != EFI_RUNTIME_SERVICES_SIGNATURE) {
kprintf("Bad EFI runtime table signature\n");
break;
}
hdr_cksum = runtime->Hdr.CRC32;
runtime->Hdr.CRC32 = 0;
cksum = crc32(0L, runtime, runtime->Hdr.HeaderSize);
DPRINTF("Runtime table calculated CRC32 = 0x%x, header = 0x%x\n", cksum, hdr_cksum);
runtime->Hdr.CRC32 = hdr_cksum;
if (cksum != hdr_cksum) {
kprintf("Bad EFI runtime table checksum\n");
break;
}
gPEEFIRuntimeServices = runtime;
}while (FALSE);
}
static void
efi_set_tables_32(EFI_SYSTEM_TABLE_32 * system_table)
{
EFI_RUNTIME_SERVICES_32 *runtime;
uint32_t hdr_cksum;
uint32_t cksum;
DPRINTF("Processing 32-bit EFI tables at %p\n", system_table);
do {
DPRINTF("Header:\n");
DPRINTF(" Signature: 0x%016llx\n", system_table->Hdr.Signature);
DPRINTF(" Revision: 0x%08x\n", system_table->Hdr.Revision);
DPRINTF(" HeaderSize: 0x%08x\n", system_table->Hdr.HeaderSize);
DPRINTF(" CRC32: 0x%08x\n", system_table->Hdr.CRC32);
DPRINTF("RuntimeServices: 0x%08x\n", system_table->RuntimeServices);
if (system_table->Hdr.Signature != EFI_SYSTEM_TABLE_SIGNATURE) {
kprintf("Bad EFI system table signature\n");
break;
}
hdr_cksum = system_table->Hdr.CRC32;
system_table->Hdr.CRC32 = 0;
DPRINTF("System table at %p HeaderSize 0x%x\n", system_table, system_table->Hdr.HeaderSize);
cksum = crc32(0L, system_table, system_table->Hdr.HeaderSize);
DPRINTF("System table calculated CRC32 = 0x%x, header = 0x%x\n", cksum, hdr_cksum);
system_table->Hdr.CRC32 = hdr_cksum;
if (cksum != hdr_cksum) {
kprintf("Bad EFI system table checksum\n");
break;
}
gPEEFISystemTable = system_table;
if (system_table->RuntimeServices == 0) {
kprintf("No runtime table present\n");
break;
}
DPRINTF("RuntimeServices table at 0x%x\n", system_table->RuntimeServices);
runtime = (EFI_RUNTIME_SERVICES_32 *)
(system_table->RuntimeServices | VM_MIN_KERNEL_ADDRESS);
DPRINTF("Runtime table addressed at %p\n", runtime);
if (runtime->Hdr.Signature != EFI_RUNTIME_SERVICES_SIGNATURE) {
kprintf("Bad EFI runtime table signature\n");
break;
}
hdr_cksum = runtime->Hdr.CRC32;
runtime->Hdr.CRC32 = 0;
cksum = crc32(0L, runtime, runtime->Hdr.HeaderSize);
DPRINTF("Runtime table calculated CRC32 = 0x%x, header = 0x%x\n", cksum, hdr_cksum);
runtime->Hdr.CRC32 = hdr_cksum;
if (cksum != hdr_cksum) {
kprintf("Bad EFI runtime table checksum\n");
break;
}
DPRINTF("Runtime functions\n");
DPRINTF(" GetTime : 0x%x\n", runtime->GetTime);
DPRINTF(" SetTime : 0x%x\n", runtime->SetTime);
DPRINTF(" GetWakeupTime : 0x%x\n", runtime->GetWakeupTime);
DPRINTF(" SetWakeupTime : 0x%x\n", runtime->SetWakeupTime);
DPRINTF(" SetVirtualAddressMap : 0x%x\n", runtime->SetVirtualAddressMap);
DPRINTF(" ConvertPointer : 0x%x\n", runtime->ConvertPointer);
DPRINTF(" GetVariable : 0x%x\n", runtime->GetVariable);
DPRINTF(" GetNextVariableName : 0x%x\n", runtime->GetNextVariableName);
DPRINTF(" SetVariable : 0x%x\n", runtime->SetVariable);
DPRINTF(" GetNextHighMonotonicCount: 0x%x\n", runtime->GetNextHighMonotonicCount);
DPRINTF(" ResetSystem : 0x%x\n", runtime->ResetSystem);
gPEEFIRuntimeServices = runtime;
}while (FALSE);
}
static void
efi_init(void)
{
boot_args *args = (boot_args *)PE_state.bootArgs;
kprintf("Initializing EFI runtime services\n");
do{
vm_offset_t vm_size, vm_addr;
vm_map_offset_t phys_addr;
EfiMemoryRange *mptr;
unsigned int msize, mcount;
unsigned int i;
msize = args->MemoryMapDescriptorSize;
mcount = args->MemoryMapSize / msize;
DPRINTF("efi_init() kernel base: 0x%x size: 0x%x\n",
args->kaddr, args->ksize);
DPRINTF(" efiSystemTable physical: 0x%x virtual: %p\n",
args->efiSystemTable,
(void *) ml_static_ptovirt(args->efiSystemTable));
DPRINTF(" efiRuntimeServicesPageStart: 0x%x\n",
args->efiRuntimeServicesPageStart);
DPRINTF(" efiRuntimeServicesPageCount: 0x%x\n",
args->efiRuntimeServicesPageCount);
DPRINTF(" efiRuntimeServicesVirtualPageStart: 0x%016llx\n",
args->efiRuntimeServicesVirtualPageStart);
mptr = (EfiMemoryRange *)ml_static_ptovirt(args->MemoryMap);
for (i = 0; i < mcount; i++, mptr = (EfiMemoryRange *)(((vm_offset_t)mptr) + msize)) {
if (((mptr->Attribute & EFI_MEMORY_RUNTIME) == EFI_MEMORY_RUNTIME)) {
vm_size = (vm_offset_t)i386_ptob((uint32_t)mptr->NumberOfPages);
vm_addr = (vm_offset_t) mptr->VirtualStart;
if (vm_addr < VM_MIN_KERNEL_ADDRESS) {
vm_addr |= VM_MIN_KERNEL_ADDRESS;
}
phys_addr = (vm_map_offset_t) mptr->PhysicalStart;
DPRINTF(" Type: %x phys: %p EFIv: %p kv: %p size: %p\n",
mptr->Type,
(void *) (uintptr_t) phys_addr,
(void *) (uintptr_t) mptr->VirtualStart,
(void *) vm_addr,
(void *) vm_size);
pmap_map_bd(vm_addr, phys_addr, phys_addr + round_page(vm_size),
(mptr->Type == kEfiRuntimeServicesCode) ? VM_PROT_READ | VM_PROT_EXECUTE : VM_PROT_READ | VM_PROT_WRITE,
(mptr->Type == EfiMemoryMappedIO) ? VM_WIMG_IO : VM_WIMG_USE_DEFAULT);
}
}
if (args->Version != kBootArgsVersion2) {
panic("Incompatible boot args version %d revision %d\n", args->Version, args->Revision);
}
DPRINTF("Boot args version %d revision %d mode %d\n", args->Version, args->Revision, args->efiMode);
if (args->efiMode == kBootArgsEfiMode64) {
efi_set_tables_64((EFI_SYSTEM_TABLE_64 *) ml_static_ptovirt(args->efiSystemTable));
} else {
efi_set_tables_32((EFI_SYSTEM_TABLE_32 *) ml_static_ptovirt(args->efiSystemTable));
}
}while (FALSE);
return;
}
boolean_t
bootloader_valid_page(ppnum_t ppn)
{
boot_args *args = (boot_args *)PE_state.bootArgs;
ppnum_t pstart = args->efiRuntimeServicesPageStart;
ppnum_t pend = pstart + args->efiRuntimeServicesPageCount;
return pstart <= ppn && ppn < pend;
}
void
hibernate_newruntime_map(void * map, vm_size_t map_size, uint32_t system_table_offset)
{
boot_args *args = (boot_args *)PE_state.bootArgs;
kprintf("Reinitializing EFI runtime services\n");
do{
vm_offset_t vm_size, vm_addr;
vm_map_offset_t phys_addr;
EfiMemoryRange *mptr;
unsigned int msize, mcount;
unsigned int i;
gPEEFISystemTable = 0;
gPEEFIRuntimeServices = 0;
system_table_offset += ptoa_32(args->efiRuntimeServicesPageStart);
kprintf("Old system table 0x%x, new 0x%x\n",
(uint32_t)args->efiSystemTable, system_table_offset);
args->efiSystemTable = system_table_offset;
kprintf("Old map:\n");
msize = args->MemoryMapDescriptorSize;
mcount = args->MemoryMapSize / msize;
mptr = (EfiMemoryRange *)ml_static_ptovirt(args->MemoryMap);
for (i = 0; i < mcount; i++, mptr = (EfiMemoryRange *)(((vm_offset_t)mptr) + msize)) {
if ((mptr->Attribute & EFI_MEMORY_RUNTIME) == EFI_MEMORY_RUNTIME) {
vm_size = (vm_offset_t)i386_ptob((uint32_t)mptr->NumberOfPages);
vm_addr = (vm_offset_t) mptr->VirtualStart;
if (vm_addr < VM_MIN_KERNEL_ADDRESS) {
vm_addr |= VM_MIN_KERNEL_ADDRESS;
}
phys_addr = (vm_map_offset_t) mptr->PhysicalStart;
kprintf("mapping[%u] %qx @ %lx, %llu\n", mptr->Type, phys_addr, (unsigned long)vm_addr, mptr->NumberOfPages);
}
}
pmap_remove(kernel_pmap, i386_ptob(args->efiRuntimeServicesPageStart),
i386_ptob(args->efiRuntimeServicesPageStart + args->efiRuntimeServicesPageCount));
kprintf("New map:\n");
msize = args->MemoryMapDescriptorSize;
mcount = (unsigned int)(map_size / msize);
mptr = map;
for (i = 0; i < mcount; i++, mptr = (EfiMemoryRange *)(((vm_offset_t)mptr) + msize)) {
if ((mptr->Attribute & EFI_MEMORY_RUNTIME) == EFI_MEMORY_RUNTIME) {
vm_size = (vm_offset_t)i386_ptob((uint32_t)mptr->NumberOfPages);
vm_addr = (vm_offset_t) mptr->VirtualStart;
if (vm_addr < VM_MIN_KERNEL_ADDRESS) {
vm_addr |= VM_MIN_KERNEL_ADDRESS;
}
phys_addr = (vm_map_offset_t) mptr->PhysicalStart;
kprintf("mapping[%u] %qx @ %lx, %llu\n", mptr->Type, phys_addr, (unsigned long)vm_addr, mptr->NumberOfPages);
pmap_map(vm_addr, phys_addr, phys_addr + round_page(vm_size),
(mptr->Type == kEfiRuntimeServicesCode) ? VM_PROT_READ | VM_PROT_EXECUTE : VM_PROT_READ | VM_PROT_WRITE,
(mptr->Type == EfiMemoryMappedIO) ? VM_WIMG_IO : VM_WIMG_USE_DEFAULT);
}
}
if (args->Version != kBootArgsVersion2) {
panic("Incompatible boot args version %d revision %d\n", args->Version, args->Revision);
}
kprintf("Boot args version %d revision %d mode %d\n", args->Version, args->Revision, args->efiMode);
if (args->efiMode == kBootArgsEfiMode64) {
efi_set_tables_64((EFI_SYSTEM_TABLE_64 *) ml_static_ptovirt(args->efiSystemTable));
} else {
efi_set_tables_32((EFI_SYSTEM_TABLE_32 *) ml_static_ptovirt(args->efiSystemTable));
}
}while (FALSE);
kprintf("Done reinitializing EFI runtime services\n");
return;
}
void
machine_init(void)
{
cpu_data_realloc();
debug_log_init();
cpuid_cpu_display("CPU identification");
cpuid_feature_display("CPU features");
cpuid_extfeature_display("CPU extended features");
efi_init();
smp_init();
init_fpu();
clock_config();
#if CONFIG_MTRR
mtrr_init();
pat_init();
#endif
pmap_lowmem_finalize();
}
void
halt_cpu(void)
{
halt_all_cpus(FALSE);
}
int reset_mem_on_reboot = 1;
__attribute__((noreturn))
void
halt_all_cpus(boolean_t reboot)
{
if (reboot) {
printf("MACH Reboot\n");
PEHaltRestart( kPERestartCPU );
} else {
printf("CPU halted\n");
PEHaltRestart( kPEHaltCPU );
}
while (1) {
;
}
}
void
panic_io_port_read(void)
{
if (panic_io_port) {
(void)inb(panic_io_port);
}
}
uint64_t panic_restart_timeout = ~(0ULL);
#define PANIC_RESTART_TIMEOUT (3ULL * NSEC_PER_SEC)
void
RecordPanicStackshot()
{
int err = 0;
size_t bytes_traced = 0, bytes_used = 0, bytes_remaining = 0;
char *stackshot_begin_loc = NULL;
if (begun_panic_stackshot) {
if (panic_info->mph_other_log_offset == 0) {
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
}
return;
}
begun_panic_stackshot = TRUE;
if (panic_info->mph_panic_log_len == 0) {
kdb_printf("Found zero length panic log, skipping capturing panic stackshot\n");
if (panic_info->mph_other_log_offset == 0) {
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
}
return;
}
if (stackshot_active()) {
panic_info->mph_panic_flags |= MACOS_PANIC_HEADER_FLAG_STACKSHOT_FAILED_NESTED;
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
kdb_printf("Panicked during stackshot, skipping panic stackshot\n");
return;
}
if (extended_debug_log_enabled) {
stackshot_begin_loc = debug_buf_ptr;
bytes_remaining = debug_buf_size - (unsigned int)((uintptr_t)stackshot_begin_loc - (uintptr_t)debug_buf_base);
} else if (panic_stackshot_buf != 0) {
stackshot_begin_loc = (char *) panic_stackshot_buf;
bytes_remaining = panic_stackshot_buf_len;
} else {
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
return;
}
err = kcdata_memory_static_init(&kc_panic_data, (mach_vm_address_t)stackshot_begin_loc,
KCDATA_BUFFER_BEGIN_STACKSHOT, (unsigned int) bytes_remaining, KCFLAG_USE_MEMCOPY);
if (err != KERN_SUCCESS) {
panic_info->mph_panic_flags |= MACOS_PANIC_HEADER_FLAG_STACKSHOT_FAILED_ERROR;
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
kdb_printf("Failed to initialize kcdata buffer for in-memory panic stackshot, skipping ...\n");
return;
}
uint32_t stackshot_flags = (STACKSHOT_SAVE_KEXT_LOADINFO | STACKSHOT_SAVE_LOADINFO | STACKSHOT_KCDATA_FORMAT |
STACKSHOT_ENABLE_BT_FAULTING | STACKSHOT_ENABLE_UUID_FAULTING | STACKSHOT_FROM_PANIC |
STACKSHOT_NO_IO_STATS | STACKSHOT_THREAD_WAITINFO);
#if DEVELOPMENT
stackshot_flags |= STACKSHOT_COLLECT_SHAREDCACHE_LAYOUT;
#endif
kdp_snapshot_preflight(-1, (void *) stackshot_begin_loc, (uint32_t) bytes_remaining, stackshot_flags, &kc_panic_data, 0);
err = do_stackshot(NULL);
bytes_traced = (int) kdp_stack_snapshot_bytes_traced();
bytes_used = (int) kcdata_memory_get_used_bytes(&kc_panic_data);
if ((err != KERN_SUCCESS) && (bytes_used > 0)) {
panic_info->mph_panic_flags |= MACOS_PANIC_HEADER_FLAG_STACKSHOT_KERNEL_ONLY;
panic_stackshot_reset_state();
memset(stackshot_begin_loc, '\0', bytes_used);
err = kcdata_memory_static_init(&kc_panic_data, (mach_vm_address_t)stackshot_begin_loc,
KCDATA_BUFFER_BEGIN_STACKSHOT, (unsigned int) bytes_remaining, KCFLAG_USE_MEMCOPY);
if (err != KERN_SUCCESS) {
panic_info->mph_panic_flags |= MACOS_PANIC_HEADER_FLAG_STACKSHOT_FAILED_ERROR;
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
kdb_printf("Failed to re-initialize kcdata buffer for kernel only in-memory panic stackshot, skipping ...\n");
return;
}
stackshot_flags = (STACKSHOT_SAVE_KEXT_LOADINFO | STACKSHOT_KCDATA_FORMAT | STACKSHOT_FROM_PANIC |
STACKSHOT_NO_IO_STATS | STACKSHOT_THREAD_WAITINFO | STACKSHOT_ACTIVE_KERNEL_THREADS_ONLY);
#if DEVELOPMENT
stackshot_flags |= STACKSHOT_COLLECT_SHAREDCACHE_LAYOUT;
#endif
kdp_snapshot_preflight(-1, (void *) stackshot_begin_loc, (uint32_t) bytes_remaining, stackshot_flags, &kc_panic_data, 0);
err = do_stackshot(NULL);
bytes_traced = (int) kdp_stack_snapshot_bytes_traced();
bytes_used = (int) kcdata_memory_get_used_bytes(&kc_panic_data);
}
if (err == KERN_SUCCESS) {
if (extended_debug_log_enabled) {
debug_buf_ptr += bytes_traced;
}
panic_info->mph_panic_flags |= MACOS_PANIC_HEADER_FLAG_STACKSHOT_SUCCEEDED;
panic_info->mph_stackshot_offset = PE_get_offset_into_panic_region(stackshot_begin_loc);
panic_info->mph_stackshot_len = (uint32_t) bytes_traced;
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
kdb_printf("\n** In Memory Panic Stackshot Succeeded ** Bytes Traced %zu **\n", bytes_traced);
panic_stackshot_buf = (vm_offset_t) stackshot_begin_loc;
panic_stackshot_len = bytes_traced;
if (!extended_debug_log_enabled && (gIOPolledCoreFileMode == kIOPolledCoreFileModeStackshot)) {
kern_dump(KERN_DUMP_STACKSHOT_DISK);
}
} else {
if (bytes_used > 0) {
memset(stackshot_begin_loc, '\0', bytes_used);
panic_info->mph_panic_flags |= MACOS_PANIC_HEADER_FLAG_STACKSHOT_FAILED_INCOMPLETE;
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
kdb_printf("\n** In Memory Panic Stackshot Incomplete ** Bytes Filled %zu ** Err %d\n", bytes_used, err);
} else {
bzero(stackshot_begin_loc, bytes_used);
panic_info->mph_panic_flags |= MACOS_PANIC_HEADER_FLAG_STACKSHOT_FAILED_ERROR;
panic_info->mph_other_log_offset = PE_get_offset_into_panic_region(debug_buf_ptr);
kdb_printf("\n** In Memory Panic Stackshot Failed ** Bytes Traced %zu, err %d\n", bytes_traced, err);
}
}
return;
}
void
SavePanicInfo(
__unused const char *message, void *panic_data, uint64_t panic_options)
{
void *stackptr = NULL;
thread_t thread_to_trace = (thread_t) panic_data;
cframe_t synthetic_stack_frame = { };
char *debugger_msg = NULL;
int cn = cpu_number();
panic_io_port_read();
if (panic_options & DEBUGGER_INTERNAL_OPTION_THREAD_BACKTRACE) {
if (!mp_kdp_all_cpus_halted()) {
debugger_msg = "Backtracing panicked thread because failed to halt all CPUs\n";
} else if (thread_to_trace == THREAD_NULL) {
debugger_msg = "Backtracing panicked thread because no thread pointer provided\n";
} else if (kvtophys((vm_offset_t)thread_to_trace) == 0ULL) {
debugger_msg = "Backtracing panicked thread because unable to access specified thread\n";
} else if (thread_to_trace->kernel_stack == 0) {
debugger_msg = "Backtracing panicked thread because kernel_stack is NULL for specified thread\n";
} else if (kvtophys(STACK_IKS(thread_to_trace->kernel_stack) == 0ULL)) {
debugger_msg = "Backtracing panicked thread because unable to access kernel_stack for specified thread\n";
} else {
debugger_msg = "Backtracing specified thread\n";
synthetic_stack_frame.prev = (cframe_t *)STACK_IKS(thread_to_trace->kernel_stack)->k_rbp;
synthetic_stack_frame.caller = (uintptr_t) STACK_IKS(thread_to_trace->kernel_stack)->k_rip;
stackptr = (void *) &synthetic_stack_frame;
}
}
if (stackptr == NULL) {
__asm__ volatile ("movq %%rbp, %0" : "=m" (stackptr));
}
if (panic_options & DEBUGGER_OPTION_INITPROC_PANIC) {
print_launchd_info();
} else {
panic_i386_backtrace(stackptr, ((panic_double_fault_cpu == cn) ? 80: 48), debugger_msg, FALSE, NULL);
}
if (panic_options & DEBUGGER_OPTION_COPROC_INITIATED_PANIC) {
panic_info->mph_panic_flags |= MACOS_PANIC_HEADER_FLAG_COPROC_INITIATED_PANIC;
}
if (PE_get_offset_into_panic_region(debug_buf_ptr) < panic_info->mph_panic_log_offset) {
kdb_printf("Invalid panic log offset found (not properly initialized?): debug_buf_ptr : 0x%p, panic_info: 0x%p mph_panic_log_offset: 0x%x\n",
debug_buf_ptr, panic_info, panic_info->mph_panic_log_offset);
panic_info->mph_panic_log_len = 0;
} else {
panic_info->mph_panic_log_len = PE_get_offset_into_panic_region(debug_buf_ptr) - panic_info->mph_panic_log_offset;
}
paniclog_flush_internal(kPaniclogFlushBase);
RecordPanicStackshot();
paniclog_flush_internal(kPaniclogFlushStackshot);
}
void
paniclog_flush_internal(paniclog_flush_type_t variant)
{
if (panic_info->mph_other_log_offset != 0) {
panic_info->mph_other_log_len = PE_get_offset_into_panic_region(debug_buf_ptr) - panic_info->mph_other_log_offset;
}
if (coprocessor_paniclog_flush) {
uint32_t overall_buffer_size = debug_buf_size;
uint32_t size_to_flush = 0, offset_to_flush = 0;
if (extended_debug_log_enabled) {
overall_buffer_size = debug_buf_size + sizeof(struct macos_panic_header);
}
panic_info->mph_crc = crc32(0L, &panic_info->mph_version, (overall_buffer_size - offsetof(struct macos_panic_header, mph_version)));
if (variant == kPaniclogFlushBase) {
kprintf("Flushing base panic log\n");
size_to_flush = ROUNDUP((panic_info->mph_panic_log_offset + panic_info->mph_panic_log_len), PANIC_FLUSH_BOUNDARY);
offset_to_flush = 0;
PESavePanicInfoAction(panic_info, offset_to_flush, size_to_flush);
} else if ((variant == kPaniclogFlushStackshot) || (variant == kPaniclogFlushOtherLog)) {
if (variant == kPaniclogFlushStackshot) {
kprintf("Flushing panic log stackshot\n");
offset_to_flush = ROUNDDOWN((panic_info->mph_panic_log_offset + panic_info->mph_panic_log_len), PANIC_FLUSH_BOUNDARY);
size_to_flush = ROUNDUP((panic_info->mph_stackshot_len + (panic_info->mph_stackshot_offset - offset_to_flush)), PANIC_FLUSH_BOUNDARY);
PESavePanicInfoAction(panic_info, offset_to_flush, size_to_flush);
}
kprintf("Flushing panic 'other' log\n");
offset_to_flush = ROUNDDOWN((panic_info->mph_stackshot_offset + panic_info->mph_stackshot_len), PANIC_FLUSH_BOUNDARY);
size_to_flush = ROUNDUP((panic_info->mph_other_log_len + (panic_info->mph_other_log_offset - offset_to_flush)), PANIC_FLUSH_BOUNDARY);
PESavePanicInfoAction(panic_info, offset_to_flush, size_to_flush);
kprintf("Flushing panic log header\n");
size_to_flush = ROUNDUP(panic_info->mph_panic_log_offset, PANIC_FLUSH_BOUNDARY);
offset_to_flush = 0;
PESavePanicInfoAction(panic_info, offset_to_flush, size_to_flush);
}
} else if (commit_paniclog_to_nvram) {
assert(debug_buf_size != 0);
unsigned int bufpos;
unsigned long pi_size = 0;
uintptr_t cr0;
debug_putc(0);
bufpos = packA(debug_buf_base, (unsigned int) (debug_buf_ptr - debug_buf_base),
debug_buf_size);
pi_size = bufpos ? bufpos : (unsigned) (debug_buf_ptr - debug_buf_base);
cr0 = get_cr0();
clear_ts();
kprintf("Attempting to commit panic log to NVRAM\n");
pi_size = PESavePanicInfo((unsigned char *)debug_buf_base,
(uint32_t)pi_size );
set_cr0(cr0);
if (bufpos) {
unpackA(debug_buf_base, bufpos);
}
}
}
void
paniclog_flush()
{
paniclog_flush_internal(kPaniclogFlushOtherLog);
return;
}
char *
machine_boot_info(char *buf, __unused vm_size_t size)
{
*buf = '\0';
return buf;
}
static int
panic_print_macho_symbol_name(kernel_mach_header_t *mh, vm_address_t search, const char *module_name)
{
kernel_nlist_t *sym = NULL;
struct load_command *cmd;
kernel_segment_command_t *orig_ts = NULL, *orig_le = NULL;
struct symtab_command *orig_st = NULL;
unsigned int i;
char *strings, *bestsym = NULL;
vm_address_t bestaddr = 0, diff, curdiff;
cmd = (struct load_command *) &mh[1];
for (i = 0; i < mh->ncmds; i++) {
if (cmd->cmd == LC_SEGMENT_KERNEL) {
kernel_segment_command_t *orig_sg = (kernel_segment_command_t *) cmd;
if (strncmp(SEG_TEXT, orig_sg->segname,
sizeof(orig_sg->segname)) == 0) {
orig_ts = orig_sg;
} else if (strncmp(SEG_LINKEDIT, orig_sg->segname,
sizeof(orig_sg->segname)) == 0) {
orig_le = orig_sg;
} else if (strncmp("", orig_sg->segname,
sizeof(orig_sg->segname)) == 0) {
orig_ts = orig_sg;
}
} else if (cmd->cmd == LC_SYMTAB) {
orig_st = (struct symtab_command *) cmd;
}
cmd = (struct load_command *) ((uintptr_t) cmd + cmd->cmdsize);
}
if ((orig_ts == NULL) || (orig_st == NULL) || (orig_le == NULL)) {
return 0;
}
if ((search < orig_ts->vmaddr) ||
(search >= orig_ts->vmaddr + orig_ts->vmsize)) {
return 0;
}
sym = (kernel_nlist_t *)(uintptr_t)(orig_le->vmaddr + orig_st->symoff - orig_le->fileoff);
strings = (char *)(uintptr_t)(orig_le->vmaddr + orig_st->stroff - orig_le->fileoff);
diff = search;
for (i = 0; i < orig_st->nsyms; i++) {
if (sym[i].n_type & N_STAB) {
continue;
}
if (sym[i].n_value <= search) {
curdiff = search - (vm_address_t)sym[i].n_value;
if (curdiff < diff) {
diff = curdiff;
bestaddr = sym[i].n_value;
bestsym = strings + sym[i].n_un.n_strx;
}
}
}
if (bestsym != NULL) {
if (diff != 0) {
paniclog_append_noflush("%s : %s + 0x%lx", module_name, bestsym, (unsigned long)diff);
} else {
paniclog_append_noflush("%s : %s", module_name, bestsym);
}
return 1;
}
return 0;
}
extern kmod_info_t * kmod;
static void
panic_print_kmod_symbol_name(vm_address_t search)
{
u_int i;
if (gLoadedKextSummaries == NULL) {
return;
}
for (i = 0; i < gLoadedKextSummaries->numSummaries; ++i) {
OSKextLoadedKextSummary *summary = gLoadedKextSummaries->summaries + i;
if ((search >= summary->address) &&
(search < (summary->address + summary->size))) {
kernel_mach_header_t *header = (kernel_mach_header_t *)(uintptr_t) summary->address;
if (panic_print_macho_symbol_name(header, search, summary->name) == 0) {
paniclog_append_noflush("%s + %llu", summary->name, (unsigned long)search - summary->address);
}
break;
}
}
}
void
panic_print_symbol_name(vm_address_t search)
{
if (panic_print_macho_symbol_name(&_mh_execute_header, search, "mach_kernel") == 0) {
panic_print_kmod_symbol_name(search);
}
}
#define DUMPFRAMES 32
#define PBT_TIMEOUT_CYCLES (5 * 1000 * 1000 * 1000ULL)
void
panic_i386_backtrace(void *_frame, int nframes, const char *msg, boolean_t regdump, x86_saved_state_t *regs)
{
cframe_t *frame = (cframe_t *)_frame;
vm_offset_t raddrs[DUMPFRAMES];
vm_offset_t PC = 0;
int frame_index;
volatile uint32_t *ppbtcnt = &pbtcnt;
uint64_t bt_tsc_timeout;
boolean_t keepsyms = FALSE;
int cn = cpu_number();
boolean_t old_doprnt_hide_pointers = doprnt_hide_pointers;
if (pbtcpu != cn) {
os_atomic_inc(&pbtcnt, relaxed);
hw_lock_to(&pbtlock, ~0U, LCK_GRP_NULL);
pbtcpu = cn;
}
if (__improbable(doprnt_hide_pointers == TRUE)) {
doprnt_hide_pointers = FALSE;
}
panic_check_hook();
PE_parse_boot_argn("keepsyms", &keepsyms, sizeof(keepsyms));
if (msg != NULL) {
paniclog_append_noflush("%s", msg);
}
if ((regdump == TRUE) && (regs != NULL)) {
x86_saved_state64_t *ss64p = saved_state64(regs);
paniclog_append_noflush(
"RAX: 0x%016llx, RBX: 0x%016llx, RCX: 0x%016llx, RDX: 0x%016llx\n"
"RSP: 0x%016llx, RBP: 0x%016llx, RSI: 0x%016llx, RDI: 0x%016llx\n"
"R8: 0x%016llx, R9: 0x%016llx, R10: 0x%016llx, R11: 0x%016llx\n"
"R12: 0x%016llx, R13: 0x%016llx, R14: 0x%016llx, R15: 0x%016llx\n"
"RFL: 0x%016llx, RIP: 0x%016llx, CS: 0x%016llx, SS: 0x%016llx\n",
ss64p->rax, ss64p->rbx, ss64p->rcx, ss64p->rdx,
ss64p->isf.rsp, ss64p->rbp, ss64p->rsi, ss64p->rdi,
ss64p->r8, ss64p->r9, ss64p->r10, ss64p->r11,
ss64p->r12, ss64p->r13, ss64p->r14, ss64p->r15,
ss64p->isf.rflags, ss64p->isf.rip, ss64p->isf.cs,
ss64p->isf.ss);
PC = ss64p->isf.rip;
}
paniclog_append_noflush("Backtrace (CPU %d), "
#if PRINT_ARGS_FROM_STACK_FRAME
"Frame : Return Address (4 potential args on stack)\n", cn);
#else
"Frame : Return Address\n", cn);
#endif
for (frame_index = 0; frame_index < nframes; frame_index++) {
vm_offset_t curframep = (vm_offset_t) frame;
if (!curframep) {
break;
}
if (curframep & 0x3) {
paniclog_append_noflush("Unaligned frame\n");
goto invalid;
}
if (!kvtophys(curframep) ||
!kvtophys(curframep + sizeof(cframe_t) - 1)) {
paniclog_append_noflush("No mapping exists for frame pointer\n");
goto invalid;
}
paniclog_append_noflush("%p : 0x%lx ", frame, frame->caller);
if (frame_index < DUMPFRAMES) {
raddrs[frame_index] = frame->caller;
}
#if PRINT_ARGS_FROM_STACK_FRAME
if (kvtophys((vm_offset_t)&(frame->args[3]))) {
paniclog_append_noflush("(0x%x 0x%x 0x%x 0x%x) ",
frame->args[0], frame->args[1],
frame->args[2], frame->args[3]);
}
#endif
if (keepsyms) {
panic_print_symbol_name((vm_address_t)frame->caller);
}
paniclog_append_noflush("\n");
frame = frame->prev;
}
if (frame_index >= nframes) {
paniclog_append_noflush("\tBacktrace continues...\n");
}
goto out;
invalid:
paniclog_append_noflush("Backtrace terminated-invalid frame pointer %p\n", frame);
out:
if (frame_index) {
kmod_panic_dump((vm_offset_t *)&raddrs[0], frame_index);
}
if (PC != 0) {
kmod_panic_dump(&PC, 1);
}
panic_display_system_configuration(FALSE);
doprnt_hide_pointers = old_doprnt_hide_pointers;
hw_lock_unlock(&pbtlock);
os_atomic_dec(&pbtcnt, relaxed);
bt_tsc_timeout = rdtsc64() + PBT_TIMEOUT_CYCLES;
while (*ppbtcnt && (rdtsc64() < bt_tsc_timeout)) {
;
}
}
static boolean_t
debug_copyin(pmap_t p, uint64_t uaddr, void *dest, size_t size)
{
size_t rem = size;
char *kvaddr = dest;
while (rem) {
ppnum_t upn = pmap_find_phys(p, uaddr);
uint64_t phys_src = ptoa_64(upn) | (uaddr & PAGE_MASK);
uint64_t phys_dest = kvtophys((vm_offset_t)kvaddr);
uint64_t src_rem = PAGE_SIZE - (phys_src & PAGE_MASK);
uint64_t dst_rem = PAGE_SIZE - (phys_dest & PAGE_MASK);
size_t cur_size = (uint32_t) MIN(src_rem, dst_rem);
cur_size = MIN(cur_size, rem);
if (upn && pmap_valid_page(upn) && phys_dest) {
bcopy_phys(phys_src, phys_dest, cur_size);
} else {
break;
}
uaddr += cur_size;
kvaddr += cur_size;
rem -= cur_size;
}
return rem == 0;
}
void
print_threads_registers(thread_t thread)
{
x86_saved_state_t *savestate;
savestate = get_user_regs(thread);
paniclog_append_noflush(
"\nRAX: 0x%016llx, RBX: 0x%016llx, RCX: 0x%016llx, RDX: 0x%016llx\n"
"RSP: 0x%016llx, RBP: 0x%016llx, RSI: 0x%016llx, RDI: 0x%016llx\n"
"R8: 0x%016llx, R9: 0x%016llx, R10: 0x%016llx, R11: 0x%016llx\n"
"R12: 0x%016llx, R13: 0x%016llx, R14: 0x%016llx, R15: 0x%016llx\n"
"RFL: 0x%016llx, RIP: 0x%016llx, CS: 0x%016llx, SS: 0x%016llx\n\n",
savestate->ss_64.rax, savestate->ss_64.rbx, savestate->ss_64.rcx, savestate->ss_64.rdx,
savestate->ss_64.isf.rsp, savestate->ss_64.rbp, savestate->ss_64.rsi, savestate->ss_64.rdi,
savestate->ss_64.r8, savestate->ss_64.r9, savestate->ss_64.r10, savestate->ss_64.r11,
savestate->ss_64.r12, savestate->ss_64.r13, savestate->ss_64.r14, savestate->ss_64.r15,
savestate->ss_64.isf.rflags, savestate->ss_64.isf.rip, savestate->ss_64.isf.cs,
savestate->ss_64.isf.ss);
}
void
print_tasks_user_threads(task_t task)
{
thread_t thread = current_thread();
x86_saved_state_t *savestate;
pmap_t pmap = 0;
uint64_t rbp;
const char *cur_marker = 0;
int j;
for (j = 0, thread = (thread_t) queue_first(&task->threads); j < task->thread_count;
++j, thread = (thread_t) queue_next(&thread->task_threads)) {
paniclog_append_noflush("Thread %d: %p\n", j, thread);
pmap = get_task_pmap(task);
savestate = get_user_regs(thread);
rbp = savestate->ss_64.rbp;
paniclog_append_noflush("\t0x%016llx\n", savestate->ss_64.isf.rip);
print_one_backtrace(pmap, (vm_offset_t)rbp, cur_marker, TRUE);
paniclog_append_noflush("\n");
}
}
void
print_thread_num_that_crashed(task_t task)
{
thread_t c_thread = current_thread();
thread_t thread;
int j;
for (j = 0, thread = (thread_t) queue_first(&task->threads); j < task->thread_count;
++j, thread = (thread_t) queue_next(&thread->task_threads)) {
if (c_thread == thread) {
paniclog_append_noflush("\nThread %d crashed\n", j);
break;
}
}
}
#define PANICLOG_UUID_BUF_SIZE 256
void
print_uuid_info(task_t task)
{
uint32_t uuid_info_count = 0;
mach_vm_address_t uuid_info_addr = 0;
boolean_t have_map = (task->map != NULL) && (ml_validate_nofault((vm_offset_t)(task->map), sizeof(struct _vm_map)));
boolean_t have_pmap = have_map && (task->map->pmap != NULL) && (ml_validate_nofault((vm_offset_t)(task->map->pmap), sizeof(struct pmap)));
int task_pid = pid_from_task(task);
char uuidbuf[PANICLOG_UUID_BUF_SIZE] = {0};
char *uuidbufptr = uuidbuf;
uint32_t k;
if (have_pmap && task->active && task_pid > 0) {
struct user64_dyld_all_image_infos task_image_infos;
if (debug_copyin(task->map->pmap, task->all_image_info_addr,
&task_image_infos, sizeof(struct user64_dyld_all_image_infos))) {
uuid_info_count = (uint32_t)task_image_infos.uuidArrayCount;
uuid_info_addr = task_image_infos.uuidArray;
}
if (!uuid_info_addr) {
uuid_info_count = 0;
}
}
if (task_pid > 0 && uuid_info_count > 0) {
uint32_t uuid_info_size = sizeof(struct user64_dyld_uuid_info);
uint32_t uuid_array_size = uuid_info_count * uuid_info_size;
uint32_t uuid_copy_size = 0;
uint32_t uuid_image_count = 0;
char *current_uuid_buffer = NULL;
paniclog_append_noflush("\nuuid info:\n");
while (uuid_array_size) {
if (uuid_array_size <= PANICLOG_UUID_BUF_SIZE) {
uuid_copy_size = uuid_array_size;
uuid_image_count = uuid_array_size / uuid_info_size;
} else {
uuid_image_count = PANICLOG_UUID_BUF_SIZE / uuid_info_size;
uuid_copy_size = uuid_image_count * uuid_info_size;
}
if (have_pmap && !debug_copyin(task->map->pmap, uuid_info_addr, uuidbufptr,
uuid_copy_size)) {
paniclog_append_noflush("Error!! Failed to copy UUID info for task %p pid %d\n", task, task_pid);
uuid_image_count = 0;
break;
}
if (uuid_image_count > 0) {
current_uuid_buffer = uuidbufptr;
for (k = 0; k < uuid_image_count; k++) {
paniclog_append_noflush(" %#llx", *(uint64_t *)current_uuid_buffer);
current_uuid_buffer += sizeof(uint64_t);
uint8_t *uuid = (uint8_t *)current_uuid_buffer;
paniclog_append_noflush("\tuuid = <%02x%02x%02x%02x-%02x%02x-%02x%02x-%02x%02x-%02x%02x%02x%02x%02x%02x>\n",
uuid[0], uuid[1], uuid[2], uuid[3], uuid[4], uuid[5], uuid[6], uuid[7], uuid[8],
uuid[9], uuid[10], uuid[11], uuid[12], uuid[13], uuid[14], uuid[15]);
current_uuid_buffer += 16;
}
bzero(&uuidbuf, sizeof(uuidbuf));
}
uuid_info_addr += uuid_copy_size;
uuid_array_size -= uuid_copy_size;
}
}
}
void
print_launchd_info(void)
{
task_t task = current_task();
thread_t thread = current_thread();
volatile uint32_t *ppbtcnt = &pbtcnt;
uint64_t bt_tsc_timeout;
int cn = cpu_number();
if (pbtcpu != cn) {
os_atomic_inc(&pbtcnt, relaxed);
hw_lock_to(&pbtlock, ~0U, LCK_GRP_NULL);
pbtcpu = cn;
}
print_uuid_info(task);
print_thread_num_that_crashed(task);
print_threads_registers(thread);
print_tasks_user_threads(task);
panic_display_system_configuration(TRUE);
hw_lock_unlock(&pbtlock);
os_atomic_dec(&pbtcnt, relaxed);
bt_tsc_timeout = rdtsc64() + PBT_TIMEOUT_CYCLES;
while (*ppbtcnt && (rdtsc64() < bt_tsc_timeout)) {
;
}
}