#include <zone_debug.h>
#include <zone_alias_addr.h>
#include <mach/mach_types.h>
#include <mach/vm_param.h>
#include <mach/kern_return.h>
#include <mach/machine/vm_types.h>
#include <mach_debug/zone_info.h>
#include <mach/vm_map.h>
#include <kern/kern_types.h>
#include <kern/assert.h>
#include <kern/sched.h>
#include <kern/locks.h>
#include <kern/misc_protos.h>
#include <kern/zalloc.h>
#include <kern/kalloc.h>
#include <vm/pmap.h>
#include <vm/vm_map.h>
#include <vm/vm_kern.h>
#include <vm/vm_page.h>
#include <pexpert/pexpert.h>
#include <machine/machparam.h>
#include <libkern/OSDebug.h>
#include <libkern/OSAtomic.h>
#include <sys/kdebug.h>
extern boolean_t vm_kernel_ready, kmem_ready;
boolean_t gzalloc_mode = FALSE;
uint32_t pdzalloc_count, pdzfree_count;
#define GZALLOC_MIN_DEFAULT (1024)
#define GZDEADZONE ((zone_t) 0xDEAD201E)
#define GZALLOC_SIGNATURE (0xABADCAFE)
#define GZALLOC_RESERVE_SIZE_DEFAULT (2 * 1024 * 1024)
#define GZFC_DEFAULT_SIZE (1024)
char gzalloc_fill_pattern = 0x67;
uint32_t gzalloc_min = ~0U;
uint32_t gzalloc_max = 0;
uint32_t gzalloc_size = 0;
uint64_t gzalloc_allocated, gzalloc_freed, gzalloc_early_alloc, gzalloc_early_free, gzalloc_wasted;
boolean_t gzalloc_uf_mode = FALSE, gzalloc_consistency_checks = TRUE;
vm_prot_t gzalloc_prot = VM_PROT_NONE;
uint32_t gzalloc_guard = KMA_GUARD_LAST;
uint32_t gzfc_size = GZFC_DEFAULT_SIZE;
vm_map_t gzalloc_map;
vm_offset_t gzalloc_map_min, gzalloc_map_max;
vm_offset_t gzalloc_reserve;
vm_size_t gzalloc_reserve_size;
typedef struct gzalloc_header {
zone_t gzone;
uint32_t gzsize;
uint32_t gzsig;
} gzhdr_t;
#define GZHEADER_SIZE (sizeof(gzhdr_t))
extern zone_t vm_page_zone;
void gzalloc_reconfigure(__unused zone_t z) {
}
boolean_t gzalloc_enabled(void) {
return gzalloc_mode;
}
void gzalloc_zone_init(zone_t z) {
if (gzalloc_mode) {
bzero(&z->gz, sizeof(z->gz));
if (gzfc_size && (z->elem_size >= gzalloc_min) && (z->elem_size <= gzalloc_max) && (z->gzalloc_exempt == FALSE)) {
vm_size_t gzfcsz = round_page(sizeof(*z->gz.gzfc) * gzfc_size);
if (!kmem_ready) {
if (gzalloc_reserve_size < gzfcsz)
panic("gzalloc reserve exhausted");
z->gz.gzfc = (vm_offset_t *)gzalloc_reserve;
gzalloc_reserve += gzfcsz;
gzalloc_reserve_size -= gzfcsz;
} else {
kern_return_t kr;
if ((kr = kernel_memory_allocate(kernel_map, (vm_offset_t *)&z->gz.gzfc, gzfcsz, 0, KMA_KOBJECT)) != KERN_SUCCESS) {
panic("zinit/gzalloc: kernel_memory_allocate failed (%d) for 0x%lx bytes", kr, (unsigned long) gzfcsz);
}
}
bzero((void *)z->gz.gzfc, gzfcsz);
}
}
}
void gzalloc_configure(void) {
char temp_buf[16];
if (PE_parse_boot_argn("-gzalloc_mode", temp_buf, sizeof (temp_buf))) {
gzalloc_mode = TRUE;
gzalloc_min = GZALLOC_MIN_DEFAULT;
#if ZONE_DEBUG
gzalloc_min += (typeof(gzalloc_min))ZONE_DEBUG_OFFSET;
#endif
gzalloc_max = ~0U;
}
if (PE_parse_boot_argn("gzalloc_min", &gzalloc_min, sizeof(gzalloc_min))) {
#if ZONE_DEBUG
gzalloc_min += (typeof(gzalloc_min))ZONE_DEBUG_OFFSET;
#endif
gzalloc_mode = TRUE;
gzalloc_max = ~0U;
}
if (PE_parse_boot_argn("gzalloc_max", &gzalloc_max, sizeof(gzalloc_max))) {
#if ZONE_DEBUG
gzalloc_max += (typeof(gzalloc_min))ZONE_DEBUG_OFFSET;
#endif
gzalloc_mode = TRUE;
if (gzalloc_min == ~0U)
gzalloc_min = 0;
}
if (PE_parse_boot_argn("gzalloc_size", &gzalloc_size, sizeof(gzalloc_size))) {
#if ZONE_DEBUG
gzalloc_size += (typeof(gzalloc_min))ZONE_DEBUG_OFFSET;
#endif
gzalloc_min = gzalloc_max = gzalloc_size;
gzalloc_mode = TRUE;
}
(void)PE_parse_boot_argn("gzalloc_fc_size", &gzfc_size, sizeof(gzfc_size));
if (PE_parse_boot_argn("-gzalloc_wp", temp_buf, sizeof (temp_buf))) {
gzalloc_prot = VM_PROT_READ;
}
if (PE_parse_boot_argn("-gzalloc_uf_mode", temp_buf, sizeof (temp_buf))) {
gzalloc_uf_mode = TRUE;
gzalloc_guard = KMA_GUARD_FIRST;
}
if (PE_parse_boot_argn("-gzalloc_noconsistency", temp_buf, sizeof (temp_buf))) {
gzalloc_consistency_checks = FALSE;
}
#if DEBUG
if (gzalloc_mode == FALSE) {
gzalloc_min = 8192;
gzalloc_max = 16384;
gzalloc_prot = VM_PROT_READ;
gzalloc_mode = TRUE;
}
#endif
if (PE_parse_boot_argn("-nogzalloc_mode", temp_buf, sizeof (temp_buf)))
gzalloc_mode = FALSE;
if (gzalloc_mode) {
gzalloc_reserve_size = GZALLOC_RESERVE_SIZE_DEFAULT;
gzalloc_reserve = (vm_offset_t) pmap_steal_memory(gzalloc_reserve_size);
}
}
void gzalloc_init(vm_size_t max_zonemap_size) {
kern_return_t retval;
if (gzalloc_mode) {
retval = kmem_suballoc(kernel_map, &gzalloc_map_min, (max_zonemap_size << 2),
FALSE, VM_FLAGS_ANYWHERE | VM_FLAGS_PERMANENT,
&gzalloc_map);
if (retval != KERN_SUCCESS)
panic("zone_init: kmem_suballoc(gzalloc) failed");
gzalloc_map_max = gzalloc_map_min + (max_zonemap_size << 2);
}
}
vm_offset_t
gzalloc_alloc(zone_t zone, boolean_t canblock) {
vm_offset_t addr = 0;
if (__improbable(gzalloc_mode &&
(((zone->elem_size >= gzalloc_min) &&
(zone->elem_size <= gzalloc_max))) &&
(zone->gzalloc_exempt == 0))) {
if (get_preemption_level() != 0) {
if (canblock == TRUE) {
pdzalloc_count++;
}
else
return 0;
}
vm_offset_t rounded_size = round_page(zone->elem_size + GZHEADER_SIZE);
vm_offset_t residue = rounded_size - zone->elem_size;
vm_offset_t gzaddr = 0;
gzhdr_t *gzh;
if (!kmem_ready || (vm_page_zone == ZONE_NULL)) {
if (gzalloc_reserve_size < rounded_size)
panic("gzalloc reserve exhausted");
gzaddr = gzalloc_reserve;
gzalloc_reserve += rounded_size + PAGE_SIZE;
gzalloc_reserve_size -= rounded_size + PAGE_SIZE;
OSAddAtomic64((SInt32) (rounded_size), &gzalloc_early_alloc);
}
else {
kern_return_t kr = kernel_memory_allocate(gzalloc_map,
&gzaddr, rounded_size + (1*PAGE_SIZE),
0, KMA_KOBJECT | gzalloc_guard);
if (kr != KERN_SUCCESS)
panic("gzalloc: kernel_memory_allocate for size 0x%llx failed with %d", (uint64_t)rounded_size, kr);
}
if (gzalloc_uf_mode) {
gzaddr += PAGE_SIZE;
gzh = (gzhdr_t *) (gzaddr + zone->elem_size);
addr = gzaddr;
} else {
gzh = (gzhdr_t *) (gzaddr + residue - GZHEADER_SIZE);
addr = (gzaddr + residue);
}
memset((void *)gzaddr, gzalloc_fill_pattern, rounded_size);
gzh->gzone = (kmem_ready && vm_page_zone) ? zone : GZDEADZONE;
gzh->gzsize = (uint32_t) zone->elem_size;
gzh->gzsig = GZALLOC_SIGNATURE;
lock_zone(zone);
zone->count++;
zone->sum_count++;
zone->cur_size += rounded_size;
unlock_zone(zone);
OSAddAtomic64((SInt32) rounded_size, &gzalloc_allocated);
OSAddAtomic64((SInt32) (rounded_size - zone->elem_size), &gzalloc_wasted);
}
return addr;
}
boolean_t gzalloc_free(zone_t zone, void *addr) {
boolean_t gzfreed = FALSE;
kern_return_t kr;
if (__improbable(gzalloc_mode &&
(((zone->elem_size >= gzalloc_min) &&
(zone->elem_size <= gzalloc_max))) &&
(zone->gzalloc_exempt == 0))) {
gzhdr_t *gzh;
vm_offset_t rounded_size = round_page(zone->elem_size + GZHEADER_SIZE);
vm_offset_t residue = rounded_size - zone->elem_size;
vm_offset_t saddr;
vm_offset_t free_addr = 0;
if (gzalloc_uf_mode) {
gzh = (gzhdr_t *)((vm_offset_t)addr + zone->elem_size);
saddr = (vm_offset_t) addr - PAGE_SIZE;
} else {
gzh = (gzhdr_t *)((vm_offset_t)addr - GZHEADER_SIZE);
saddr = ((vm_offset_t)addr) - residue;
}
assert((saddr & PAGE_MASK) == 0);
if (gzalloc_consistency_checks) {
if (gzh->gzsig != GZALLOC_SIGNATURE) {
panic("GZALLOC signature mismatch for element %p, expected 0x%x, found 0x%x", addr, GZALLOC_SIGNATURE, gzh->gzsig);
}
if (gzh->gzone != zone && (gzh->gzone != GZDEADZONE))
panic("%s: Mismatched zone or under/overflow, current zone: %p, recorded zone: %p, address: %p", __FUNCTION__, zone, gzh->gzone, (void *)addr);
if (gzh->gzsize != zone->elem_size) {
panic("Mismatched zfree or under/overflow for zone %p, recorded size: 0x%x, element size: 0x%x, address: %p\n", zone, gzh->gzsize, (uint32_t) zone->elem_size, (void *)addr);
}
}
if (!kmem_ready || gzh->gzone == GZDEADZONE) {
OSAddAtomic64((SInt32) (rounded_size), &gzalloc_early_free);
return TRUE;
}
if (get_preemption_level() != 0) {
pdzfree_count++;
}
if (gzfc_size) {
kr = vm_map_protect(
gzalloc_map,
saddr,
saddr + rounded_size + (1 * PAGE_SIZE),
gzalloc_prot,
FALSE);
if (kr != KERN_SUCCESS)
panic("%s: vm_map_protect: %p, 0x%x", __FUNCTION__, (void *)saddr, kr);
} else {
free_addr = saddr;
}
lock_zone(zone);
if (gzfc_size) {
if (zone->gz.gzfc_index >= gzfc_size) {
zone->gz.gzfc_index = 0;
}
free_addr = zone->gz.gzfc[zone->gz.gzfc_index];
zone->gz.gzfc[zone->gz.gzfc_index++] = saddr;
}
if (free_addr) {
zone->count--;
zone->cur_size -= rounded_size;
}
unlock_zone(zone);
if (free_addr) {
kr = vm_map_remove(
gzalloc_map,
free_addr,
free_addr + rounded_size + (1 * PAGE_SIZE),
VM_MAP_REMOVE_KUNWIRE);
if (kr != KERN_SUCCESS)
panic("gzfree: vm_map_remove: %p, 0x%x", (void *)free_addr, kr);
OSAddAtomic64((SInt32)rounded_size, &gzalloc_freed);
OSAddAtomic64(-((SInt32) (rounded_size - zone->elem_size)), &gzalloc_wasted);
}
gzfreed = TRUE;
}
return gzfreed;
}