temporary workarounds for bugs

Daniel Micay 2019-01-01 14:45:27 -05:00
parent 1984cb3b3d
commit dcb9981160
1 changed files with 32 additions and 7 deletions

View File

@ -76,6 +76,9 @@ static union {
#ifdef USE_PKEY
int metadata_pkey;
#endif
bool zero_on_free;
bool purge_slabs;
bool region_quarantine_protect;
};
char padding[PAGE_SIZE];
} ro __attribute__((aligned(PAGE_SIZE)));
@ -429,7 +432,7 @@ static void *slot_pointer(size_t size, void *slab, size_t slot) {
}
static void write_after_free_check(const char *p, size_t size) {
if (!WRITE_AFTER_FREE_CHECK) {
if (!WRITE_AFTER_FREE_CHECK || !ro.zero_on_free) {
return;
}
@ -671,7 +674,7 @@ static inline void deallocate_small(void *p, const size_t *expected_size) {
}
}
if (ZERO_ON_FREE) {
if (ro.zero_on_free) {
memset(p, 0, size - canary_size);
}
}
@ -748,7 +751,7 @@ static inline void deallocate_small(void *p, const size_t *expected_size) {
metadata->prev = NULL;
if (c->empty_slabs_total + slab_size > max_empty_slabs_total) {
if (!memory_map_fixed(slab, slab_size)) {
if (ro.purge_slabs && !memory_map_fixed(slab, slab_size)) {
label_slab(slab, slab_size, class);
stats_slab_deallocate(c, slab_size);
enqueue_free_slab(c, metadata);
@ -832,9 +835,13 @@ static void regions_quarantine_deallocate_pages(void *p, size_t size, size_t gua
return;
}
if (unlikely(memory_map_fixed(p, size))) {
deallocate_pages(p, size, guard_size);
return;
if (ro.region_quarantine_protect) {
if (unlikely(memory_map_fixed(p, size))) {
deallocate_pages(p, size, guard_size);
return;
}
} else {
madvise(p, size, MADV_DONTNEED);
}
memory_set_name(p, size, "malloc large quarantine");
@ -1048,6 +1055,19 @@ static inline void enforce_init(void) {
}
}
COLD static void handle_hal_bugs(void) {
char path[256];
if (readlink("/proc/self/exe", path, sizeof(path)) == -1) {
return;
}
const char camera_provider[] = "/vendor/bin/hw/android.hardware.camera.provider@2.4-service_64";
if (strcmp(camera_provider, path) == 0) {
ro.zero_on_free = false;
ro.purge_slabs = false;
ro.region_quarantine_protect = false;
}
}
COLD static void init_slow_path(void) {
static struct mutex lock = MUTEX_INITIALIZER;
@ -1062,6 +1082,11 @@ COLD static void init_slow_path(void) {
ro.metadata_pkey = pkey_alloc(0, 0);
#endif
ro.purge_slabs = true;
ro.zero_on_free = ZERO_ON_FREE;
ro.region_quarantine_protect = true;
handle_hal_bugs();
if (sysconf(_SC_PAGESIZE) != PAGE_SIZE) {
fatal_error("runtime page size does not match compile-time page size which is not supported");
}
@ -1333,7 +1358,7 @@ EXPORT void *h_calloc(size_t nmemb, size_t size) {
total_size = adjust_size_for_canaries(total_size);
void *p = allocate(arena, total_size);
thread_seal_metadata();
if (!ZERO_ON_FREE && likely(p != NULL) && total_size && total_size <= MAX_SLAB_SIZE_CLASS) {
if (!ro.zero_on_free && likely(p != NULL) && total_size && total_size <= MAX_SLAB_SIZE_CLASS) {
memset(p, 0, total_size - canary_size);
}
return p;