| // Copyright 2016 The Chromium Authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #include "base/allocator/allocator_shim.h" |
| |
| #include <stdlib.h> |
| #include <string.h> |
| |
| #include <memory> |
| #include <new> |
| #include <vector> |
| |
| #include "starboard/types.h" |
| |
| #include "starboard/common/string.h" |
| |
| #include "starboard/memory.h" |
| |
| #include "base/allocator/buildflags.h" |
| #include "base/allocator/partition_allocator/partition_alloc.h" |
| #include "base/atomicops.h" |
| #include "base/process/process_metrics.h" |
| #include "base/synchronization/waitable_event.h" |
| #include "base/threading/platform_thread.h" |
| #include "base/threading/thread_local.h" |
| #include "build/build_config.h" |
| #include "testing/gmock/include/gmock/gmock.h" |
| #include "testing/gtest/include/gtest/gtest.h" |
| |
| #if defined(OS_WIN) |
| #include <windows.h> |
| #elif defined(OS_MACOSX) |
| #include <malloc/malloc.h> |
| #include "base/allocator/allocator_interception_mac.h" |
| #include "base/mac/mac_util.h" |
| #include "third_party/apple_apsl/malloc.h" |
| #else |
| #include <malloc.h> |
| #endif |
| |
| #if !defined(OS_WIN) |
| #include <unistd.h> |
| #endif |
| |
| // Some new Android NDKs (64 bit) does not expose (p)valloc anymore. These |
| // functions are implemented at the shim-layer level. |
| #if defined(OS_ANDROID) |
| extern "C" { |
| void* valloc(size_t size); |
| void* pvalloc(size_t size); |
| } |
| #endif |
| |
| namespace base { |
| namespace allocator { |
| namespace { |
| |
| using testing::MockFunction; |
| using testing::_; |
| |
| class AllocatorShimTest : public testing::Test { |
| public: |
| static const size_t kMaxSizeTracked = 2 * base::kSystemPageSize; |
| AllocatorShimTest() : testing::Test() {} |
| |
| static size_t Hash(const void* ptr) { |
| return reinterpret_cast<uintptr_t>(ptr) % kMaxSizeTracked; |
| } |
| |
| static void* MockAlloc(const AllocatorDispatch* self, |
| size_t size, |
| void* context) { |
| if (instance_ && size < kMaxSizeTracked) |
| ++(instance_->allocs_intercepted_by_size[size]); |
| return self->next->alloc_function(self->next, size, context); |
| } |
| |
| static void* MockAllocZeroInit(const AllocatorDispatch* self, |
| size_t n, |
| size_t size, |
| void* context) { |
| const size_t real_size = n * size; |
| if (instance_ && real_size < kMaxSizeTracked) |
| ++(instance_->zero_allocs_intercepted_by_size[real_size]); |
| return self->next->alloc_zero_initialized_function(self->next, n, size, |
| context); |
| } |
| |
| static void* MockAllocAligned(const AllocatorDispatch* self, |
| size_t alignment, |
| size_t size, |
| void* context) { |
| if (instance_) { |
| if (size < kMaxSizeTracked) |
| ++(instance_->aligned_allocs_intercepted_by_size[size]); |
| if (alignment < kMaxSizeTracked) |
| ++(instance_->aligned_allocs_intercepted_by_alignment[alignment]); |
| } |
| return self->next->alloc_aligned_function(self->next, alignment, size, |
| context); |
| } |
| |
| static void* MockRealloc(const AllocatorDispatch* self, |
| void* address, |
| size_t size, |
| void* context) { |
| if (instance_) { |
| // Size 0xFEED a special sentinel for the NewHandlerConcurrency test. |
| // Hitting it for the first time will cause a failure, causing the |
| // invocation of the std::new_handler. |
| if (size == 0xFEED) { |
| if (!instance_->did_fail_realloc_0xfeed_once->Get()) { |
| instance_->did_fail_realloc_0xfeed_once->Set(true); |
| return nullptr; |
| } |
| return address; |
| } |
| |
| if (size < kMaxSizeTracked) |
| ++(instance_->reallocs_intercepted_by_size[size]); |
| ++instance_->reallocs_intercepted_by_addr[Hash(address)]; |
| } |
| return self->next->realloc_function(self->next, address, size, context); |
| } |
| |
| static void MockFree(const AllocatorDispatch* self, |
| void* address, |
| void* context) { |
| if (instance_) { |
| ++instance_->frees_intercepted_by_addr[Hash(address)]; |
| } |
| self->next->free_function(self->next, address, context); |
| } |
| |
| static size_t MockGetSizeEstimate(const AllocatorDispatch* self, |
| void* address, |
| void* context) { |
| return self->next->get_size_estimate_function(self->next, address, context); |
| } |
| |
| static unsigned MockBatchMalloc(const AllocatorDispatch* self, |
| size_t size, |
| void** results, |
| unsigned num_requested, |
| void* context) { |
| if (instance_) { |
| instance_->batch_mallocs_intercepted_by_size[size] = |
| instance_->batch_mallocs_intercepted_by_size[size] + num_requested; |
| } |
| return self->next->batch_malloc_function(self->next, size, results, |
| num_requested, context); |
| } |
| |
| static void MockBatchFree(const AllocatorDispatch* self, |
| void** to_be_freed, |
| unsigned num_to_be_freed, |
| void* context) { |
| if (instance_) { |
| for (unsigned i = 0; i < num_to_be_freed; ++i) { |
| ++instance_->batch_frees_intercepted_by_addr[Hash(to_be_freed[i])]; |
| } |
| } |
| self->next->batch_free_function(self->next, to_be_freed, num_to_be_freed, |
| context); |
| } |
| |
| static void MockFreeDefiniteSize(const AllocatorDispatch* self, |
| void* ptr, |
| size_t size, |
| void* context) { |
| if (instance_) { |
| ++instance_->frees_intercepted_by_addr[Hash(ptr)]; |
| ++instance_->free_definite_sizes_intercepted_by_size[size]; |
| } |
| self->next->free_definite_size_function(self->next, ptr, size, context); |
| } |
| |
| static void NewHandler() { |
| if (!instance_) |
| return; |
| subtle::Barrier_AtomicIncrement(&instance_->num_new_handler_calls, 1); |
| } |
| |
| int32_t GetNumberOfNewHandlerCalls() { |
| return subtle::Acquire_Load(&instance_->num_new_handler_calls); |
| } |
| |
| void SetUp() override { |
| const size_t array_size = kMaxSizeTracked * sizeof(size_t); |
| SbMemorySet(&allocs_intercepted_by_size, 0, array_size); |
| SbMemorySet(&zero_allocs_intercepted_by_size, 0, array_size); |
| SbMemorySet(&aligned_allocs_intercepted_by_size, 0, array_size); |
| SbMemorySet(&aligned_allocs_intercepted_by_alignment, 0, array_size); |
| SbMemorySet(&reallocs_intercepted_by_size, 0, array_size); |
| SbMemorySet(&frees_intercepted_by_addr, 0, array_size); |
| SbMemorySet(&batch_mallocs_intercepted_by_size, 0, array_size); |
| SbMemorySet(&batch_frees_intercepted_by_addr, 0, array_size); |
| SbMemorySet(&free_definite_sizes_intercepted_by_size, 0, array_size); |
| did_fail_realloc_0xfeed_once.reset(new ThreadLocalBoolean()); |
| subtle::Release_Store(&num_new_handler_calls, 0); |
| instance_ = this; |
| |
| #if defined(OS_MACOSX) |
| InitializeAllocatorShim(); |
| #endif |
| } |
| |
| void TearDown() override { |
| instance_ = nullptr; |
| #if defined(OS_MACOSX) |
| UninterceptMallocZonesForTesting(); |
| #endif |
| } |
| |
| protected: |
| size_t allocs_intercepted_by_size[kMaxSizeTracked]; |
| size_t zero_allocs_intercepted_by_size[kMaxSizeTracked]; |
| size_t aligned_allocs_intercepted_by_size[kMaxSizeTracked]; |
| size_t aligned_allocs_intercepted_by_alignment[kMaxSizeTracked]; |
| size_t reallocs_intercepted_by_size[kMaxSizeTracked]; |
| size_t reallocs_intercepted_by_addr[kMaxSizeTracked]; |
| size_t frees_intercepted_by_addr[kMaxSizeTracked]; |
| size_t batch_mallocs_intercepted_by_size[kMaxSizeTracked]; |
| size_t batch_frees_intercepted_by_addr[kMaxSizeTracked]; |
| size_t free_definite_sizes_intercepted_by_size[kMaxSizeTracked]; |
| std::unique_ptr<ThreadLocalBoolean> did_fail_realloc_0xfeed_once; |
| subtle::Atomic32 num_new_handler_calls; |
| |
| private: |
| static AllocatorShimTest* instance_; |
| }; |
| |
| struct TestStruct1 { |
| uint32_t ignored; |
| uint8_t ignored_2; |
| }; |
| |
| struct TestStruct2 { |
| uint64_t ignored; |
| uint8_t ignored_3; |
| }; |
| |
| class ThreadDelegateForNewHandlerTest : public PlatformThread::Delegate { |
| public: |
| ThreadDelegateForNewHandlerTest(WaitableEvent* event) : event_(event) {} |
| |
| void ThreadMain() override { |
| event_->Wait(); |
| void* temp = SbMemoryAllocate(1); |
| void* res = SbMemoryReallocate(temp, 0xFEED); |
| EXPECT_EQ(temp, res); |
| } |
| |
| private: |
| WaitableEvent* event_; |
| }; |
| |
| AllocatorShimTest* AllocatorShimTest::instance_ = nullptr; |
| |
| AllocatorDispatch g_mock_dispatch = { |
| &AllocatorShimTest::MockAlloc, /* alloc_function */ |
| &AllocatorShimTest::MockAllocZeroInit, /* alloc_zero_initialized_function */ |
| &AllocatorShimTest::MockAllocAligned, /* alloc_aligned_function */ |
| &AllocatorShimTest::MockRealloc, /* realloc_function */ |
| &AllocatorShimTest::MockFree, /* free_function */ |
| &AllocatorShimTest::MockGetSizeEstimate, /* get_size_estimate_function */ |
| &AllocatorShimTest::MockBatchMalloc, /* batch_malloc_function */ |
| &AllocatorShimTest::MockBatchFree, /* batch_free_function */ |
| &AllocatorShimTest::MockFreeDefiniteSize, /* free_definite_size_function */ |
| nullptr, /* next */ |
| }; |
| |
| TEST_F(AllocatorShimTest, InterceptLibcSymbols) { |
| InsertAllocatorDispatch(&g_mock_dispatch); |
| |
| void* alloc_ptr = SbMemoryAllocate(19); |
| ASSERT_NE(nullptr, alloc_ptr); |
| ASSERT_GE(allocs_intercepted_by_size[19], 1u); |
| |
| void* zero_alloc_ptr = calloc(2, 23); |
| ASSERT_NE(nullptr, zero_alloc_ptr); |
| ASSERT_GE(zero_allocs_intercepted_by_size[2 * 23], 1u); |
| |
| #if !defined(OS_WIN) |
| const size_t kPageSize = base::GetPageSize(); |
| void* posix_memalign_ptr = nullptr; |
| int res = posix_memalign(&posix_memalign_ptr, 256, 59); |
| ASSERT_EQ(0, res); |
| ASSERT_NE(nullptr, posix_memalign_ptr); |
| ASSERT_EQ(0u, reinterpret_cast<uintptr_t>(posix_memalign_ptr) % 256); |
| ASSERT_GE(aligned_allocs_intercepted_by_alignment[256], 1u); |
| ASSERT_GE(aligned_allocs_intercepted_by_size[59], 1u); |
| |
| void* valloc_ptr = valloc(61); |
| ASSERT_NE(nullptr, valloc_ptr); |
| ASSERT_EQ(0u, reinterpret_cast<uintptr_t>(valloc_ptr) % kPageSize); |
| ASSERT_GE(aligned_allocs_intercepted_by_alignment[kPageSize], 1u); |
| ASSERT_GE(aligned_allocs_intercepted_by_size[61], 1u); |
| #endif // !OS_WIN |
| |
| #if !defined(OS_WIN) && !defined(OS_MACOSX) |
| void* memalign_ptr = memalign(128, 53); |
| ASSERT_NE(nullptr, memalign_ptr); |
| ASSERT_EQ(0u, reinterpret_cast<uintptr_t>(memalign_ptr) % 128); |
| ASSERT_GE(aligned_allocs_intercepted_by_alignment[128], 1u); |
| ASSERT_GE(aligned_allocs_intercepted_by_size[53], 1u); |
| |
| void* pvalloc_ptr = pvalloc(67); |
| ASSERT_NE(nullptr, pvalloc_ptr); |
| ASSERT_EQ(0u, reinterpret_cast<uintptr_t>(pvalloc_ptr) % kPageSize); |
| ASSERT_GE(aligned_allocs_intercepted_by_alignment[kPageSize], 1u); |
| // pvalloc rounds the size up to the next page. |
| ASSERT_GE(aligned_allocs_intercepted_by_size[kPageSize], 1u); |
| #endif // !OS_WIN && !OS_MACOSX |
| |
| char* realloc_ptr = static_cast<char*>(SbMemoryAllocate(10)); |
| SbStringCopyUnsafe(realloc_ptr, "foobar"); |
| void* old_realloc_ptr = realloc_ptr; |
| SbMemoryReallocate_ptr = |
| static_cast<char*>(SbMemoryReallocate(SbMemoryReallocate_ptr, 73)); |
| ASSERT_GE(reallocs_intercepted_by_size[73], 1u); |
| ASSERT_GE(reallocs_intercepted_by_addr[Hash(old_realloc_ptr)], 1u); |
| ASSERT_EQ(0, strcmp(realloc_ptr, "foobar")); |
| |
| SbMemoryDeallocate(alloc_ptr); |
| ASSERT_GE(frees_intercepted_by_addr[Hash(alloc_ptr)], 1u); |
| |
| SbMemoryDeallocate(zero_alloc_ptr); |
| ASSERT_GE(frees_intercepted_by_addr[Hash(zero_alloc_ptr)], 1u); |
| |
| #if !defined(OS_WIN) && !defined(OS_MACOSX) |
| SbMemoryDeallocate(memalign_ptr); |
| ASSERT_GE(frees_intercepted_by_addr[Hash(memalign_ptr)], 1u); |
| |
| SbMemoryDeallocate(pvalloc_ptr); |
| ASSERT_GE(frees_intercepted_by_addr[Hash(pvalloc_ptr)], 1u); |
| #endif // !OS_WIN && !OS_MACOSX |
| |
| #if !defined(OS_WIN) |
| SbMemoryDeallocate(posix_memalign_ptr); |
| ASSERT_GE(frees_intercepted_by_addr[Hash(posix_memalign_ptr)], 1u); |
| |
| SbMemoryDeallocate(valloc_ptr); |
| ASSERT_GE(frees_intercepted_by_addr[Hash(valloc_ptr)], 1u); |
| #endif // !OS_WIN |
| |
| SbMemoryDeallocate(realloc_ptr); |
| ASSERT_GE(frees_intercepted_by_addr[Hash(realloc_ptr)], 1u); |
| |
| RemoveAllocatorDispatchForTesting(&g_mock_dispatch); |
| |
| void* non_hooked_ptr = SbMemoryAllocate(4095); |
| ASSERT_NE(nullptr, non_hooked_ptr); |
| ASSERT_EQ(0u, allocs_intercepted_by_size[4095]); |
| SbMemoryDeallocate(non_hooked_ptr); |
| } |
| |
| #if defined(OS_MACOSX) |
| TEST_F(AllocatorShimTest, InterceptLibcSymbolsBatchMallocFree) { |
| InsertAllocatorDispatch(&g_mock_dispatch); |
| |
| unsigned count = 13; |
| std::vector<void*> results; |
| results.resize(count); |
| unsigned result_count = malloc_zone_batch_malloc(malloc_default_zone(), 99, |
| results.data(), count); |
| ASSERT_EQ(count, result_count); |
| |
| // TODO(erikchen): On macOS 10.12+, batch_malloc in the default zone may |
| // forward to another zone, which we've also shimmed, resulting in |
| // MockBatchMalloc getting called twice as often as we'd expect. This |
| // re-entrancy into the allocator shim is a bug that needs to be fixed. |
| // https://crbug.com/693237. |
| // ASSERT_EQ(count, batch_mallocs_intercepted_by_size[99]); |
| |
| std::vector<void*> results_copy(results); |
| malloc_zone_batch_free(malloc_default_zone(), results.data(), count); |
| for (void* result : results_copy) { |
| ASSERT_GE(batch_frees_intercepted_by_addr[Hash(result)], 1u); |
| } |
| RemoveAllocatorDispatchForTesting(&g_mock_dispatch); |
| } |
| |
| TEST_F(AllocatorShimTest, InterceptLibcSymbolsFreeDefiniteSize) { |
| InsertAllocatorDispatch(&g_mock_dispatch); |
| |
| void* alloc_ptr = SbMemoryAllocate(19); |
| ASSERT_NE(nullptr, alloc_ptr); |
| ASSERT_GE(allocs_intercepted_by_size[19], 1u); |
| |
| ChromeMallocZone* default_zone = |
| reinterpret_cast<ChromeMallocZone*>(malloc_default_zone()); |
| default_zone->free_definite_size(malloc_default_zone(), alloc_ptr, 19); |
| ASSERT_GE(free_definite_sizes_intercepted_by_size[19], 1u); |
| RemoveAllocatorDispatchForTesting(&g_mock_dispatch); |
| } |
| #endif // defined(OS_MACOSX) |
| |
| TEST_F(AllocatorShimTest, InterceptCppSymbols) { |
| InsertAllocatorDispatch(&g_mock_dispatch); |
| |
| TestStruct1* new_ptr = new TestStruct1; |
| ASSERT_NE(nullptr, new_ptr); |
| ASSERT_GE(allocs_intercepted_by_size[sizeof(TestStruct1)], 1u); |
| |
| TestStruct1* new_array_ptr = new TestStruct1[3]; |
| ASSERT_NE(nullptr, new_array_ptr); |
| ASSERT_GE(allocs_intercepted_by_size[sizeof(TestStruct1) * 3], 1u); |
| |
| TestStruct2* new_nt_ptr = new (std::nothrow) TestStruct2; |
| ASSERT_NE(nullptr, new_nt_ptr); |
| ASSERT_GE(allocs_intercepted_by_size[sizeof(TestStruct2)], 1u); |
| |
| TestStruct2* new_array_nt_ptr = new TestStruct2[3]; |
| ASSERT_NE(nullptr, new_array_nt_ptr); |
| ASSERT_GE(allocs_intercepted_by_size[sizeof(TestStruct2) * 3], 1u); |
| |
| delete new_ptr; |
| ASSERT_GE(frees_intercepted_by_addr[Hash(new_ptr)], 1u); |
| |
| delete[] new_array_ptr; |
| ASSERT_GE(frees_intercepted_by_addr[Hash(new_array_ptr)], 1u); |
| |
| delete new_nt_ptr; |
| ASSERT_GE(frees_intercepted_by_addr[Hash(new_nt_ptr)], 1u); |
| |
| delete[] new_array_nt_ptr; |
| ASSERT_GE(frees_intercepted_by_addr[Hash(new_array_nt_ptr)], 1u); |
| |
| RemoveAllocatorDispatchForTesting(&g_mock_dispatch); |
| } |
| |
| // This test exercises the case of concurrent OOM failure, which would end up |
| // invoking std::new_handler concurrently. This is to cover the CallNewHandler() |
| // paths of allocator_shim.cc and smoke-test its thread safey. |
| // The test creates kNumThreads threads. Each of them mallocs some memory, and |
| // then does a realloc(<new memory>, 0xFEED). |
| // The shim intercepts such realloc and makes it fail only once on each thread. |
| // We expect to see excactly kNumThreads invocations of the new_handler. |
| TEST_F(AllocatorShimTest, NewHandlerConcurrency) { |
| const int kNumThreads = 32; |
| PlatformThreadHandle threads[kNumThreads]; |
| |
| // The WaitableEvent here is used to attempt to trigger all the threads at |
| // the same time, after they have been initialized. |
| WaitableEvent event(WaitableEvent::ResetPolicy::MANUAL, |
| WaitableEvent::InitialState::NOT_SIGNALED); |
| |
| ThreadDelegateForNewHandlerTest mock_thread_main(&event); |
| |
| for (int i = 0; i < kNumThreads; ++i) |
| PlatformThread::Create(0, &mock_thread_main, &threads[i]); |
| |
| std::set_new_handler(&AllocatorShimTest::NewHandler); |
| SetCallNewHandlerOnMallocFailure(true); // It's going to fail on realloc(). |
| InsertAllocatorDispatch(&g_mock_dispatch); |
| event.Signal(); |
| for (int i = 0; i < kNumThreads; ++i) |
| PlatformThread::Join(threads[i]); |
| RemoveAllocatorDispatchForTesting(&g_mock_dispatch); |
| ASSERT_EQ(kNumThreads, GetNumberOfNewHandlerCalls()); |
| } |
| |
| #if defined(OS_WIN) && BUILDFLAG(USE_ALLOCATOR_SHIM) |
| TEST_F(AllocatorShimTest, ShimReplacesCRTHeapWhenEnabled) { |
| ASSERT_NE(::GetProcessHeap(), reinterpret_cast<HANDLE>(_get_heap_handle())); |
| } |
| #endif // defined(OS_WIN) && BUILDFLAG(USE_ALLOCATOR_SHIM) |
| |
| } // namespace |
| } // namespace allocator |
| } // namespace base |