416 lines
		
	
	
		
			15 KiB
		
	
	
	
		
			C++
		
	
	
	
			
		
		
	
	
			416 lines
		
	
	
		
			15 KiB
		
	
	
	
		
			C++
		
	
	
	
| /*
 | |
|  * Copyright (C) 2011 The Android Open Source Project
 | |
|  *
 | |
|  * Licensed under the Apache License, Version 2.0 (the "License");
 | |
|  * you may not use this file except in compliance with the License.
 | |
|  * You may obtain a copy of the License at
 | |
|  *
 | |
|  *      http://www.apache.org/licenses/LICENSE-2.0
 | |
|  *
 | |
|  * Unless required by applicable law or agreed to in writing, software
 | |
|  * distributed under the License is distributed on an "AS IS" BASIS,
 | |
|  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
|  * See the License for the specific language governing permissions and
 | |
|  * limitations under the License.
 | |
|  */
 | |
| 
 | |
| #include "dlmalloc_space-inl.h"
 | |
| 
 | |
| #include "base/logging.h"  // For VLOG.
 | |
| #include "base/time_utils.h"
 | |
| #include "base/utils.h"
 | |
| #include "gc/accounting/card_table.h"
 | |
| #include "gc/accounting/space_bitmap-inl.h"
 | |
| #include "gc/heap.h"
 | |
| #include "jit/jit.h"
 | |
| #include "jit/jit_code_cache.h"
 | |
| #include "memory_tool_malloc_space-inl.h"
 | |
| #include "mirror/class-inl.h"
 | |
| #include "mirror/object-inl.h"
 | |
| #include "runtime.h"
 | |
| #include "scoped_thread_state_change-inl.h"
 | |
| #include "thread.h"
 | |
| #include "thread_list.h"
 | |
| 
 | |
| namespace art {
 | |
| namespace gc {
 | |
| namespace space {
 | |
| 
 | |
| static constexpr bool kPrefetchDuringDlMallocFreeList = true;
 | |
| 
 | |
| DlMallocSpace::DlMallocSpace(MemMap&& mem_map,
 | |
|                              size_t initial_size,
 | |
|                              const std::string& name,
 | |
|                              void* mspace,
 | |
|                              uint8_t* begin,
 | |
|                              uint8_t* end,
 | |
|                              uint8_t* limit,
 | |
|                              size_t growth_limit,
 | |
|                              bool can_move_objects,
 | |
|                              size_t starting_size)
 | |
|     : MallocSpace(name,
 | |
|                   std::move(mem_map),
 | |
|                   begin,
 | |
|                   end,
 | |
|                   limit,
 | |
|                   growth_limit,
 | |
|                   /* create_bitmaps= */ true,
 | |
|                   can_move_objects,
 | |
|                   starting_size, initial_size),
 | |
|       mspace_(mspace) {
 | |
|   CHECK(mspace != nullptr);
 | |
| }
 | |
| 
 | |
| DlMallocSpace* DlMallocSpace::CreateFromMemMap(MemMap&& mem_map,
 | |
|                                                const std::string& name,
 | |
|                                                size_t starting_size,
 | |
|                                                size_t initial_size,
 | |
|                                                size_t growth_limit,
 | |
|                                                size_t capacity,
 | |
|                                                bool can_move_objects) {
 | |
|   DCHECK(mem_map.IsValid());
 | |
|   void* mspace = CreateMspace(mem_map.Begin(), starting_size, initial_size);
 | |
|   if (mspace == nullptr) {
 | |
|     LOG(ERROR) << "Failed to initialize mspace for alloc space (" << name << ")";
 | |
|     return nullptr;
 | |
|   }
 | |
| 
 | |
|   // Protect memory beyond the starting size. morecore will add r/w permissions when necessory
 | |
|   uint8_t* end = mem_map.Begin() + starting_size;
 | |
|   if (capacity - starting_size > 0) {
 | |
|     CheckedCall(mprotect, name.c_str(), end, capacity - starting_size, PROT_NONE);
 | |
|   }
 | |
| 
 | |
|   // Everything is set so record in immutable structure and leave
 | |
|   uint8_t* begin = mem_map.Begin();
 | |
|   if (Runtime::Current()->IsRunningOnMemoryTool()) {
 | |
|     return new MemoryToolMallocSpace<DlMallocSpace, kDefaultMemoryToolRedZoneBytes, true, false>(
 | |
|         std::move(mem_map),
 | |
|         initial_size,
 | |
|         name,
 | |
|         mspace,
 | |
|         begin,
 | |
|         end,
 | |
|         begin + capacity, growth_limit,
 | |
|         can_move_objects,
 | |
|         starting_size);
 | |
|   } else {
 | |
|     return new DlMallocSpace(std::move(mem_map),
 | |
|                              initial_size,
 | |
|                              name,
 | |
|                              mspace,
 | |
|                              begin,
 | |
|                              end,
 | |
|                              begin + capacity,
 | |
|                              growth_limit,
 | |
|                              can_move_objects,
 | |
|                              starting_size);
 | |
|   }
 | |
| }
 | |
| 
 | |
| DlMallocSpace* DlMallocSpace::Create(const std::string& name,
 | |
|                                      size_t initial_size,
 | |
|                                      size_t growth_limit,
 | |
|                                      size_t capacity,
 | |
|                                      bool can_move_objects) {
 | |
|   uint64_t start_time = 0;
 | |
|   if (VLOG_IS_ON(heap) || VLOG_IS_ON(startup)) {
 | |
|     start_time = NanoTime();
 | |
|     LOG(INFO) << "DlMallocSpace::Create entering " << name
 | |
|         << " initial_size=" << PrettySize(initial_size)
 | |
|         << " growth_limit=" << PrettySize(growth_limit)
 | |
|         << " capacity=" << PrettySize(capacity);
 | |
|   }
 | |
| 
 | |
|   // Memory we promise to dlmalloc before it asks for morecore.
 | |
|   // Note: making this value large means that large allocations are unlikely to succeed as dlmalloc
 | |
|   // will ask for this memory from sys_alloc which will fail as the footprint (this value plus the
 | |
|   // size of the large allocation) will be greater than the footprint limit.
 | |
|   size_t starting_size = kPageSize;
 | |
|   MemMap mem_map = CreateMemMap(name, starting_size, &initial_size, &growth_limit, &capacity);
 | |
|   if (!mem_map.IsValid()) {
 | |
|     LOG(ERROR) << "Failed to create mem map for alloc space (" << name << ") of size "
 | |
|                << PrettySize(capacity);
 | |
|     return nullptr;
 | |
|   }
 | |
|   DlMallocSpace* space = CreateFromMemMap(std::move(mem_map),
 | |
|                                           name,
 | |
|                                           starting_size,
 | |
|                                           initial_size,
 | |
|                                           growth_limit,
 | |
|                                           capacity,
 | |
|                                           can_move_objects);
 | |
|   // We start out with only the initial size possibly containing objects.
 | |
|   if (VLOG_IS_ON(heap) || VLOG_IS_ON(startup)) {
 | |
|     LOG(INFO) << "DlMallocSpace::Create exiting (" << PrettyDuration(NanoTime() - start_time)
 | |
|         << " ) " << *space;
 | |
|   }
 | |
|   return space;
 | |
| }
 | |
| 
 | |
| void* DlMallocSpace::CreateMspace(void* begin, size_t morecore_start, size_t initial_size) {
 | |
|   // clear errno to allow PLOG on error
 | |
|   errno = 0;
 | |
|   // create mspace using our backing storage starting at begin and with a footprint of
 | |
|   // morecore_start. Don't use an internal dlmalloc lock (as we already hold heap lock). When
 | |
|   // morecore_start bytes of memory is exhaused morecore will be called.
 | |
|   void* msp = create_mspace_with_base(begin, morecore_start, 0 /*locked*/);
 | |
|   if (msp != nullptr) {
 | |
|     // Do not allow morecore requests to succeed beyond the initial size of the heap
 | |
|     mspace_set_footprint_limit(msp, initial_size);
 | |
|   } else {
 | |
|     PLOG(ERROR) << "create_mspace_with_base failed";
 | |
|   }
 | |
|   return msp;
 | |
| }
 | |
| 
 | |
| mirror::Object* DlMallocSpace::AllocWithGrowth(Thread* self, size_t num_bytes,
 | |
|                                                size_t* bytes_allocated, size_t* usable_size,
 | |
|                                                size_t* bytes_tl_bulk_allocated) {
 | |
|   mirror::Object* result;
 | |
|   {
 | |
|     MutexLock mu(self, lock_);
 | |
|     // Grow as much as possible within the space.
 | |
|     size_t max_allowed = Capacity();
 | |
|     mspace_set_footprint_limit(mspace_, max_allowed);
 | |
|     // Try the allocation.
 | |
|     result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size,
 | |
|                                       bytes_tl_bulk_allocated);
 | |
|     // Shrink back down as small as possible.
 | |
|     size_t footprint = mspace_footprint(mspace_);
 | |
|     mspace_set_footprint_limit(mspace_, footprint);
 | |
|   }
 | |
|   if (result != nullptr) {
 | |
|     // Zero freshly allocated memory, done while not holding the space's lock.
 | |
|     memset(result, 0, num_bytes);
 | |
|     // Check that the result is contained in the space.
 | |
|     CHECK_IMPLIES(kDebugSpaces, Contains(result));
 | |
|   }
 | |
|   return result;
 | |
| }
 | |
| 
 | |
| MallocSpace* DlMallocSpace::CreateInstance(MemMap&& mem_map,
 | |
|                                            const std::string& name,
 | |
|                                            void* allocator,
 | |
|                                            uint8_t* begin,
 | |
|                                            uint8_t* end,
 | |
|                                            uint8_t* limit,
 | |
|                                            size_t growth_limit,
 | |
|                                            bool can_move_objects) {
 | |
|   if (Runtime::Current()->IsRunningOnMemoryTool()) {
 | |
|     return new MemoryToolMallocSpace<DlMallocSpace, kDefaultMemoryToolRedZoneBytes, true, false>(
 | |
|         std::move(mem_map),
 | |
|         initial_size_,
 | |
|         name,
 | |
|         allocator,
 | |
|         begin,
 | |
|         end,
 | |
|         limit,
 | |
|         growth_limit,
 | |
|         can_move_objects,
 | |
|         starting_size_);
 | |
|   } else {
 | |
|     return new DlMallocSpace(std::move(mem_map),
 | |
|                              initial_size_,
 | |
|                              name,
 | |
|                              allocator,
 | |
|                              begin,
 | |
|                              end,
 | |
|                              limit,
 | |
|                              growth_limit,
 | |
|                              can_move_objects,
 | |
|                              starting_size_);
 | |
|   }
 | |
| }
 | |
| 
 | |
| size_t DlMallocSpace::Free(Thread* self, mirror::Object* ptr) {
 | |
|   MutexLock mu(self, lock_);
 | |
|   if (kDebugSpaces) {
 | |
|     CHECK(ptr != nullptr);
 | |
|     CHECK(Contains(ptr)) << "Free (" << ptr << ") not in bounds of heap " << *this;
 | |
|   }
 | |
|   const size_t bytes_freed = AllocationSizeNonvirtual(ptr, nullptr);
 | |
|   if (kRecentFreeCount > 0) {
 | |
|     RegisterRecentFree(ptr);
 | |
|   }
 | |
|   mspace_free(mspace_, ptr);
 | |
|   return bytes_freed;
 | |
| }
 | |
| 
 | |
| size_t DlMallocSpace::FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) {
 | |
|   DCHECK(ptrs != nullptr);
 | |
| 
 | |
|   // Don't need the lock to calculate the size of the freed pointers.
 | |
|   size_t bytes_freed = 0;
 | |
|   for (size_t i = 0; i < num_ptrs; i++) {
 | |
|     mirror::Object* ptr = ptrs[i];
 | |
|     const size_t look_ahead = 8;
 | |
|     if (kPrefetchDuringDlMallocFreeList && i + look_ahead < num_ptrs) {
 | |
|       // The head of chunk for the allocation is sizeof(size_t) behind the allocation.
 | |
|       __builtin_prefetch(reinterpret_cast<char*>(ptrs[i + look_ahead]) - sizeof(size_t));
 | |
|     }
 | |
|     bytes_freed += AllocationSizeNonvirtual(ptr, nullptr);
 | |
|   }
 | |
| 
 | |
|   if (kRecentFreeCount > 0) {
 | |
|     MutexLock mu(self, lock_);
 | |
|     for (size_t i = 0; i < num_ptrs; i++) {
 | |
|       RegisterRecentFree(ptrs[i]);
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   if (kDebugSpaces) {
 | |
|     size_t num_broken_ptrs = 0;
 | |
|     for (size_t i = 0; i < num_ptrs; i++) {
 | |
|       if (!Contains(ptrs[i])) {
 | |
|         num_broken_ptrs++;
 | |
|         LOG(ERROR) << "FreeList[" << i << "] (" << ptrs[i] << ") not in bounds of heap " << *this;
 | |
|       } else {
 | |
|         size_t size = mspace_usable_size(ptrs[i]);
 | |
|         memset(ptrs[i], 0xEF, size);
 | |
|       }
 | |
|     }
 | |
|     CHECK_EQ(num_broken_ptrs, 0u);
 | |
|   }
 | |
| 
 | |
|   {
 | |
|     MutexLock mu(self, lock_);
 | |
|     mspace_bulk_free(mspace_, reinterpret_cast<void**>(ptrs), num_ptrs);
 | |
|     return bytes_freed;
 | |
|   }
 | |
| }
 | |
| 
 | |
| size_t DlMallocSpace::Trim() {
 | |
|   MutexLock mu(Thread::Current(), lock_);
 | |
|   // Trim to release memory at the end of the space.
 | |
|   mspace_trim(mspace_, 0);
 | |
|   // Visit space looking for page-sized holes to advise the kernel we don't need.
 | |
|   size_t reclaimed = 0;
 | |
|   mspace_inspect_all(mspace_, DlmallocMadviseCallback, &reclaimed);
 | |
|   return reclaimed;
 | |
| }
 | |
| 
 | |
| void DlMallocSpace::Walk(void(*callback)(void *start, void *end, size_t num_bytes, void* callback_arg),
 | |
|                       void* arg) {
 | |
|   MutexLock mu(Thread::Current(), lock_);
 | |
|   mspace_inspect_all(mspace_, callback, arg);
 | |
|   callback(nullptr, nullptr, 0, arg);  // Indicate end of a space.
 | |
| }
 | |
| 
 | |
| size_t DlMallocSpace::GetFootprint() {
 | |
|   MutexLock mu(Thread::Current(), lock_);
 | |
|   return mspace_footprint(mspace_);
 | |
| }
 | |
| 
 | |
| size_t DlMallocSpace::GetFootprintLimit() {
 | |
|   MutexLock mu(Thread::Current(), lock_);
 | |
|   return mspace_footprint_limit(mspace_);
 | |
| }
 | |
| 
 | |
| void DlMallocSpace::SetFootprintLimit(size_t new_size) {
 | |
|   MutexLock mu(Thread::Current(), lock_);
 | |
|   VLOG(heap) << "DlMallocSpace::SetFootprintLimit " << PrettySize(new_size);
 | |
|   // Compare against the actual footprint, rather than the Size(), because the heap may not have
 | |
|   // grown all the way to the allowed size yet.
 | |
|   size_t current_space_size = mspace_footprint(mspace_);
 | |
|   if (new_size < current_space_size) {
 | |
|     // Don't let the space grow any more.
 | |
|     new_size = current_space_size;
 | |
|   }
 | |
|   mspace_set_footprint_limit(mspace_, new_size);
 | |
| }
 | |
| 
 | |
| uint64_t DlMallocSpace::GetBytesAllocated() {
 | |
|   MutexLock mu(Thread::Current(), lock_);
 | |
|   size_t bytes_allocated = 0;
 | |
|   mspace_inspect_all(mspace_, DlmallocBytesAllocatedCallback, &bytes_allocated);
 | |
|   return bytes_allocated;
 | |
| }
 | |
| 
 | |
| uint64_t DlMallocSpace::GetObjectsAllocated() {
 | |
|   MutexLock mu(Thread::Current(), lock_);
 | |
|   size_t objects_allocated = 0;
 | |
|   mspace_inspect_all(mspace_, DlmallocObjectsAllocatedCallback, &objects_allocated);
 | |
|   return objects_allocated;
 | |
| }
 | |
| 
 | |
| void DlMallocSpace::Clear() {
 | |
|   size_t footprint_limit = GetFootprintLimit();
 | |
|   madvise(GetMemMap()->Begin(), GetMemMap()->Size(), MADV_DONTNEED);
 | |
|   live_bitmap_.Clear();
 | |
|   mark_bitmap_.Clear();
 | |
|   SetEnd(Begin() + starting_size_);
 | |
|   mspace_ = CreateMspace(mem_map_.Begin(), starting_size_, initial_size_);
 | |
|   SetFootprintLimit(footprint_limit);
 | |
| }
 | |
| 
 | |
| #ifndef NDEBUG
 | |
| void DlMallocSpace::CheckMoreCoreForPrecondition() {
 | |
|   lock_.AssertHeld(Thread::Current());
 | |
| }
 | |
| #endif
 | |
| 
 | |
| static void MSpaceChunkCallback(void* start, void* end, size_t used_bytes, void* arg) {
 | |
|   size_t chunk_size = reinterpret_cast<uint8_t*>(end) - reinterpret_cast<uint8_t*>(start);
 | |
|   if (used_bytes < chunk_size) {
 | |
|     size_t chunk_free_bytes = chunk_size - used_bytes;
 | |
|     size_t& max_contiguous_allocation = *reinterpret_cast<size_t*>(arg);
 | |
|     max_contiguous_allocation = std::max(max_contiguous_allocation, chunk_free_bytes);
 | |
|   }
 | |
| }
 | |
| 
 | |
| bool DlMallocSpace::LogFragmentationAllocFailure(std::ostream& os,
 | |
|                                                  size_t failed_alloc_bytes) {
 | |
|   Thread* const self = Thread::Current();
 | |
|   size_t max_contiguous_allocation = 0;
 | |
|   // To allow the Walk/InspectAll() to exclusively-lock the mutator
 | |
|   // lock, temporarily release the shared access to the mutator
 | |
|   // lock here by transitioning to the suspended state.
 | |
|   Locks::mutator_lock_->AssertSharedHeld(self);
 | |
|   ScopedThreadSuspension sts(self, ThreadState::kSuspended);
 | |
|   Walk(MSpaceChunkCallback, &max_contiguous_allocation);
 | |
|   if (failed_alloc_bytes > max_contiguous_allocation) {
 | |
|     os << "; failed due to fragmentation (largest possible contiguous allocation "
 | |
|        <<  max_contiguous_allocation << " bytes)";
 | |
|     return true;
 | |
|   }
 | |
|   return false;
 | |
| }
 | |
| 
 | |
| }  // namespace space
 | |
| 
 | |
| namespace allocator {
 | |
| 
 | |
| // Implement the dlmalloc morecore callback.
 | |
| void* ArtDlMallocMoreCore(void* mspace, intptr_t increment) REQUIRES_SHARED(Locks::mutator_lock_) {
 | |
|   Runtime* runtime = Runtime::Current();
 | |
|   Heap* heap = runtime->GetHeap();
 | |
|   ::art::gc::space::DlMallocSpace* dlmalloc_space = heap->GetDlMallocSpace();
 | |
|   // Support for multiple DlMalloc provided by a slow path.
 | |
|   if (UNLIKELY(dlmalloc_space == nullptr || dlmalloc_space->GetMspace() != mspace)) {
 | |
|     if (LIKELY(runtime->GetJitCodeCache() != nullptr)) {
 | |
|       jit::JitCodeCache* code_cache = runtime->GetJitCodeCache();
 | |
|       if (code_cache->OwnsSpace(mspace)) {
 | |
|         return code_cache->MoreCore(mspace, increment);
 | |
|       }
 | |
|     }
 | |
|     dlmalloc_space = nullptr;
 | |
|     for (space::ContinuousSpace* space : heap->GetContinuousSpaces()) {
 | |
|       if (space->IsDlMallocSpace()) {
 | |
|         ::art::gc::space::DlMallocSpace* cur_dlmalloc_space = space->AsDlMallocSpace();
 | |
|         if (cur_dlmalloc_space->GetMspace() == mspace) {
 | |
|           dlmalloc_space = cur_dlmalloc_space;
 | |
|           break;
 | |
|         }
 | |
|       }
 | |
|     }
 | |
|     CHECK(dlmalloc_space != nullptr) << "Couldn't find DlmMallocSpace with mspace=" << mspace;
 | |
|   }
 | |
|   return dlmalloc_space->MoreCore(increment);
 | |
| }
 | |
| 
 | |
| }  // namespace allocator
 | |
| 
 | |
| }  // namespace gc
 | |
| }  // namespace art
 |