289 lines
		
	
	
		
			9.1 KiB
		
	
	
	
		
			C++
		
	
	
	
			
		
		
	
	
			289 lines
		
	
	
		
			9.1 KiB
		
	
	
	
		
			C++
		
	
	
	
| /*
 | |
|  * Copyright 2019 The Android Open Source Project
 | |
|  *
 | |
|  * Licensed under the Apache License, Version 2.0 (the "License");
 | |
|  * you may not use this file except in compliance with the License.
 | |
|  * You may obtain a copy of the License at
 | |
|  *
 | |
|  *      http://www.apache.org/licenses/LICENSE-2.0
 | |
|  *
 | |
|  * Unless required by applicable law or agreed to in writing, software
 | |
|  * distributed under the License is distributed on an "AS IS" BASIS,
 | |
|  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
|  * See the License for the specific language governing permissions and
 | |
|  * limitations under the License.
 | |
|  */
 | |
| 
 | |
| #ifndef ART_RUNTIME_JIT_JIT_MEMORY_REGION_H_
 | |
| #define ART_RUNTIME_JIT_JIT_MEMORY_REGION_H_
 | |
| 
 | |
| #include <string>
 | |
| 
 | |
| #include "arch/instruction_set.h"
 | |
| #include "base/globals.h"
 | |
| #include "base/locks.h"
 | |
| #include "base/mem_map.h"
 | |
| #include "gc_root-inl.h"
 | |
| #include "handle.h"
 | |
| 
 | |
| namespace art {
 | |
| 
 | |
| namespace mirror {
 | |
| class Object;
 | |
| }
 | |
| 
 | |
| namespace jit {
 | |
| 
 | |
| class TestZygoteMemory;
 | |
| 
 | |
| // Number of bytes represented by a bit in the CodeCacheBitmap. Value is reasonable for all
 | |
| // architectures.
 | |
| static constexpr int kJitCodeAccountingBytes = 16;
 | |
| 
 | |
| // Helper to get the size required for emitting `number_of_roots` in the
 | |
| // data portion of a JIT memory region.
 | |
| uint32_t inline ComputeRootTableSize(uint32_t number_of_roots) {
 | |
|   return sizeof(uint32_t) + number_of_roots * sizeof(GcRoot<mirror::Object>);
 | |
| }
 | |
| 
 | |
| // Represents a memory region for the JIT, where code and data are stored. This class
 | |
| // provides allocation and deallocation primitives.
 | |
| class JitMemoryRegion {
 | |
|  public:
 | |
|   JitMemoryRegion()
 | |
|       : initial_capacity_(0),
 | |
|         max_capacity_(0),
 | |
|         current_capacity_(0),
 | |
|         data_end_(0),
 | |
|         exec_end_(0),
 | |
|         used_memory_for_code_(0),
 | |
|         used_memory_for_data_(0),
 | |
|         data_pages_(),
 | |
|         writable_data_pages_(),
 | |
|         exec_pages_(),
 | |
|         non_exec_pages_(),
 | |
|         data_mspace_(nullptr),
 | |
|         exec_mspace_(nullptr) {}
 | |
| 
 | |
|   bool Initialize(size_t initial_capacity,
 | |
|                   size_t max_capacity,
 | |
|                   bool rwx_memory_allowed,
 | |
|                   bool is_zygote,
 | |
|                   std::string* error_msg)
 | |
|       REQUIRES(Locks::jit_lock_);
 | |
| 
 | |
|   // Try to increase the current capacity of the code cache. Return whether we
 | |
|   // succeeded at doing so.
 | |
|   bool IncreaseCodeCacheCapacity() REQUIRES(Locks::jit_lock_);
 | |
| 
 | |
|   // Set the footprint limit of the code cache.
 | |
|   void SetFootprintLimit(size_t new_footprint) REQUIRES(Locks::jit_lock_);
 | |
| 
 | |
|   const uint8_t* AllocateCode(size_t code_size) REQUIRES(Locks::jit_lock_);
 | |
|   void FreeCode(const uint8_t* code) REQUIRES(Locks::jit_lock_);
 | |
|   const uint8_t* AllocateData(size_t data_size) REQUIRES(Locks::jit_lock_);
 | |
|   void FreeData(const uint8_t* data) REQUIRES(Locks::jit_lock_);
 | |
|   void FreeData(uint8_t* writable_data) REQUIRES(Locks::jit_lock_) = delete;
 | |
|   void FreeWritableData(uint8_t* writable_data) REQUIRES(Locks::jit_lock_);
 | |
| 
 | |
|   // Emit header and code into the memory pointed by `reserved_code` (despite it being const).
 | |
|   // Returns pointer to copied code (within reserved_code region; after OatQuickMethodHeader).
 | |
|   const uint8_t* CommitCode(ArrayRef<const uint8_t> reserved_code,
 | |
|                             ArrayRef<const uint8_t> code,
 | |
|                             const uint8_t* stack_map,
 | |
|                             bool has_should_deoptimize_flag)
 | |
|       REQUIRES(Locks::jit_lock_);
 | |
| 
 | |
|   // Emit roots and stack map into the memory pointed by `roots_data` (despite it being const).
 | |
|   bool CommitData(ArrayRef<const uint8_t> reserved_data,
 | |
|                   const std::vector<Handle<mirror::Object>>& roots,
 | |
|                   ArrayRef<const uint8_t> stack_map)
 | |
|       REQUIRES(Locks::jit_lock_)
 | |
|       REQUIRES_SHARED(Locks::mutator_lock_);
 | |
| 
 | |
|   void ResetWritableMappings() REQUIRES(Locks::jit_lock_) {
 | |
|     non_exec_pages_.ResetInForkedProcess();
 | |
|     writable_data_pages_.ResetInForkedProcess();
 | |
|     // Also clear the mspaces, which, in their implementation,
 | |
|     // point to the discarded mappings.
 | |
|     exec_mspace_ = nullptr;
 | |
|     data_mspace_ = nullptr;
 | |
|   }
 | |
| 
 | |
|   bool IsValid() const NO_THREAD_SAFETY_ANALYSIS {
 | |
|     return exec_mspace_ != nullptr || data_mspace_ != nullptr;
 | |
|   }
 | |
| 
 | |
|   template <typename T>
 | |
|   void FillData(const T* address, size_t n, const T& t)  REQUIRES(Locks::jit_lock_) {
 | |
|     std::fill_n(GetWritableDataAddress(address), n, t);
 | |
|   }
 | |
| 
 | |
|   // Generic helper for writing abritrary data in the data portion of the
 | |
|   // region.
 | |
|   template <typename T>
 | |
|   void WriteData(const T* address, const T& value) {
 | |
|     *GetWritableDataAddress(address) = value;
 | |
|   }
 | |
| 
 | |
|   bool HasDualCodeMapping() const {
 | |
|     return non_exec_pages_.IsValid();
 | |
|   }
 | |
| 
 | |
|   bool HasDualDataMapping() const {
 | |
|     return writable_data_pages_.IsValid();
 | |
|   }
 | |
| 
 | |
|   bool HasCodeMapping() const {
 | |
|     return exec_pages_.IsValid();
 | |
|   }
 | |
| 
 | |
|   bool IsInDataSpace(const void* ptr) const {
 | |
|     return data_pages_.HasAddress(ptr);
 | |
|   }
 | |
| 
 | |
|   bool IsInExecSpace(const void* ptr) const {
 | |
|     return exec_pages_.HasAddress(ptr);
 | |
|   }
 | |
| 
 | |
|   const MemMap* GetExecPages() const {
 | |
|     return &exec_pages_;
 | |
|   }
 | |
| 
 | |
|   void* MoreCore(const void* mspace, intptr_t increment);
 | |
| 
 | |
|   bool OwnsSpace(const void* mspace) const NO_THREAD_SAFETY_ANALYSIS {
 | |
|     return mspace == data_mspace_ || mspace == exec_mspace_;
 | |
|   }
 | |
| 
 | |
|   size_t GetCurrentCapacity() const REQUIRES(Locks::jit_lock_) {
 | |
|     return current_capacity_;
 | |
|   }
 | |
| 
 | |
|   size_t GetMaxCapacity() const REQUIRES(Locks::jit_lock_) {
 | |
|     return max_capacity_;
 | |
|   }
 | |
| 
 | |
|   size_t GetUsedMemoryForCode() const REQUIRES(Locks::jit_lock_) {
 | |
|     return used_memory_for_code_;
 | |
|   }
 | |
| 
 | |
|   size_t GetResidentMemoryForCode() const REQUIRES(Locks::jit_lock_) {
 | |
|     return exec_end_;
 | |
|   }
 | |
| 
 | |
|   size_t GetUsedMemoryForData() const REQUIRES(Locks::jit_lock_) {
 | |
|     return used_memory_for_data_;
 | |
|   }
 | |
| 
 | |
|   size_t GetResidentMemoryForData() const REQUIRES(Locks::jit_lock_) {
 | |
|     return data_end_;
 | |
|   }
 | |
| 
 | |
|   template <typename T> T* GetWritableDataAddress(const T* src_ptr) {
 | |
|     if (!HasDualDataMapping()) {
 | |
|       return const_cast<T*>(src_ptr);
 | |
|     }
 | |
|     return const_cast<T*>(TranslateAddress(src_ptr, data_pages_, writable_data_pages_));
 | |
|   }
 | |
| 
 | |
|  private:
 | |
|   template <typename T>
 | |
|   T* TranslateAddress(T* src_ptr, const MemMap& src, const MemMap& dst) {
 | |
|     CHECK(src.HasAddress(src_ptr)) << reinterpret_cast<const void*>(src_ptr);
 | |
|     const uint8_t* const raw_src_ptr = reinterpret_cast<const uint8_t*>(src_ptr);
 | |
|     return reinterpret_cast<T*>(raw_src_ptr - src.Begin() + dst.Begin());
 | |
|   }
 | |
| 
 | |
|   const MemMap* GetUpdatableCodeMapping() const {
 | |
|     if (HasDualCodeMapping()) {
 | |
|       return &non_exec_pages_;
 | |
|     } else if (HasCodeMapping()) {
 | |
|       return &exec_pages_;
 | |
|     } else {
 | |
|       return nullptr;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   const MemMap* GetWritableDataMapping() const {
 | |
|     if (HasDualDataMapping()) {
 | |
|       return &writable_data_pages_;
 | |
|     } else {
 | |
|       return &data_pages_;
 | |
|     }
 | |
|   }
 | |
| 
 | |
|   template <typename T> T* GetNonWritableDataAddress(T* src_ptr) {
 | |
|     if (!HasDualDataMapping()) {
 | |
|       return src_ptr;
 | |
|     }
 | |
|     return TranslateAddress(src_ptr, writable_data_pages_, data_pages_);
 | |
|   }
 | |
| 
 | |
|   template <typename T> T* GetExecutableAddress(T* src_ptr) {
 | |
|     if (!HasDualCodeMapping()) {
 | |
|       return src_ptr;
 | |
|     }
 | |
|     return TranslateAddress(src_ptr, non_exec_pages_, exec_pages_);
 | |
|   }
 | |
| 
 | |
|   template <typename T> T* GetNonExecutableAddress(T* src_ptr) {
 | |
|     if (!HasDualCodeMapping()) {
 | |
|       return src_ptr;
 | |
|     }
 | |
|     return TranslateAddress(src_ptr, exec_pages_, non_exec_pages_);
 | |
|   }
 | |
| 
 | |
|   static int CreateZygoteMemory(size_t capacity, std::string* error_msg);
 | |
|   static bool ProtectZygoteMemory(int fd, std::string* error_msg);
 | |
| 
 | |
|   // The initial capacity in bytes this code region starts with.
 | |
|   size_t initial_capacity_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   // The maximum capacity in bytes this region can go to.
 | |
|   size_t max_capacity_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   // The current capacity in bytes of the region.
 | |
|   size_t current_capacity_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   // The current footprint in bytes of the data portion of the region.
 | |
|   size_t data_end_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   // The current footprint in bytes of the code portion of the region.
 | |
|   size_t exec_end_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   // The size in bytes of used memory for the code portion of the region.
 | |
|   size_t used_memory_for_code_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   // The size in bytes of used memory for the data portion of the region.
 | |
|   size_t used_memory_for_data_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   // Mem map which holds data (stack maps and profiling info).
 | |
|   MemMap data_pages_;
 | |
| 
 | |
|   // Mem map which holds data with writable permission. Only valid for dual view
 | |
|   // JIT when this is the writable view and data_pages_ is the readable view.
 | |
|   MemMap writable_data_pages_;
 | |
| 
 | |
|   // Mem map which holds code and has executable permission.
 | |
|   MemMap exec_pages_;
 | |
| 
 | |
|   // Mem map which holds code with non executable permission. Only valid for dual view JIT when
 | |
|   // this is the non-executable view of code used to write updates.
 | |
|   MemMap non_exec_pages_;
 | |
| 
 | |
|   // The opaque mspace for allocating data.
 | |
|   void* data_mspace_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   // The opaque mspace for allocating code.
 | |
|   void* exec_mspace_ GUARDED_BY(Locks::jit_lock_);
 | |
| 
 | |
|   friend class ScopedCodeCacheWrite;  // For GetUpdatableCodeMapping
 | |
|   friend class TestZygoteMemory;
 | |
| };
 | |
| 
 | |
| }  // namespace jit
 | |
| }  // namespace art
 | |
| 
 | |
| #endif  // ART_RUNTIME_JIT_JIT_MEMORY_REGION_H_
 |