Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Copyright (C) 2008 The Android Open Source Project |
| 3 | * |
| 4 | * Licensed under the Apache License, Version 2.0 (the "License"); |
| 5 | * you may not use this file except in compliance with the License. |
| 6 | * You may obtain a copy of the License at |
| 7 | * |
| 8 | * http://www.apache.org/licenses/LICENSE-2.0 |
| 9 | * |
| 10 | * Unless required by applicable law or agreed to in writing, software |
| 11 | * distributed under the License is distributed on an "AS IS" BASIS, |
| 12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| 13 | * See the License for the specific language governing permissions and |
| 14 | * limitations under the License. |
| 15 | */ |
| 16 | |
David Sehr | 79e2607 | 2018-04-06 17:58:50 -0700 | [diff] [blame] | 17 | #ifndef ART_LIBARTBASE_BASE_MEM_MAP_H_ |
| 18 | #define ART_LIBARTBASE_BASE_MEM_MAP_H_ |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 19 | |
Brian Carlstrom | 27ec961 | 2011-09-19 20:20:38 -0700 | [diff] [blame] | 20 | #include <stddef.h> |
| 21 | #include <sys/types.h> |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 22 | |
Andreas Gampe | 0dfc315 | 2017-04-24 07:58:06 -0700 | [diff] [blame] | 23 | #include <map> |
Igor Murashkin | 5573c37 | 2017-11-16 13:34:30 -0800 | [diff] [blame] | 24 | #include <mutex> |
Andreas Gampe | 0dfc315 | 2017-04-24 07:58:06 -0700 | [diff] [blame] | 25 | #include <string> |
| 26 | |
| 27 | #include "android-base/thread_annotations.h" |
David Sehr | 1979c64 | 2018-04-26 14:41:18 -0700 | [diff] [blame] | 28 | #include "macros.h" |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 29 | |
| 30 | namespace art { |
| 31 | |
Steve Austin | 882ed6b | 2018-06-08 11:40:38 -0700 | [diff] [blame] | 32 | #if defined(__LP64__) && !defined(__Fuchsia__) && \ |
| 33 | (defined(__aarch64__) || defined(__mips__) || defined(__APPLE__)) |
Ian Rogers | c3ccc10 | 2014-06-25 11:52:14 -0700 | [diff] [blame] | 34 | #define USE_ART_LOW_4G_ALLOCATOR 1 |
| 35 | #else |
Steve Austin | 882ed6b | 2018-06-08 11:40:38 -0700 | [diff] [blame] | 36 | #if defined(__LP64__) && !defined(__Fuchsia__) && !defined(__x86_64__) |
Andreas Gampe | 651ba59 | 2017-06-14 14:41:33 -0700 | [diff] [blame] | 37 | #error "Unrecognized 64-bit architecture." |
| 38 | #endif |
Ian Rogers | c3ccc10 | 2014-06-25 11:52:14 -0700 | [diff] [blame] | 39 | #define USE_ART_LOW_4G_ALLOCATOR 0 |
| 40 | #endif |
| 41 | |
Ian Rogers | c5f1773 | 2014-06-05 20:48:42 -0700 | [diff] [blame] | 42 | #ifdef __linux__ |
| 43 | static constexpr bool kMadviseZeroes = true; |
Alex Light | ca97ada | 2018-02-02 09:25:31 -0800 | [diff] [blame] | 44 | #define HAVE_MREMAP_SYSCALL true |
Ian Rogers | c5f1773 | 2014-06-05 20:48:42 -0700 | [diff] [blame] | 45 | #else |
| 46 | static constexpr bool kMadviseZeroes = false; |
Alex Light | ca97ada | 2018-02-02 09:25:31 -0800 | [diff] [blame] | 47 | // We cannot ever perform MemMap::ReplaceWith on non-linux hosts since the syscall is not |
| 48 | // present. |
| 49 | #define HAVE_MREMAP_SYSCALL false |
Ian Rogers | c5f1773 | 2014-06-05 20:48:42 -0700 | [diff] [blame] | 50 | #endif |
| 51 | |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 52 | // Used to keep track of mmap segments. |
Andreas Gampe | d8f26db | 2014-05-19 17:01:13 -0700 | [diff] [blame] | 53 | // |
| 54 | // On 64b systems not supporting MAP_32BIT, the implementation of MemMap will do a linear scan |
| 55 | // for free pages. For security, the start of this scan should be randomized. This requires a |
| 56 | // dynamic initializer. |
| 57 | // For this to work, it is paramount that there are no other static initializers that access MemMap. |
| 58 | // Otherwise, calls might see uninitialized values. |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 59 | class MemMap { |
| 60 | public: |
Alex Light | ca97ada | 2018-02-02 09:25:31 -0800 | [diff] [blame] | 61 | static constexpr bool kCanReplaceMapping = HAVE_MREMAP_SYSCALL; |
| 62 | |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 63 | // Creates an invalid mapping. |
| 64 | MemMap() {} |
| 65 | |
| 66 | // Creates an invalid mapping. Used when we want to be more explicit than MemMap(). |
| 67 | static MemMap Invalid() { |
| 68 | return MemMap(); |
| 69 | } |
| 70 | |
| 71 | MemMap(MemMap&& other) REQUIRES(!MemMap::mem_maps_lock_); |
| 72 | MemMap& operator=(MemMap&& other) REQUIRES(!MemMap::mem_maps_lock_) { |
| 73 | Reset(); |
| 74 | swap(other); |
| 75 | return *this; |
| 76 | } |
| 77 | |
| 78 | // Releases the memory mapping. |
| 79 | ~MemMap() REQUIRES(!MemMap::mem_maps_lock_); |
| 80 | |
| 81 | // Swap two MemMaps. |
| 82 | void swap(MemMap& other); |
| 83 | |
| 84 | void Reset() { |
| 85 | if (IsValid()) { |
| 86 | DoReset(); |
| 87 | } |
| 88 | } |
| 89 | |
| 90 | bool IsValid() const { |
| 91 | return base_size_ != 0u; |
| 92 | } |
| 93 | |
Alex Light | ca97ada | 2018-02-02 09:25:31 -0800 | [diff] [blame] | 94 | // Replace the data in this memmmap with the data in the memmap pointed to by source. The caller |
| 95 | // relinquishes ownership of the source mmap. |
| 96 | // |
| 97 | // For the call to be successful: |
| 98 | // * The range [dest->Begin, dest->Begin() + source->Size()] must not overlap with |
| 99 | // [source->Begin(), source->End()]. |
| 100 | // * Neither source nor dest may be 'reused' mappings (they must own all the pages associated |
| 101 | // with them. |
| 102 | // * kCanReplaceMapping must be true. |
| 103 | // * Neither source nor dest may use manual redzones. |
| 104 | // * Both source and dest must have the same offset from the nearest page boundary. |
| 105 | // * mremap must succeed when called on the mappings. |
| 106 | // |
| 107 | // If this call succeeds it will return true and: |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 108 | // * Invalidate *source |
Alex Light | ca97ada | 2018-02-02 09:25:31 -0800 | [diff] [blame] | 109 | // * The protection of this will remain the same. |
| 110 | // * The size of this will be the size of the source |
| 111 | // * The data in this will be the data from source. |
| 112 | // |
| 113 | // If this call fails it will return false and make no changes to *source or this. The ownership |
| 114 | // of the source mmap is returned to the caller. |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 115 | bool ReplaceWith(/*in-out*/MemMap* source, /*out*/std::string* error); |
Alex Light | ca97ada | 2018-02-02 09:25:31 -0800 | [diff] [blame] | 116 | |
Elliott Hughes | ecd3a6f | 2012-06-06 18:16:37 -0700 | [diff] [blame] | 117 | // Request an anonymous region of length 'byte_count' and a requested base address. |
Mathieu Chartier | 2cebb24 | 2015-04-21 16:50:40 -0700 | [diff] [blame] | 118 | // Use null as the requested base address if you don't care. |
Vladimir Marko | 5c42c29 | 2015-02-25 12:02:49 +0000 | [diff] [blame] | 119 | // "reuse" allows re-mapping an address range from an existing mapping. |
Elliott Hughes | 6c9c06d | 2011-11-07 16:43:47 -0800 | [diff] [blame] | 120 | // |
| 121 | // The word "anonymous" in this context means "not backed by a file". The supplied |
Nicolas Geoffray | a25dce9 | 2016-01-12 16:41:10 +0000 | [diff] [blame] | 122 | // 'name' will be used -- on systems that support it -- to give the mapping |
Elliott Hughes | 6c9c06d | 2011-11-07 16:43:47 -0800 | [diff] [blame] | 123 | // a name. |
Brian Carlstrom | 4a289ed | 2011-08-16 17:17:49 -0700 | [diff] [blame] | 124 | // |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 125 | // On success, returns returns a valid MemMap. On failure, returns an invalid MemMap. |
| 126 | static MemMap MapAnonymous(const char* name, |
| 127 | uint8_t* addr, |
| 128 | size_t byte_count, |
| 129 | int prot, |
| 130 | bool low_4gb, |
| 131 | bool reuse, |
| 132 | std::string* error_msg, |
| 133 | bool use_ashmem = true); |
Vladimir Marko | f6985bd | 2018-08-24 09:02:28 +0100 | [diff] [blame^] | 134 | static MemMap MapAnonymous(const char* name, |
| 135 | uint8_t* addr, |
| 136 | size_t byte_count, |
| 137 | int prot, |
| 138 | bool low_4gb, |
| 139 | std::string* error_msg) { |
| 140 | return MapAnonymous(name, addr, byte_count, prot, low_4gb, /* reuse */ false, error_msg); |
| 141 | } |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 142 | |
David Srbecky | 1baabf0 | 2015-06-16 17:12:34 +0000 | [diff] [blame] | 143 | // Create placeholder for a region allocated by direct call to mmap. |
| 144 | // This is useful when we do not have control over the code calling mmap, |
| 145 | // but when we still want to keep track of it in the list. |
| 146 | // The region is not considered to be owned and will not be unmmaped. |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 147 | static MemMap MapDummy(const char* name, uint8_t* addr, size_t byte_count); |
David Srbecky | 1baabf0 | 2015-06-16 17:12:34 +0000 | [diff] [blame] | 148 | |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 149 | // Map part of a file, taking care of non-page aligned offsets. The |
| 150 | // "start" offset is absolute, not relative. |
| 151 | // |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 152 | // On success, returns returns a valid MemMap. On failure, returns an invalid MemMap. |
| 153 | static MemMap MapFile(size_t byte_count, |
| 154 | int prot, |
| 155 | int flags, |
| 156 | int fd, |
| 157 | off_t start, |
| 158 | bool low_4gb, |
| 159 | const char* filename, |
| 160 | std::string* error_msg) { |
Mathieu Chartier | 42bddce | 2015-11-09 15:16:56 -0800 | [diff] [blame] | 161 | return MapFileAtAddress(nullptr, |
| 162 | byte_count, |
| 163 | prot, |
| 164 | flags, |
| 165 | fd, |
| 166 | start, |
| 167 | /*low_4gb*/low_4gb, |
| 168 | /*reuse*/false, |
| 169 | filename, |
| 170 | error_msg); |
Brian Carlstrom | 4a289ed | 2011-08-16 17:17:49 -0700 | [diff] [blame] | 171 | } |
| 172 | |
Mathieu Chartier | ebe2dfc | 2015-11-24 13:47:52 -0800 | [diff] [blame] | 173 | // Map part of a file, taking care of non-page aligned offsets. The "start" offset is absolute, |
| 174 | // not relative. This version allows requesting a specific address for the base of the mapping. |
| 175 | // "reuse" allows us to create a view into an existing mapping where we do not take ownership of |
| 176 | // the memory. If error_msg is null then we do not print /proc/maps to the log if |
| 177 | // MapFileAtAddress fails. This helps improve performance of the fail case since reading and |
| 178 | // printing /proc/maps takes several milliseconds in the worst case. |
Brian Carlstrom | 4a289ed | 2011-08-16 17:17:49 -0700 | [diff] [blame] | 179 | // |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 180 | // On success, returns returns a valid MemMap. On failure, returns an invalid MemMap. |
| 181 | static MemMap MapFileAtAddress(uint8_t* addr, |
| 182 | size_t byte_count, |
| 183 | int prot, |
| 184 | int flags, |
| 185 | int fd, |
| 186 | off_t start, |
| 187 | bool low_4gb, |
| 188 | bool reuse, |
| 189 | const char* filename, |
| 190 | std::string* error_msg); |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 191 | |
Brian Carlstrom | 0d6adac | 2014-02-05 17:39:16 -0800 | [diff] [blame] | 192 | const std::string& GetName() const { |
| 193 | return name_; |
| 194 | } |
| 195 | |
Vladimir Marko | 9bdf108 | 2016-01-21 12:15:52 +0000 | [diff] [blame] | 196 | bool Sync(); |
| 197 | |
Logan Chien | d88fa26 | 2012-06-06 15:23:32 +0800 | [diff] [blame] | 198 | bool Protect(int prot); |
| 199 | |
Ian Rogers | c5f1773 | 2014-06-05 20:48:42 -0700 | [diff] [blame] | 200 | void MadviseDontNeedAndZero(); |
| 201 | |
Ian Rogers | 1c849e5 | 2012-06-28 14:00:33 -0700 | [diff] [blame] | 202 | int GetProtect() const { |
| 203 | return prot_; |
| 204 | } |
| 205 | |
Ian Rogers | 1373595 | 2014-10-08 12:43:28 -0700 | [diff] [blame] | 206 | uint8_t* Begin() const { |
Ian Rogers | 30fab40 | 2012-01-23 15:43:46 -0800 | [diff] [blame] | 207 | return begin_; |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 208 | } |
| 209 | |
Ian Rogers | 30fab40 | 2012-01-23 15:43:46 -0800 | [diff] [blame] | 210 | size_t Size() const { |
| 211 | return size_; |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 212 | } |
| 213 | |
Mathieu Chartier | 379d09f | 2015-01-08 11:28:13 -0800 | [diff] [blame] | 214 | // Resize the mem-map by unmapping pages at the end. Currently only supports shrinking. |
| 215 | void SetSize(size_t new_size); |
| 216 | |
Ian Rogers | 1373595 | 2014-10-08 12:43:28 -0700 | [diff] [blame] | 217 | uint8_t* End() const { |
Mathieu Chartier | 2fde533 | 2012-09-14 14:51:54 -0700 | [diff] [blame] | 218 | return Begin() + Size(); |
| 219 | } |
| 220 | |
Brian Carlstrom | 0d6adac | 2014-02-05 17:39:16 -0800 | [diff] [blame] | 221 | void* BaseBegin() const { |
| 222 | return base_begin_; |
| 223 | } |
| 224 | |
| 225 | size_t BaseSize() const { |
| 226 | return base_size_; |
| 227 | } |
| 228 | |
| 229 | void* BaseEnd() const { |
Ian Rogers | 1373595 | 2014-10-08 12:43:28 -0700 | [diff] [blame] | 230 | return reinterpret_cast<uint8_t*>(BaseBegin()) + BaseSize(); |
Brian Carlstrom | 0d6adac | 2014-02-05 17:39:16 -0800 | [diff] [blame] | 231 | } |
| 232 | |
Mathieu Chartier | 2fde533 | 2012-09-14 14:51:54 -0700 | [diff] [blame] | 233 | bool HasAddress(const void* addr) const { |
| 234 | return Begin() <= addr && addr < End(); |
Brian Carlstrom | b765be0 | 2011-08-17 23:54:10 -0700 | [diff] [blame] | 235 | } |
| 236 | |
Hiroshi Yamauchi | fd7e7f1 | 2013-10-22 14:17:48 -0700 | [diff] [blame] | 237 | // Unmap the pages at end and remap them to create another memory map. |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 238 | MemMap RemapAtEnd(uint8_t* new_end, |
| 239 | const char* tail_name, |
| 240 | int tail_prot, |
| 241 | std::string* error_msg, |
| 242 | bool use_ashmem = true); |
Mathieu Chartier | cc236d7 | 2012-07-20 10:29:05 -0700 | [diff] [blame] | 243 | |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 244 | static bool CheckNoGaps(MemMap& begin_map, MemMap& end_map) |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 245 | REQUIRES(!MemMap::mem_maps_lock_); |
Vladimir Marko | 17a924a | 2015-05-08 15:17:32 +0100 | [diff] [blame] | 246 | static void DumpMaps(std::ostream& os, bool terse = false) |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 247 | REQUIRES(!MemMap::mem_maps_lock_); |
Hiroshi Yamauchi | 3eed93d | 2014-06-04 11:43:59 -0700 | [diff] [blame] | 248 | |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 249 | // Init and Shutdown are NOT thread safe. |
| 250 | // Both may be called multiple times and MemMap objects may be created any |
| 251 | // time after the first call to Init and before the first call to Shutodwn. |
| 252 | static void Init() REQUIRES(!MemMap::mem_maps_lock_); |
| 253 | static void Shutdown() REQUIRES(!MemMap::mem_maps_lock_); |
Mathieu Chartier | 6e88ef6 | 2014-10-14 15:01:24 -0700 | [diff] [blame] | 254 | |
Hiroshi Yamauchi | 6edb9ae | 2016-02-08 14:18:21 -0800 | [diff] [blame] | 255 | // If the map is PROT_READ, try to read each page of the map to check it is in fact readable (not |
| 256 | // faulting). This is used to diagnose a bug b/19894268 where mprotect doesn't seem to be working |
| 257 | // intermittently. |
| 258 | void TryReadable(); |
| 259 | |
Hiroshi Yamauchi | 3c3c4a1 | 2017-02-21 16:49:59 -0800 | [diff] [blame] | 260 | // Align the map by unmapping the unaligned parts at the lower and the higher ends. |
| 261 | void AlignBy(size_t size); |
| 262 | |
Andreas Gampe | 0dfc315 | 2017-04-24 07:58:06 -0700 | [diff] [blame] | 263 | // For annotation reasons. |
| 264 | static std::mutex* GetMemMapsLock() RETURN_CAPABILITY(mem_maps_lock_) { |
| 265 | return nullptr; |
| 266 | } |
| 267 | |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 268 | private: |
Mathieu Chartier | 42bddce | 2015-11-09 15:16:56 -0800 | [diff] [blame] | 269 | MemMap(const std::string& name, |
| 270 | uint8_t* begin, |
| 271 | size_t size, |
| 272 | void* base_begin, |
| 273 | size_t base_size, |
| 274 | int prot, |
| 275 | bool reuse, |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 276 | size_t redzone_size = 0) REQUIRES(!MemMap::mem_maps_lock_); |
Hiroshi Yamauchi | 3eed93d | 2014-06-04 11:43:59 -0700 | [diff] [blame] | 277 | |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 278 | void DoReset(); |
| 279 | void Invalidate(); |
| 280 | void SwapMembers(MemMap& other); |
| 281 | |
Vladimir Marko | 17a924a | 2015-05-08 15:17:32 +0100 | [diff] [blame] | 282 | static void DumpMapsLocked(std::ostream& os, bool terse) |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 283 | REQUIRES(MemMap::mem_maps_lock_); |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 284 | static bool HasMemMap(MemMap& map) |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 285 | REQUIRES(MemMap::mem_maps_lock_); |
Hiroshi Yamauchi | 3eed93d | 2014-06-04 11:43:59 -0700 | [diff] [blame] | 286 | static MemMap* GetLargestMemMapAt(void* address) |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 287 | REQUIRES(MemMap::mem_maps_lock_); |
Mathieu Chartier | e58991b | 2015-10-13 07:59:34 -0700 | [diff] [blame] | 288 | static bool ContainedWithinExistingMap(uint8_t* ptr, size_t size, std::string* error_msg) |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 289 | REQUIRES(!MemMap::mem_maps_lock_); |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 290 | |
Mathieu Chartier | 42bddce | 2015-11-09 15:16:56 -0800 | [diff] [blame] | 291 | // Internal version of mmap that supports low 4gb emulation. |
| 292 | static void* MapInternal(void* addr, |
| 293 | size_t length, |
| 294 | int prot, |
| 295 | int flags, |
| 296 | int fd, |
| 297 | off_t offset, |
Andreas Gampe | 651ba59 | 2017-06-14 14:41:33 -0700 | [diff] [blame] | 298 | bool low_4gb) |
| 299 | REQUIRES(!MemMap::mem_maps_lock_); |
| 300 | static void* MapInternalArtLow4GBAllocator(size_t length, |
| 301 | int prot, |
| 302 | int flags, |
| 303 | int fd, |
| 304 | off_t offset) |
| 305 | REQUIRES(!MemMap::mem_maps_lock_); |
Mathieu Chartier | 42bddce | 2015-11-09 15:16:56 -0800 | [diff] [blame] | 306 | |
Steve Austin | 882ed6b | 2018-06-08 11:40:38 -0700 | [diff] [blame] | 307 | // member function to access real_munmap |
| 308 | static bool CheckMapRequest(uint8_t* expected_ptr, |
| 309 | void* actual_ptr, |
| 310 | size_t byte_count, |
| 311 | std::string* error_msg); |
| 312 | |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 313 | std::string name_; |
| 314 | uint8_t* begin_ = nullptr; // Start of data. May be changed by AlignBy. |
| 315 | size_t size_ = 0u; // Length of data. |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 316 | |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 317 | void* base_begin_ = nullptr; // Page-aligned base address. May be changed by AlignBy. |
| 318 | size_t base_size_ = 0u; // Length of mapping. May be changed by RemapAtEnd (ie Zygote). |
| 319 | int prot_ = 0; // Protection of the map. |
Hiroshi Yamauchi | fd7e7f1 | 2013-10-22 14:17:48 -0700 | [diff] [blame] | 320 | |
Jim_Guo | a62a588 | 2014-04-28 11:11:57 +0800 | [diff] [blame] | 321 | // When reuse_ is true, this is just a view of an existing mapping |
| 322 | // and we do not take ownership and are not responsible for |
| 323 | // unmapping. |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 324 | bool reuse_ = false; |
Jim_Guo | a62a588 | 2014-04-28 11:11:57 +0800 | [diff] [blame] | 325 | |
Alex Light | ca97ada | 2018-02-02 09:25:31 -0800 | [diff] [blame] | 326 | // When already_unmapped_ is true the destructor will not call munmap. |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 327 | bool already_unmapped_ = false; |
Alex Light | ca97ada | 2018-02-02 09:25:31 -0800 | [diff] [blame] | 328 | |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 329 | size_t redzone_size_ = 0u; |
Evgenii Stepanov | 1e13374 | 2015-05-20 12:30:59 -0700 | [diff] [blame] | 330 | |
Ian Rogers | c3ccc10 | 2014-06-25 11:52:14 -0700 | [diff] [blame] | 331 | #if USE_ART_LOW_4G_ALLOCATOR |
| 332 | static uintptr_t next_mem_pos_; // Next memory location to check for low_4g extent. |
Steve Austin | 882ed6b | 2018-06-08 11:40:38 -0700 | [diff] [blame] | 333 | |
| 334 | static void* TryMemMapLow4GB(void* ptr, |
| 335 | size_t page_aligned_byte_count, |
| 336 | int prot, |
| 337 | int flags, |
| 338 | int fd, |
| 339 | off_t offset); |
Stuart Monteith | 8dba5aa | 2014-03-12 12:44:01 +0000 | [diff] [blame] | 340 | #endif |
| 341 | |
Steve Austin | 882ed6b | 2018-06-08 11:40:38 -0700 | [diff] [blame] | 342 | static void TargetMMapInit(); |
| 343 | static void* TargetMMap(void* start, size_t len, int prot, int flags, int fd, off_t fd_off); |
| 344 | static int TargetMUnmap(void* start, size_t len); |
| 345 | |
David Sehr | 1b14fb8 | 2017-02-01 10:42:11 -0800 | [diff] [blame] | 346 | static std::mutex* mem_maps_lock_; |
| 347 | |
Hiroshi Yamauchi | fd7e7f1 | 2013-10-22 14:17:48 -0700 | [diff] [blame] | 348 | friend class MemMapTest; // To allow access to base_begin_ and base_size_. |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 349 | }; |
Mathieu Chartier | 6e6078a | 2016-10-24 15:45:41 -0700 | [diff] [blame] | 350 | |
Vladimir Marko | c34bebf | 2018-08-16 16:12:49 +0100 | [diff] [blame] | 351 | inline void swap(MemMap& lhs, MemMap& rhs) { |
| 352 | lhs.swap(rhs); |
| 353 | } |
| 354 | |
Brian Carlstrom | 0d6adac | 2014-02-05 17:39:16 -0800 | [diff] [blame] | 355 | std::ostream& operator<<(std::ostream& os, const MemMap& mem_map); |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 356 | |
Mathieu Chartier | 6e6078a | 2016-10-24 15:45:41 -0700 | [diff] [blame] | 357 | // Zero and release pages if possible, no requirements on alignments. |
| 358 | void ZeroAndReleasePages(void* address, size_t length); |
| 359 | |
Brian Carlstrom | db4d540 | 2011-08-09 12:18:28 -0700 | [diff] [blame] | 360 | } // namespace art |
| 361 | |
David Sehr | 79e2607 | 2018-04-06 17:58:50 -0700 | [diff] [blame] | 362 | #endif // ART_LIBARTBASE_BASE_MEM_MAP_H_ |