|
1 /* |
|
2 * Copyright (c) 2013, 2014, Oracle and/or its affiliates. All rights reserved. |
|
3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
|
4 * |
|
5 * This code is free software; you can redistribute it and/or modify it |
|
6 * under the terms of the GNU General Public License version 2 only, as |
|
7 * published by the Free Software Foundation. |
|
8 * |
|
9 * This code is distributed in the hope that it will be useful, but WITHOUT |
|
10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
|
11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
|
12 * version 2 for more details (a copy is included in the LICENSE file that |
|
13 * accompanied this code). |
|
14 * |
|
15 * You should have received a copy of the GNU General Public License version |
|
16 * 2 along with this work; if not, write to the Free Software Foundation, |
|
17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
|
18 * |
|
19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
|
20 * or visit www.oracle.com if you need additional information or have any |
|
21 * questions. |
|
22 * |
|
23 */ |
|
24 |
|
25 #ifndef SHARE_VM_SERVICES_VIRTUAL_MEMORY_TRACKER_HPP |
|
26 #define SHARE_VM_SERVICES_VIRTUAL_MEMORY_TRACKER_HPP |
|
27 |
|
28 #if INCLUDE_NMT |
|
29 |
|
30 #include "memory/allocation.hpp" |
|
31 #include "services/allocationSite.hpp" |
|
32 #include "services/nmtCommon.hpp" |
|
33 #include "utilities/linkedlist.hpp" |
|
34 #include "utilities/nativeCallStack.hpp" |
|
35 #include "utilities/ostream.hpp" |
|
36 |
|
37 |
|
38 /* |
|
39 * Virtual memory counter |
|
40 */ |
|
41 class VirtualMemory VALUE_OBJ_CLASS_SPEC { |
|
42 private: |
|
43 size_t _reserved; |
|
44 size_t _committed; |
|
45 |
|
46 public: |
|
47 VirtualMemory() : _reserved(0), _committed(0) { } |
|
48 |
|
49 inline void reserve_memory(size_t sz) { _reserved += sz; } |
|
50 inline void commit_memory (size_t sz) { |
|
51 _committed += sz; |
|
52 assert(_committed <= _reserved, "Sanity check"); |
|
53 } |
|
54 |
|
55 inline void release_memory (size_t sz) { |
|
56 assert(_reserved >= sz, "Negative amount"); |
|
57 _reserved -= sz; |
|
58 } |
|
59 |
|
60 inline void uncommit_memory(size_t sz) { |
|
61 assert(_committed >= sz, "Negative amount"); |
|
62 _committed -= sz; |
|
63 } |
|
64 |
|
65 inline size_t reserved() const { return _reserved; } |
|
66 inline size_t committed() const { return _committed; } |
|
67 }; |
|
68 |
|
69 // Virtual memory allocation site, keeps track where the virtual memory is reserved. |
|
70 class VirtualMemoryAllocationSite : public AllocationSite<VirtualMemory> { |
|
71 public: |
|
72 VirtualMemoryAllocationSite(const NativeCallStack& stack) : |
|
73 AllocationSite<VirtualMemory>(stack) { } |
|
74 |
|
75 inline void reserve_memory(size_t sz) { data()->reserve_memory(sz); } |
|
76 inline void commit_memory (size_t sz) { data()->commit_memory(sz); } |
|
77 inline void uncommit_memory(size_t sz) { data()->uncommit_memory(sz); } |
|
78 inline void release_memory(size_t sz) { data()->release_memory(sz); } |
|
79 inline size_t reserved() const { return peek()->reserved(); } |
|
80 inline size_t committed() const { return peek()->committed(); } |
|
81 }; |
|
82 |
|
83 class VirtualMemorySummary; |
|
84 |
|
85 // This class represents a snapshot of virtual memory at a given time. |
|
86 // The latest snapshot is saved in a static area. |
|
87 class VirtualMemorySnapshot : public ResourceObj { |
|
88 friend class VirtualMemorySummary; |
|
89 |
|
90 private: |
|
91 VirtualMemory _virtual_memory[mt_number_of_types]; |
|
92 |
|
93 public: |
|
94 inline VirtualMemory* by_type(MEMFLAGS flag) { |
|
95 int index = NMTUtil::flag_to_index(flag); |
|
96 return &_virtual_memory[index]; |
|
97 } |
|
98 |
|
99 inline VirtualMemory* by_index(int index) { |
|
100 assert(index >= 0, "Index out of bound"); |
|
101 assert(index < mt_number_of_types, "Index out of bound"); |
|
102 return &_virtual_memory[index]; |
|
103 } |
|
104 |
|
105 inline size_t total_reserved() const { |
|
106 size_t amount = 0; |
|
107 for (int index = 0; index < mt_number_of_types; index ++) { |
|
108 amount += _virtual_memory[index].reserved(); |
|
109 } |
|
110 return amount; |
|
111 } |
|
112 |
|
113 inline size_t total_committed() const { |
|
114 size_t amount = 0; |
|
115 for (int index = 0; index < mt_number_of_types; index ++) { |
|
116 amount += _virtual_memory[index].committed(); |
|
117 } |
|
118 return amount; |
|
119 } |
|
120 |
|
121 void copy_to(VirtualMemorySnapshot* s) { |
|
122 for (int index = 0; index < mt_number_of_types; index ++) { |
|
123 s->_virtual_memory[index] = _virtual_memory[index]; |
|
124 } |
|
125 } |
|
126 }; |
|
127 |
|
128 class VirtualMemorySummary : AllStatic { |
|
129 public: |
|
130 static void initialize(); |
|
131 |
|
132 static inline void record_reserved_memory(size_t size, MEMFLAGS flag) { |
|
133 as_snapshot()->by_type(flag)->reserve_memory(size); |
|
134 } |
|
135 |
|
136 static inline void record_committed_memory(size_t size, MEMFLAGS flag) { |
|
137 as_snapshot()->by_type(flag)->commit_memory(size); |
|
138 } |
|
139 |
|
140 static inline void record_uncommitted_memory(size_t size, MEMFLAGS flag) { |
|
141 as_snapshot()->by_type(flag)->uncommit_memory(size); |
|
142 } |
|
143 |
|
144 static inline void record_released_memory(size_t size, MEMFLAGS flag) { |
|
145 as_snapshot()->by_type(flag)->release_memory(size); |
|
146 } |
|
147 |
|
148 // Move virtual memory from one memory type to another. |
|
149 // Virtual memory can be reserved before it is associated with a memory type, and tagged |
|
150 // as 'unknown'. Once the memory is tagged, the virtual memory will be moved from 'unknown' |
|
151 // type to specified memory type. |
|
152 static inline void move_reserved_memory(MEMFLAGS from, MEMFLAGS to, size_t size) { |
|
153 as_snapshot()->by_type(from)->release_memory(size); |
|
154 as_snapshot()->by_type(to)->reserve_memory(size); |
|
155 } |
|
156 |
|
157 static inline void move_committed_memory(MEMFLAGS from, MEMFLAGS to, size_t size) { |
|
158 as_snapshot()->by_type(from)->uncommit_memory(size); |
|
159 as_snapshot()->by_type(to)->commit_memory(size); |
|
160 } |
|
161 |
|
162 static inline void snapshot(VirtualMemorySnapshot* s) { |
|
163 as_snapshot()->copy_to(s); |
|
164 } |
|
165 |
|
166 static VirtualMemorySnapshot* as_snapshot() { |
|
167 return (VirtualMemorySnapshot*)_snapshot; |
|
168 } |
|
169 |
|
170 private: |
|
171 static size_t _snapshot[CALC_OBJ_SIZE_IN_TYPE(VirtualMemorySnapshot, size_t)]; |
|
172 }; |
|
173 |
|
174 |
|
175 |
|
176 /* |
|
177 * A virtual memory region |
|
178 */ |
|
179 class VirtualMemoryRegion VALUE_OBJ_CLASS_SPEC { |
|
180 private: |
|
181 address _base_address; |
|
182 size_t _size; |
|
183 |
|
184 public: |
|
185 VirtualMemoryRegion(address addr, size_t size) : |
|
186 _base_address(addr), _size(size) { |
|
187 assert(addr != NULL, "Invalid address"); |
|
188 assert(size > 0, "Invalid size"); |
|
189 } |
|
190 |
|
191 inline address base() const { return _base_address; } |
|
192 inline address end() const { return base() + size(); } |
|
193 inline size_t size() const { return _size; } |
|
194 |
|
195 inline bool is_empty() const { return size() == 0; } |
|
196 |
|
197 inline bool contain_address(address addr) const { |
|
198 return (addr >= base() && addr < end()); |
|
199 } |
|
200 |
|
201 |
|
202 inline bool contain_region(address addr, size_t size) const { |
|
203 return contain_address(addr) && contain_address(addr + size - 1); |
|
204 } |
|
205 |
|
206 inline bool same_region(address addr, size_t sz) const { |
|
207 return (addr == base() && sz == size()); |
|
208 } |
|
209 |
|
210 |
|
211 inline bool overlap_region(address addr, size_t sz) const { |
|
212 VirtualMemoryRegion rgn(addr, sz); |
|
213 return contain_address(addr) || |
|
214 contain_address(addr + sz - 1) || |
|
215 rgn.contain_address(base()) || |
|
216 rgn.contain_address(end() - 1); |
|
217 } |
|
218 |
|
219 inline bool adjacent_to(address addr, size_t sz) const { |
|
220 return (addr == end() || (addr + sz) == base()); |
|
221 } |
|
222 |
|
223 void exclude_region(address addr, size_t sz) { |
|
224 assert(contain_region(addr, sz), "Not containment"); |
|
225 assert(addr == base() || addr + sz == end(), "Can not exclude from middle"); |
|
226 size_t new_size = size() - sz; |
|
227 |
|
228 if (addr == base()) { |
|
229 set_base(addr + sz); |
|
230 } |
|
231 set_size(new_size); |
|
232 } |
|
233 |
|
234 void expand_region(address addr, size_t sz) { |
|
235 assert(adjacent_to(addr, sz), "Not adjacent regions"); |
|
236 if (base() == addr + sz) { |
|
237 set_base(addr); |
|
238 } |
|
239 set_size(size() + sz); |
|
240 } |
|
241 |
|
242 protected: |
|
243 void set_base(address base) { |
|
244 assert(base != NULL, "Sanity check"); |
|
245 _base_address = base; |
|
246 } |
|
247 |
|
248 void set_size(size_t size) { |
|
249 assert(size > 0, "Sanity check"); |
|
250 _size = size; |
|
251 } |
|
252 }; |
|
253 |
|
254 |
|
255 class CommittedMemoryRegion : public VirtualMemoryRegion { |
|
256 private: |
|
257 NativeCallStack _stack; |
|
258 |
|
259 public: |
|
260 CommittedMemoryRegion(address addr, size_t size, const NativeCallStack& stack) : |
|
261 VirtualMemoryRegion(addr, size), _stack(stack) { } |
|
262 |
|
263 inline int compare(const CommittedMemoryRegion& rgn) const { |
|
264 if (overlap_region(rgn.base(), rgn.size())) { |
|
265 return 0; |
|
266 } else { |
|
267 if (base() == rgn.base()) { |
|
268 return 0; |
|
269 } else if (base() > rgn.base()) { |
|
270 return 1; |
|
271 } else { |
|
272 return -1; |
|
273 } |
|
274 } |
|
275 } |
|
276 |
|
277 inline bool equals(const CommittedMemoryRegion& rgn) const { |
|
278 return compare(rgn) == 0; |
|
279 } |
|
280 |
|
281 inline void set_call_stack(const NativeCallStack& stack) { _stack = stack; } |
|
282 inline const NativeCallStack* call_stack() const { return &_stack; } |
|
283 }; |
|
284 |
|
285 |
|
286 typedef LinkedListIterator<CommittedMemoryRegion> CommittedRegionIterator; |
|
287 |
|
288 int compare_committed_region(const CommittedMemoryRegion&, const CommittedMemoryRegion&); |
|
289 class ReservedMemoryRegion : public VirtualMemoryRegion { |
|
290 private: |
|
291 SortedLinkedList<CommittedMemoryRegion, compare_committed_region> |
|
292 _committed_regions; |
|
293 |
|
294 NativeCallStack _stack; |
|
295 MEMFLAGS _flag; |
|
296 |
|
297 bool _all_committed; |
|
298 |
|
299 public: |
|
300 ReservedMemoryRegion(address base, size_t size, const NativeCallStack& stack, |
|
301 MEMFLAGS flag = mtNone) : |
|
302 VirtualMemoryRegion(base, size), _stack(stack), _flag(flag), |
|
303 _all_committed(false) { } |
|
304 |
|
305 |
|
306 ReservedMemoryRegion(address base, size_t size) : |
|
307 VirtualMemoryRegion(base, size), _stack(NativeCallStack::EMPTY_STACK), _flag(mtNone), |
|
308 _all_committed(false) { } |
|
309 |
|
310 // Copy constructor |
|
311 ReservedMemoryRegion(const ReservedMemoryRegion& rr) : |
|
312 VirtualMemoryRegion(rr.base(), rr.size()) { |
|
313 *this = rr; |
|
314 } |
|
315 |
|
316 inline void set_call_stack(const NativeCallStack& stack) { _stack = stack; } |
|
317 inline const NativeCallStack* call_stack() const { return &_stack; } |
|
318 |
|
319 void set_flag(MEMFLAGS flag); |
|
320 inline MEMFLAGS flag() const { return _flag; } |
|
321 |
|
322 inline int compare(const ReservedMemoryRegion& rgn) const { |
|
323 if (overlap_region(rgn.base(), rgn.size())) { |
|
324 return 0; |
|
325 } else { |
|
326 if (base() == rgn.base()) { |
|
327 return 0; |
|
328 } else if (base() > rgn.base()) { |
|
329 return 1; |
|
330 } else { |
|
331 return -1; |
|
332 } |
|
333 } |
|
334 } |
|
335 |
|
336 inline bool equals(const ReservedMemoryRegion& rgn) const { |
|
337 return compare(rgn) == 0; |
|
338 } |
|
339 |
|
340 bool add_committed_region(address addr, size_t size, const NativeCallStack& stack); |
|
341 bool remove_uncommitted_region(address addr, size_t size); |
|
342 |
|
343 size_t committed_size() const; |
|
344 |
|
345 // move committed regions that higher than specified address to |
|
346 // the new region |
|
347 void move_committed_regions(address addr, ReservedMemoryRegion& rgn); |
|
348 |
|
349 inline bool all_committed() const { return _all_committed; } |
|
350 void set_all_committed(bool b); |
|
351 |
|
352 CommittedRegionIterator iterate_committed_regions() const { |
|
353 return CommittedRegionIterator(_committed_regions.head()); |
|
354 } |
|
355 |
|
356 ReservedMemoryRegion& operator= (const ReservedMemoryRegion& other) { |
|
357 set_base(other.base()); |
|
358 set_size(other.size()); |
|
359 |
|
360 _stack = *other.call_stack(); |
|
361 _flag = other.flag(); |
|
362 _all_committed = other.all_committed(); |
|
363 if (other.all_committed()) { |
|
364 set_all_committed(true); |
|
365 } else { |
|
366 CommittedRegionIterator itr = other.iterate_committed_regions(); |
|
367 const CommittedMemoryRegion* rgn = itr.next(); |
|
368 while (rgn != NULL) { |
|
369 _committed_regions.add(*rgn); |
|
370 rgn = itr.next(); |
|
371 } |
|
372 } |
|
373 return *this; |
|
374 } |
|
375 |
|
376 private: |
|
377 // The committed region contains the uncommitted region, subtract the uncommitted |
|
378 // region from this committed region |
|
379 bool remove_uncommitted_region(LinkedListNode<CommittedMemoryRegion>* node, |
|
380 address addr, size_t sz); |
|
381 |
|
382 bool add_committed_region(const CommittedMemoryRegion& rgn) { |
|
383 assert(rgn.base() != NULL, "Invalid base address"); |
|
384 assert(size() > 0, "Invalid size"); |
|
385 return _committed_regions.add(rgn) != NULL; |
|
386 } |
|
387 }; |
|
388 |
|
389 int compare_reserved_region_base(const ReservedMemoryRegion& r1, const ReservedMemoryRegion& r2); |
|
390 |
|
391 class VirtualMemoryWalker : public StackObj { |
|
392 public: |
|
393 virtual bool do_allocation_site(const ReservedMemoryRegion* rgn) { return false; } |
|
394 }; |
|
395 |
|
396 // Main class called from MemTracker to track virtual memory allocations, commits and releases. |
|
397 class VirtualMemoryTracker : AllStatic { |
|
398 public: |
|
399 static bool initialize(NMT_TrackingLevel level); |
|
400 |
|
401 // Late phase initialization |
|
402 static bool late_initialize(NMT_TrackingLevel level); |
|
403 |
|
404 static bool add_reserved_region (address base_addr, size_t size, const NativeCallStack& stack, |
|
405 MEMFLAGS flag = mtNone, bool all_committed = false); |
|
406 |
|
407 static bool add_committed_region (address base_addr, size_t size, const NativeCallStack& stack); |
|
408 static bool remove_uncommitted_region (address base_addr, size_t size); |
|
409 static bool remove_released_region (address base_addr, size_t size); |
|
410 static void set_reserved_region_type (address addr, MEMFLAGS flag); |
|
411 |
|
412 // Walk virtual memory data structure for creating baseline, etc. |
|
413 static bool walk_virtual_memory(VirtualMemoryWalker* walker); |
|
414 |
|
415 static bool transition(NMT_TrackingLevel from, NMT_TrackingLevel to); |
|
416 |
|
417 private: |
|
418 static SortedLinkedList<ReservedMemoryRegion, compare_reserved_region_base>* _reserved_regions; |
|
419 }; |
|
420 |
|
421 |
|
422 #endif // INCLUDE_NMT |
|
423 |
|
424 #endif // SHARE_VM_SERVICES_VIRTUAL_MEMORY_TRACKER_HPP |