author | dcubed |
Sat, 21 Sep 2019 10:13:25 -0400 | |
changeset 58252 | 14c1ff687621 |
parent 53970 | 1ad7c590a6e7 |
child 57464 | 32e61f51ee09 |
child 58063 | bdf136b8ae0e |
child 59292 | 95b1385dd476 |
permissions | -rw-r--r-- |
50193 | 1 |
/* |
2 |
* Copyright (c) 2018, Oracle and/or its affiliates. All rights reserved. |
|
3 |
* DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. |
|
4 |
* |
|
5 |
* This code is free software; you can redistribute it and/or modify it |
|
6 |
* under the terms of the GNU General Public License version 2 only, as |
|
7 |
* published by the Free Software Foundation. |
|
8 |
* |
|
9 |
* This code is distributed in the hope that it will be useful, but WITHOUT |
|
10 |
* ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
|
11 |
* FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License |
|
12 |
* version 2 for more details (a copy is included in the LICENSE file that |
|
13 |
* accompanied this code). |
|
14 |
* |
|
15 |
* You should have received a copy of the GNU General Public License version |
|
16 |
* 2 along with this work; if not, write to the Free Software Foundation, |
|
17 |
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. |
|
18 |
* |
|
19 |
* Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA |
|
20 |
* or visit www.oracle.com if you need additional information or have any |
|
21 |
* questions. |
|
22 |
* |
|
23 |
*/ |
|
24 |
||
25 |
#include "precompiled.hpp" |
|
26 |
||
27 |
#include "logging/log.hpp" |
|
28 |
#include "logging/logStream.hpp" |
|
29 |
#include "memory/metaspace/metachunk.hpp" |
|
30 |
#include "memory/metaspace.hpp" |
|
31 |
#include "memory/metaspace/chunkManager.hpp" |
|
53970 | 32 |
#include "memory/metaspace/metaDebug.hpp" |
50193 | 33 |
#include "memory/metaspace/metaspaceCommon.hpp" |
34 |
#include "memory/metaspace/occupancyMap.hpp" |
|
35 |
#include "memory/metaspace/virtualSpaceNode.hpp" |
|
36 |
#include "memory/virtualspace.hpp" |
|
37 |
#include "runtime/os.hpp" |
|
38 |
#include "services/memTracker.hpp" |
|
39 |
#include "utilities/copy.hpp" |
|
40 |
#include "utilities/debug.hpp" |
|
41 |
#include "utilities/globalDefinitions.hpp" |
|
42 |
||
43 |
namespace metaspace { |
|
44 |
||
45 |
// Decide if large pages should be committed when the memory is reserved. |
|
46 |
static bool should_commit_large_pages_when_reserving(size_t bytes) { |
|
47 |
if (UseLargePages && UseLargePagesInMetaspace && !os::can_commit_large_page_memory()) { |
|
48 |
size_t words = bytes / BytesPerWord; |
|
49 |
bool is_class = false; // We never reserve large pages for the class space. |
|
50 |
if (MetaspaceGC::can_expand(words, is_class) && |
|
51 |
MetaspaceGC::allowed_expansion() >= words) { |
|
52 |
return true; |
|
53 |
} |
|
54 |
} |
|
55 |
||
56 |
return false; |
|
57 |
} |
|
58 |
||
59 |
// byte_size is the size of the associated virtualspace. |
|
60 |
VirtualSpaceNode::VirtualSpaceNode(bool is_class, size_t bytes) : |
|
51334
cc2c79d22508
8208671: Runtime, JFR, Serviceability changes to allow enabling -Wreorder
tschatzl
parents:
50811
diff
changeset
|
61 |
_next(NULL), _is_class(is_class), _rs(), _top(NULL), _container_count(0), _occupancy_map(NULL) { |
50193 | 62 |
assert_is_aligned(bytes, Metaspace::reserve_alignment()); |
63 |
bool large_pages = should_commit_large_pages_when_reserving(bytes); |
|
64 |
_rs = ReservedSpace(bytes, Metaspace::reserve_alignment(), large_pages); |
|
65 |
||
66 |
if (_rs.is_reserved()) { |
|
67 |
assert(_rs.base() != NULL, "Catch if we get a NULL address"); |
|
68 |
assert(_rs.size() != 0, "Catch if we get a 0 size"); |
|
69 |
assert_is_aligned(_rs.base(), Metaspace::reserve_alignment()); |
|
70 |
assert_is_aligned(_rs.size(), Metaspace::reserve_alignment()); |
|
71 |
||
72 |
MemTracker::record_virtual_memory_type((address)_rs.base(), mtClass); |
|
73 |
} |
|
74 |
} |
|
75 |
||
76 |
void VirtualSpaceNode::purge(ChunkManager* chunk_manager) { |
|
53970 | 77 |
// When a node is purged, lets give it a thorough examination. |
78 |
DEBUG_ONLY(verify(true);) |
|
79 |
Metachunk* chunk = first_chunk(); |
|
50193 | 80 |
Metachunk* invalid_chunk = (Metachunk*) top(); |
81 |
while (chunk < invalid_chunk ) { |
|
82 |
assert(chunk->is_tagged_free(), "Should be tagged free"); |
|
83 |
MetaWord* next = ((MetaWord*)chunk) + chunk->word_size(); |
|
84 |
chunk_manager->remove_chunk(chunk); |
|
85 |
chunk->remove_sentinel(); |
|
86 |
assert(chunk->next() == NULL && |
|
87 |
chunk->prev() == NULL, |
|
88 |
"Was not removed from its list"); |
|
89 |
chunk = (Metachunk*) next; |
|
90 |
} |
|
91 |
} |
|
92 |
||
93 |
void VirtualSpaceNode::print_map(outputStream* st, bool is_class) const { |
|
94 |
||
95 |
if (bottom() == top()) { |
|
96 |
return; |
|
97 |
} |
|
98 |
||
99 |
const size_t spec_chunk_size = is_class ? ClassSpecializedChunk : SpecializedChunk; |
|
100 |
const size_t small_chunk_size = is_class ? ClassSmallChunk : SmallChunk; |
|
101 |
const size_t med_chunk_size = is_class ? ClassMediumChunk : MediumChunk; |
|
102 |
||
103 |
int line_len = 100; |
|
104 |
const size_t section_len = align_up(spec_chunk_size * line_len, med_chunk_size); |
|
105 |
line_len = (int)(section_len / spec_chunk_size); |
|
106 |
||
107 |
static const int NUM_LINES = 4; |
|
108 |
||
109 |
char* lines[NUM_LINES]; |
|
110 |
for (int i = 0; i < NUM_LINES; i ++) { |
|
111 |
lines[i] = (char*)os::malloc(line_len, mtInternal); |
|
112 |
} |
|
113 |
int pos = 0; |
|
114 |
const MetaWord* p = bottom(); |
|
115 |
const Metachunk* chunk = (const Metachunk*)p; |
|
116 |
const MetaWord* chunk_end = p + chunk->word_size(); |
|
117 |
while (p < top()) { |
|
118 |
if (pos == line_len) { |
|
119 |
pos = 0; |
|
120 |
for (int i = 0; i < NUM_LINES; i ++) { |
|
121 |
st->fill_to(22); |
|
122 |
st->print_raw(lines[i], line_len); |
|
123 |
st->cr(); |
|
124 |
} |
|
125 |
} |
|
126 |
if (pos == 0) { |
|
127 |
st->print(PTR_FORMAT ":", p2i(p)); |
|
128 |
} |
|
129 |
if (p == chunk_end) { |
|
130 |
chunk = (Metachunk*)p; |
|
131 |
chunk_end = p + chunk->word_size(); |
|
132 |
} |
|
133 |
// line 1: chunk starting points (a dot if that area is a chunk start). |
|
134 |
lines[0][pos] = p == (const MetaWord*)chunk ? '.' : ' '; |
|
135 |
||
136 |
// Line 2: chunk type (x=spec, s=small, m=medium, h=humongous), uppercase if |
|
137 |
// chunk is in use. |
|
138 |
const bool chunk_is_free = ((Metachunk*)chunk)->is_tagged_free(); |
|
139 |
if (chunk->word_size() == spec_chunk_size) { |
|
140 |
lines[1][pos] = chunk_is_free ? 'x' : 'X'; |
|
141 |
} else if (chunk->word_size() == small_chunk_size) { |
|
142 |
lines[1][pos] = chunk_is_free ? 's' : 'S'; |
|
143 |
} else if (chunk->word_size() == med_chunk_size) { |
|
144 |
lines[1][pos] = chunk_is_free ? 'm' : 'M'; |
|
145 |
} else if (chunk->word_size() > med_chunk_size) { |
|
146 |
lines[1][pos] = chunk_is_free ? 'h' : 'H'; |
|
147 |
} else { |
|
148 |
ShouldNotReachHere(); |
|
149 |
} |
|
150 |
||
151 |
// Line 3: chunk origin |
|
152 |
const ChunkOrigin origin = chunk->get_origin(); |
|
153 |
lines[2][pos] = origin == origin_normal ? ' ' : '0' + (int) origin; |
|
154 |
||
155 |
// Line 4: Virgin chunk? Virgin chunks are chunks created as a byproduct of padding or splitting, |
|
156 |
// but were never used. |
|
157 |
lines[3][pos] = chunk->get_use_count() > 0 ? ' ' : 'v'; |
|
158 |
||
159 |
p += spec_chunk_size; |
|
160 |
pos ++; |
|
161 |
} |
|
162 |
if (pos > 0) { |
|
163 |
for (int i = 0; i < NUM_LINES; i ++) { |
|
164 |
st->fill_to(22); |
|
165 |
st->print_raw(lines[i], line_len); |
|
166 |
st->cr(); |
|
167 |
} |
|
168 |
} |
|
169 |
for (int i = 0; i < NUM_LINES; i ++) { |
|
170 |
os::free(lines[i]); |
|
171 |
} |
|
172 |
} |
|
173 |
||
174 |
||
175 |
#ifdef ASSERT |
|
176 |
||
177 |
// Verify counters, all chunks in this list node and the occupancy map. |
|
53970 | 178 |
void VirtualSpaceNode::verify(bool slow) { |
179 |
log_trace(gc, metaspace, freelist)("verifying %s virtual space node (%s).", |
|
180 |
(is_class() ? "class space" : "metaspace"), (slow ? "slow" : "quick")); |
|
181 |
// Fast mode: just verify chunk counters and basic geometry |
|
182 |
// Slow mode: verify chunks and occupancy map |
|
50193 | 183 |
uintx num_in_use_chunks = 0; |
184 |
Metachunk* chunk = first_chunk(); |
|
185 |
Metachunk* invalid_chunk = (Metachunk*) top(); |
|
186 |
||
187 |
// Iterate the chunks in this node and verify each chunk. |
|
188 |
while (chunk < invalid_chunk ) { |
|
53970 | 189 |
if (slow) { |
190 |
do_verify_chunk(chunk); |
|
191 |
} |
|
192 |
if (!chunk->is_tagged_free()) { |
|
193 |
num_in_use_chunks ++; |
|
194 |
} |
|
195 |
const size_t s = chunk->word_size(); |
|
196 |
// Prevent endless loop on invalid chunk size. |
|
197 |
assert(is_valid_chunksize(is_class(), s), "Invalid chunk size: " SIZE_FORMAT ".", s); |
|
198 |
MetaWord* next = ((MetaWord*)chunk) + s; |
|
50193 | 199 |
chunk = (Metachunk*) next; |
200 |
} |
|
201 |
assert(_container_count == num_in_use_chunks, "Container count mismatch (real: " UINTX_FORMAT |
|
202 |
", counter: " UINTX_FORMAT ".", num_in_use_chunks, _container_count); |
|
203 |
// Also verify the occupancy map. |
|
53970 | 204 |
if (slow) { |
205 |
occupancy_map()->verify(bottom(), top()); |
|
206 |
} |
|
50193 | 207 |
} |
208 |
||
209 |
// Verify that all free chunks in this node are ideally merged |
|
210 |
// (there not should be multiple small chunks where a large chunk could exist.) |
|
211 |
void VirtualSpaceNode::verify_free_chunks_are_ideally_merged() { |
|
212 |
Metachunk* chunk = first_chunk(); |
|
213 |
Metachunk* invalid_chunk = (Metachunk*) top(); |
|
214 |
// Shorthands. |
|
215 |
const size_t size_med = (is_class() ? ClassMediumChunk : MediumChunk) * BytesPerWord; |
|
216 |
const size_t size_small = (is_class() ? ClassSmallChunk : SmallChunk) * BytesPerWord; |
|
217 |
int num_free_chunks_since_last_med_boundary = -1; |
|
218 |
int num_free_chunks_since_last_small_boundary = -1; |
|
53970 | 219 |
bool error = false; |
220 |
char err[256]; |
|
221 |
while (!error && chunk < invalid_chunk ) { |
|
50193 | 222 |
// Test for missed chunk merge opportunities: count number of free chunks since last chunk boundary. |
223 |
// Reset the counter when encountering a non-free chunk. |
|
224 |
if (chunk->get_chunk_type() != HumongousIndex) { |
|
225 |
if (chunk->is_tagged_free()) { |
|
226 |
// Count successive free, non-humongous chunks. |
|
227 |
if (is_aligned(chunk, size_small)) { |
|
53970 | 228 |
if (num_free_chunks_since_last_small_boundary > 0) { |
229 |
error = true; |
|
230 |
jio_snprintf(err, sizeof(err), "Missed chunk merge opportunity to merge a small chunk preceding " PTR_FORMAT ".", p2i(chunk)); |
|
231 |
} else { |
|
232 |
num_free_chunks_since_last_small_boundary = 0; |
|
233 |
} |
|
50193 | 234 |
} else if (num_free_chunks_since_last_small_boundary != -1) { |
235 |
num_free_chunks_since_last_small_boundary ++; |
|
236 |
} |
|
237 |
if (is_aligned(chunk, size_med)) { |
|
53970 | 238 |
if (num_free_chunks_since_last_med_boundary > 0) { |
239 |
error = true; |
|
240 |
jio_snprintf(err, sizeof(err), "Missed chunk merge opportunity to merge a medium chunk preceding " PTR_FORMAT ".", p2i(chunk)); |
|
241 |
} else { |
|
242 |
num_free_chunks_since_last_med_boundary = 0; |
|
243 |
} |
|
50193 | 244 |
} else if (num_free_chunks_since_last_med_boundary != -1) { |
245 |
num_free_chunks_since_last_med_boundary ++; |
|
246 |
} |
|
247 |
} else { |
|
248 |
// Encountering a non-free chunk, reset counters. |
|
249 |
num_free_chunks_since_last_med_boundary = -1; |
|
250 |
num_free_chunks_since_last_small_boundary = -1; |
|
251 |
} |
|
252 |
} else { |
|
253 |
// One cannot merge areas with a humongous chunk in the middle. Reset counters. |
|
254 |
num_free_chunks_since_last_med_boundary = -1; |
|
255 |
num_free_chunks_since_last_small_boundary = -1; |
|
256 |
} |
|
257 |
||
53970 | 258 |
if (error) { |
259 |
print_map(tty, is_class()); |
|
260 |
fatal("%s", err); |
|
261 |
} |
|
262 |
||
50193 | 263 |
MetaWord* next = ((MetaWord*)chunk) + chunk->word_size(); |
264 |
chunk = (Metachunk*) next; |
|
265 |
} |
|
266 |
} |
|
267 |
#endif // ASSERT |
|
268 |
||
269 |
void VirtualSpaceNode::inc_container_count() { |
|
270 |
assert_lock_strong(MetaspaceExpand_lock); |
|
271 |
_container_count++; |
|
272 |
} |
|
273 |
||
274 |
void VirtualSpaceNode::dec_container_count() { |
|
275 |
assert_lock_strong(MetaspaceExpand_lock); |
|
276 |
_container_count--; |
|
277 |
} |
|
278 |
||
279 |
VirtualSpaceNode::~VirtualSpaceNode() { |
|
280 |
_rs.release(); |
|
281 |
if (_occupancy_map != NULL) { |
|
282 |
delete _occupancy_map; |
|
283 |
} |
|
284 |
#ifdef ASSERT |
|
285 |
size_t word_size = sizeof(*this) / BytesPerWord; |
|
286 |
Copy::fill_to_words((HeapWord*) this, word_size, 0xf1f1f1f1); |
|
287 |
#endif |
|
288 |
} |
|
289 |
||
290 |
size_t VirtualSpaceNode::used_words_in_vs() const { |
|
291 |
return pointer_delta(top(), bottom(), sizeof(MetaWord)); |
|
292 |
} |
|
293 |
||
294 |
// Space committed in the VirtualSpace |
|
295 |
size_t VirtualSpaceNode::capacity_words_in_vs() const { |
|
296 |
return pointer_delta(end(), bottom(), sizeof(MetaWord)); |
|
297 |
} |
|
298 |
||
299 |
size_t VirtualSpaceNode::free_words_in_vs() const { |
|
300 |
return pointer_delta(end(), top(), sizeof(MetaWord)); |
|
301 |
} |
|
302 |
||
303 |
// Given an address larger than top(), allocate padding chunks until top is at the given address. |
|
304 |
void VirtualSpaceNode::allocate_padding_chunks_until_top_is_at(MetaWord* target_top) { |
|
305 |
||
306 |
assert(target_top > top(), "Sanity"); |
|
307 |
||
308 |
// Padding chunks are added to the freelist. |
|
53970 | 309 |
ChunkManager* const chunk_manager = Metaspace::get_chunk_manager(is_class()); |
50193 | 310 |
|
311 |
// shorthands |
|
312 |
const size_t spec_word_size = chunk_manager->specialized_chunk_word_size(); |
|
313 |
const size_t small_word_size = chunk_manager->small_chunk_word_size(); |
|
314 |
const size_t med_word_size = chunk_manager->medium_chunk_word_size(); |
|
315 |
||
316 |
while (top() < target_top) { |
|
317 |
||
318 |
// We could make this coding more generic, but right now we only deal with two possible chunk sizes |
|
319 |
// for padding chunks, so it is not worth it. |
|
320 |
size_t padding_chunk_word_size = small_word_size; |
|
321 |
if (is_aligned(top(), small_word_size * sizeof(MetaWord)) == false) { |
|
322 |
assert_is_aligned(top(), spec_word_size * sizeof(MetaWord)); // Should always hold true. |
|
323 |
padding_chunk_word_size = spec_word_size; |
|
324 |
} |
|
325 |
MetaWord* here = top(); |
|
326 |
assert_is_aligned(here, padding_chunk_word_size * sizeof(MetaWord)); |
|
327 |
inc_top(padding_chunk_word_size); |
|
328 |
||
329 |
// Create new padding chunk. |
|
330 |
ChunkIndex padding_chunk_type = get_chunk_type_by_size(padding_chunk_word_size, is_class()); |
|
331 |
assert(padding_chunk_type == SpecializedIndex || padding_chunk_type == SmallIndex, "sanity"); |
|
332 |
||
333 |
Metachunk* const padding_chunk = |
|
334 |
::new (here) Metachunk(padding_chunk_type, is_class(), padding_chunk_word_size, this); |
|
335 |
assert(padding_chunk == (Metachunk*)here, "Sanity"); |
|
336 |
DEBUG_ONLY(padding_chunk->set_origin(origin_pad);) |
|
337 |
log_trace(gc, metaspace, freelist)("Created padding chunk in %s at " |
|
338 |
PTR_FORMAT ", size " SIZE_FORMAT_HEX ".", |
|
339 |
(is_class() ? "class space " : "metaspace"), |
|
340 |
p2i(padding_chunk), padding_chunk->word_size() * sizeof(MetaWord)); |
|
341 |
||
342 |
// Mark chunk start in occupancy map. |
|
343 |
occupancy_map()->set_chunk_starts_at_address((MetaWord*)padding_chunk, true); |
|
344 |
||
345 |
// Chunks are born as in-use (see MetaChunk ctor). So, before returning |
|
346 |
// the padding chunk to its chunk manager, mark it as in use (ChunkManager |
|
347 |
// will assert that). |
|
348 |
do_update_in_use_info_for_chunk(padding_chunk, true); |
|
349 |
||
350 |
// Return Chunk to freelist. |
|
351 |
inc_container_count(); |
|
352 |
chunk_manager->return_single_chunk(padding_chunk); |
|
353 |
// Please note: at this point, ChunkManager::return_single_chunk() |
|
354 |
// may already have merged the padding chunk with neighboring chunks, so |
|
355 |
// it may have vanished at this point. Do not reference the padding |
|
356 |
// chunk beyond this point. |
|
357 |
} |
|
358 |
||
359 |
assert(top() == target_top, "Sanity"); |
|
360 |
||
361 |
} // allocate_padding_chunks_until_top_is_at() |
|
362 |
||
363 |
// Allocates the chunk from the virtual space only. |
|
364 |
// This interface is also used internally for debugging. Not all |
|
365 |
// chunks removed here are necessarily used for allocation. |
|
366 |
Metachunk* VirtualSpaceNode::take_from_committed(size_t chunk_word_size) { |
|
367 |
// Non-humongous chunks are to be allocated aligned to their chunk |
|
368 |
// size. So, start addresses of medium chunks are aligned to medium |
|
369 |
// chunk size, those of small chunks to small chunk size and so |
|
370 |
// forth. This facilitates merging of free chunks and reduces |
|
371 |
// fragmentation. Chunk sizes are spec < small < medium, with each |
|
372 |
// larger chunk size being a multiple of the next smaller chunk |
|
373 |
// size. |
|
374 |
// Because of this alignment, me may need to create a number of padding |
|
375 |
// chunks. These chunks are created and added to the freelist. |
|
376 |
||
377 |
// The chunk manager to which we will give our padding chunks. |
|
53970 | 378 |
ChunkManager* const chunk_manager = Metaspace::get_chunk_manager(is_class()); |
50193 | 379 |
|
380 |
// shorthands |
|
381 |
const size_t spec_word_size = chunk_manager->specialized_chunk_word_size(); |
|
382 |
const size_t small_word_size = chunk_manager->small_chunk_word_size(); |
|
383 |
const size_t med_word_size = chunk_manager->medium_chunk_word_size(); |
|
384 |
||
385 |
assert(chunk_word_size == spec_word_size || chunk_word_size == small_word_size || |
|
386 |
chunk_word_size >= med_word_size, "Invalid chunk size requested."); |
|
387 |
||
388 |
// Chunk alignment (in bytes) == chunk size unless humongous. |
|
389 |
// Humongous chunks are aligned to the smallest chunk size (spec). |
|
390 |
const size_t required_chunk_alignment = (chunk_word_size > med_word_size ? |
|
391 |
spec_word_size : chunk_word_size) * sizeof(MetaWord); |
|
392 |
||
393 |
// Do we have enough space to create the requested chunk plus |
|
394 |
// any padding chunks needed? |
|
395 |
MetaWord* const next_aligned = |
|
396 |
static_cast<MetaWord*>(align_up(top(), required_chunk_alignment)); |
|
397 |
if (!is_available((next_aligned - top()) + chunk_word_size)) { |
|
398 |
return NULL; |
|
399 |
} |
|
400 |
||
401 |
// Before allocating the requested chunk, allocate padding chunks if necessary. |
|
402 |
// We only need to do this for small or medium chunks: specialized chunks are the |
|
403 |
// smallest size, hence always aligned. Homungous chunks are allocated unaligned |
|
404 |
// (implicitly, also aligned to smallest chunk size). |
|
405 |
if ((chunk_word_size == med_word_size || chunk_word_size == small_word_size) && next_aligned > top()) { |
|
406 |
log_trace(gc, metaspace, freelist)("Creating padding chunks in %s between %p and %p...", |
|
407 |
(is_class() ? "class space " : "metaspace"), |
|
408 |
top(), next_aligned); |
|
409 |
allocate_padding_chunks_until_top_is_at(next_aligned); |
|
410 |
// Now, top should be aligned correctly. |
|
411 |
assert_is_aligned(top(), required_chunk_alignment); |
|
412 |
} |
|
413 |
||
414 |
// Now, top should be aligned correctly. |
|
415 |
assert_is_aligned(top(), required_chunk_alignment); |
|
416 |
||
417 |
// Bottom of the new chunk |
|
418 |
MetaWord* chunk_limit = top(); |
|
419 |
assert(chunk_limit != NULL, "Not safe to call this method"); |
|
420 |
||
421 |
// The virtual spaces are always expanded by the |
|
422 |
// commit granularity to enforce the following condition. |
|
423 |
// Without this the is_available check will not work correctly. |
|
424 |
assert(_virtual_space.committed_size() == _virtual_space.actual_committed_size(), |
|
425 |
"The committed memory doesn't match the expanded memory."); |
|
426 |
||
427 |
if (!is_available(chunk_word_size)) { |
|
50811
f533eb5e7430
8205664: Move detailed metaspace logging from debug to trace
pliden
parents:
50286
diff
changeset
|
428 |
LogTarget(Trace, gc, metaspace, freelist) lt; |
50193 | 429 |
if (lt.is_enabled()) { |
430 |
LogStream ls(lt); |
|
431 |
ls.print("VirtualSpaceNode::take_from_committed() not available " SIZE_FORMAT " words ", chunk_word_size); |
|
432 |
// Dump some information about the virtual space that is nearly full |
|
433 |
print_on(&ls); |
|
434 |
} |
|
435 |
return NULL; |
|
436 |
} |
|
437 |
||
438 |
// Take the space (bump top on the current virtual space). |
|
439 |
inc_top(chunk_word_size); |
|
440 |
||
441 |
// Initialize the chunk |
|
442 |
ChunkIndex chunk_type = get_chunk_type_by_size(chunk_word_size, is_class()); |
|
443 |
Metachunk* result = ::new (chunk_limit) Metachunk(chunk_type, is_class(), chunk_word_size, this); |
|
444 |
assert(result == (Metachunk*)chunk_limit, "Sanity"); |
|
445 |
occupancy_map()->set_chunk_starts_at_address((MetaWord*)result, true); |
|
446 |
do_update_in_use_info_for_chunk(result, true); |
|
447 |
||
448 |
inc_container_count(); |
|
449 |
||
53970 | 450 |
#ifdef ASSERT |
451 |
EVERY_NTH(VerifyMetaspaceInterval) |
|
452 |
chunk_manager->locked_verify(true); |
|
453 |
verify(true); |
|
454 |
END_EVERY_NTH |
|
455 |
do_verify_chunk(result); |
|
456 |
#endif |
|
50193 | 457 |
|
458 |
result->inc_use_count(); |
|
459 |
||
460 |
return result; |
|
461 |
} |
|
462 |
||
463 |
||
464 |
// Expand the virtual space (commit more of the reserved space) |
|
465 |
bool VirtualSpaceNode::expand_by(size_t min_words, size_t preferred_words) { |
|
466 |
size_t min_bytes = min_words * BytesPerWord; |
|
467 |
size_t preferred_bytes = preferred_words * BytesPerWord; |
|
468 |
||
469 |
size_t uncommitted = virtual_space()->reserved_size() - virtual_space()->actual_committed_size(); |
|
470 |
||
471 |
if (uncommitted < min_bytes) { |
|
472 |
return false; |
|
473 |
} |
|
474 |
||
475 |
size_t commit = MIN2(preferred_bytes, uncommitted); |
|
476 |
bool result = virtual_space()->expand_by(commit, false); |
|
477 |
||
478 |
if (result) { |
|
479 |
log_trace(gc, metaspace, freelist)("Expanded %s virtual space list node by " SIZE_FORMAT " words.", |
|
480 |
(is_class() ? "class" : "non-class"), commit); |
|
481 |
DEBUG_ONLY(Atomic::inc(&g_internal_statistics.num_committed_space_expanded)); |
|
482 |
} else { |
|
483 |
log_trace(gc, metaspace, freelist)("Failed to expand %s virtual space list node by " SIZE_FORMAT " words.", |
|
484 |
(is_class() ? "class" : "non-class"), commit); |
|
485 |
} |
|
486 |
||
487 |
assert(result, "Failed to commit memory"); |
|
488 |
||
489 |
return result; |
|
490 |
} |
|
491 |
||
492 |
Metachunk* VirtualSpaceNode::get_chunk_vs(size_t chunk_word_size) { |
|
493 |
assert_lock_strong(MetaspaceExpand_lock); |
|
494 |
Metachunk* result = take_from_committed(chunk_word_size); |
|
495 |
return result; |
|
496 |
} |
|
497 |
||
498 |
bool VirtualSpaceNode::initialize() { |
|
499 |
||
500 |
if (!_rs.is_reserved()) { |
|
501 |
return false; |
|
502 |
} |
|
503 |
||
504 |
// These are necessary restriction to make sure that the virtual space always |
|
505 |
// grows in steps of Metaspace::commit_alignment(). If both base and size are |
|
506 |
// aligned only the middle alignment of the VirtualSpace is used. |
|
507 |
assert_is_aligned(_rs.base(), Metaspace::commit_alignment()); |
|
508 |
assert_is_aligned(_rs.size(), Metaspace::commit_alignment()); |
|
509 |
||
510 |
// ReservedSpaces marked as special will have the entire memory |
|
511 |
// pre-committed. Setting a committed size will make sure that |
|
512 |
// committed_size and actual_committed_size agrees. |
|
513 |
size_t pre_committed_size = _rs.special() ? _rs.size() : 0; |
|
514 |
||
515 |
bool result = virtual_space()->initialize_with_granularity(_rs, pre_committed_size, |
|
516 |
Metaspace::commit_alignment()); |
|
517 |
if (result) { |
|
518 |
assert(virtual_space()->committed_size() == virtual_space()->actual_committed_size(), |
|
519 |
"Checking that the pre-committed memory was registered by the VirtualSpace"); |
|
520 |
||
521 |
set_top((MetaWord*)virtual_space()->low()); |
|
522 |
} |
|
523 |
||
524 |
// Initialize Occupancy Map. |
|
525 |
const size_t smallest_chunk_size = is_class() ? ClassSpecializedChunk : SpecializedChunk; |
|
526 |
_occupancy_map = new OccupancyMap(bottom(), reserved_words(), smallest_chunk_size); |
|
527 |
||
528 |
return result; |
|
529 |
} |
|
530 |
||
531 |
void VirtualSpaceNode::print_on(outputStream* st, size_t scale) const { |
|
532 |
size_t used_words = used_words_in_vs(); |
|
533 |
size_t commit_words = committed_words(); |
|
534 |
size_t res_words = reserved_words(); |
|
535 |
VirtualSpace* vs = virtual_space(); |
|
536 |
||
537 |
st->print("node @" PTR_FORMAT ": ", p2i(this)); |
|
538 |
st->print("reserved="); |
|
539 |
print_scaled_words(st, res_words, scale); |
|
540 |
st->print(", committed="); |
|
541 |
print_scaled_words_and_percentage(st, commit_words, res_words, scale); |
|
542 |
st->print(", used="); |
|
543 |
print_scaled_words_and_percentage(st, used_words, res_words, scale); |
|
544 |
st->cr(); |
|
545 |
st->print(" [" PTR_FORMAT ", " PTR_FORMAT ", " |
|
546 |
PTR_FORMAT ", " PTR_FORMAT ")", |
|
547 |
p2i(bottom()), p2i(top()), p2i(end()), |
|
548 |
p2i(vs->high_boundary())); |
|
549 |
} |
|
550 |
||
551 |
#ifdef ASSERT |
|
552 |
void VirtualSpaceNode::mangle() { |
|
553 |
size_t word_size = capacity_words_in_vs(); |
|
554 |
Copy::fill_to_words((HeapWord*) low(), word_size, 0xf1f1f1f1); |
|
555 |
} |
|
556 |
#endif // ASSERT |
|
557 |
||
558 |
void VirtualSpaceNode::retire(ChunkManager* chunk_manager) { |
|
53970 | 559 |
assert(is_class() == chunk_manager->is_class(), "Wrong ChunkManager?"); |
560 |
#ifdef ASSERT |
|
561 |
verify(false); |
|
562 |
EVERY_NTH(VerifyMetaspaceInterval) |
|
563 |
verify(true); |
|
564 |
END_EVERY_NTH |
|
565 |
#endif |
|
50193 | 566 |
for (int i = (int)MediumIndex; i >= (int)ZeroIndex; --i) { |
567 |
ChunkIndex index = (ChunkIndex)i; |
|
568 |
size_t chunk_size = chunk_manager->size_by_index(index); |
|
569 |
||
570 |
while (free_words_in_vs() >= chunk_size) { |
|
571 |
Metachunk* chunk = get_chunk_vs(chunk_size); |
|
572 |
// Chunk will be allocated aligned, so allocation may require |
|
573 |
// additional padding chunks. That may cause above allocation to |
|
574 |
// fail. Just ignore the failed allocation and continue with the |
|
575 |
// next smaller chunk size. As the VirtualSpaceNode comitted |
|
576 |
// size should be a multiple of the smallest chunk size, we |
|
577 |
// should always be able to fill the VirtualSpace completely. |
|
578 |
if (chunk == NULL) { |
|
579 |
break; |
|
580 |
} |
|
581 |
chunk_manager->return_single_chunk(chunk); |
|
582 |
} |
|
583 |
} |
|
584 |
assert(free_words_in_vs() == 0, "should be empty now"); |
|
585 |
} |
|
586 |
||
587 |
} // namespace metaspace |
|
588 |