equal
deleted
inserted
replaced
79 } |
79 } |
80 // attempt_allocation_slow will also unlock the heap lock when appropriate. |
80 // attempt_allocation_slow will also unlock the heap lock when appropriate. |
81 return attempt_allocation_slow(word_size, permit_collection_pause); |
81 return attempt_allocation_slow(word_size, permit_collection_pause); |
82 } |
82 } |
83 |
83 |
84 inline RefToScanQueue* G1CollectedHeap::task_queue(int i) { |
84 inline RefToScanQueue* G1CollectedHeap::task_queue(int i) const { |
85 return _task_queues->queue(i); |
85 return _task_queues->queue(i); |
86 } |
86 } |
87 |
|
88 |
87 |
89 inline bool G1CollectedHeap::isMarkedPrev(oop obj) const { |
88 inline bool G1CollectedHeap::isMarkedPrev(oop obj) const { |
90 return _cm->prevMarkBitMap()->isMarked((HeapWord *)obj); |
89 return _cm->prevMarkBitMap()->isMarked((HeapWord *)obj); |
91 } |
90 } |
92 |
91 |