index

src/share/vm/gc_implementation/g1/g1CollectedHeap.inline.hpp

Print this page
rev 8024 : imported patch event1
* * *
imported patch event2

*** 130,140 **** return r != NULL && r->in_collection_set(); } inline HeapWord* G1CollectedHeap::attempt_allocation(size_t word_size, uint* gc_count_before_ret, ! uint* gclocker_retry_count_ret) { assert_heap_not_locked_and_not_at_safepoint(); assert(!is_humongous(word_size), "attempt_allocation() should not " "be called for humongous allocation requests"); AllocationContext_t context = AllocationContext::current(); --- 130,141 ---- return r != NULL && r->in_collection_set(); } inline HeapWord* G1CollectedHeap::attempt_allocation(size_t word_size, uint* gc_count_before_ret, ! uint* gclocker_retry_count_ret, ! uint* gc_attempt) { assert_heap_not_locked_and_not_at_safepoint(); assert(!is_humongous(word_size), "attempt_allocation() should not " "be called for humongous allocation requests"); AllocationContext_t context = AllocationContext::current();
*** 142,152 **** false /* bot_updates */); if (result == NULL) { result = attempt_allocation_slow(word_size, context, gc_count_before_ret, ! gclocker_retry_count_ret); } assert_heap_not_locked(); if (result != NULL) { dirty_young_block(result, word_size); } --- 143,154 ---- false /* bot_updates */); if (result == NULL) { result = attempt_allocation_slow(word_size, context, gc_count_before_ret, ! gclocker_retry_count_ret, ! gc_attempt); } assert_heap_not_locked(); if (result != NULL) { dirty_young_block(result, word_size); }
index