src/share/vm/gc_implementation/concurrentMarkSweep/compactibleFreeListSpace.hpp

Print this page

        

@@ -229,20 +229,20 @@
   // Split a free chunk into a smaller free chunk of size "new_size".
   // Return the smaller free chunk and return the remainder to the
   // free lists.
   FreeChunk* splitChunkAndReturnRemainder(FreeChunk* chunk, size_t new_size);
   // Add a chunk to the free lists.
-  void       addChunkToFreeLists(HeapWord* chunk, size_t size);
+  void       addChunkToFreeLists(HeapWord* chunk, size_t size, bool deallocate_pages);
   // Add a chunk to the free lists, preferring to suffix it
   // to the last free chunk at end of space if possible, and
   // updating the block census stats as well as block offset table.
   // Take any locks as appropriate if we are multithreaded.
   void       addChunkToFreeListsAtEndRecordingStats(HeapWord* chunk, size_t size);
   // Add a free chunk to the indexed free lists.
-  void       returnChunkToFreeList(FreeChunk* chunk);
+  void       returnChunkToFreeList(FreeChunk* chunk, bool deallocate_pages);
   // Add a free chunk to the dictionary.
-  void       returnChunkToDictionary(FreeChunk* chunk);
+  void       returnChunkToDictionary(FreeChunk* chunk, bool deallocate_pages);
 
   // Functions for maintaining the linear allocation buffers (LinAB).
   // Repairing a linear allocation block refers to operations
   // performed on the remainder of a LinAB after an allocation
   // has been made from it.

@@ -474,11 +474,11 @@
   }
   // Minimum size of a free block.
   virtual size_t minimum_free_block_size() const { return MinChunkSize; }
   void      removeFreeChunkFromFreeLists(FreeChunk* chunk);
   void      addChunkAndRepairOffsetTable(HeapWord* chunk, size_t size,
-              bool coalesced);
+              bool coalesced, bool deallocate_pages);
 
   // Support for decisions regarding concurrent collection policy
   bool should_concurrent_collect() const;
 
   // Support for compaction