< prev index next >

src/share/vm/gc/cms/compactibleFreeListSpace.cpp

Print this page

        

*** 62,72 **** // Set CMS global values assert(MinChunkSize == 0, "already set"); // MinChunkSize should be a multiple of MinObjAlignment and be large enough // for chunks to contain a FreeChunk. ! size_t min_chunk_size_in_bytes = align_size_up(sizeof(FreeChunk), MinObjAlignmentInBytes); MinChunkSize = min_chunk_size_in_bytes / BytesPerWord; assert(IndexSetStart == 0 && IndexSetStride == 0, "already set"); IndexSetStart = MinChunkSize; IndexSetStride = MinObjAlignment; --- 62,72 ---- // Set CMS global values assert(MinChunkSize == 0, "already set"); // MinChunkSize should be a multiple of MinObjAlignment and be large enough // for chunks to contain a FreeChunk. ! size_t min_chunk_size_in_bytes = align_up(sizeof(FreeChunk), MinObjAlignmentInBytes); MinChunkSize = min_chunk_size_in_bytes / BytesPerWord; assert(IndexSetStart == 0 && IndexSetStride == 0, "already set"); IndexSetStart = MinChunkSize; IndexSetStride = MinObjAlignment;
*** 2871,2882 **** MemRegion span = _old_gen->reserved(); if (low != NULL) { if (span.contains(low)) { // Align low down to a card boundary so that // we can use block_offset_careful() on span boundaries. ! HeapWord* aligned_low = (HeapWord*)align_size_down((uintptr_t)low, ! CardTableModRefBS::card_size); // Clip span prefix at aligned_low span = span.intersection(MemRegion(aligned_low, span.end())); } else if (low > span.end()) { span = MemRegion(low, low); // Null region } // else use entire span --- 2871,2881 ---- MemRegion span = _old_gen->reserved(); if (low != NULL) { if (span.contains(low)) { // Align low down to a card boundary so that // we can use block_offset_careful() on span boundaries. ! HeapWord* aligned_low = align_down(low, CardTableModRefBS::card_size); // Clip span prefix at aligned_low span = span.intersection(MemRegion(aligned_low, span.end())); } else if (low > span.end()) { span = MemRegion(low, low); // Null region } // else use entire span
< prev index next >