1 /* 2 * Copyright (c) 2005, 2015, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef SHARE_VM_GC_IMPLEMENTATION_PARALLELSCAVENGE_PSCOMPACTIONMANAGER_HPP 26 #define SHARE_VM_GC_IMPLEMENTATION_PARALLELSCAVENGE_PSCOMPACTIONMANAGER_HPP 27 28 #include "memory/allocation.hpp" 29 #include "utilities/stack.hpp" 30 #include "utilities/taskqueue.hpp" 31 32 // Move to some global location 33 #define HAS_BEEN_MOVED 0x1501d01d 34 // End move to some global location 35 36 37 class MutableSpace; 38 class PSOldGen; 39 class ParCompactionManager; 40 class ObjectStartArray; 41 class ParallelCompactData; 42 class ParMarkBitMap; 43 44 class ParCompactionManager : public CHeapObj<mtGC> { 45 friend class ParallelTaskTerminator; 46 friend class ParMarkBitMap; 47 friend class PSParallelCompact; 48 friend class StealRegionCompactionTask; 49 friend class UpdateAndFillClosure; 50 friend class RefProcTaskExecutor; 51 friend class IdleGCTask; 52 53 public: 54 55 // ------------------------ Don't putback if not needed 56 // Actions that the compaction manager should take. 57 enum Action { 58 Update, 59 Copy, 60 UpdateAndCopy, 61 CopyAndUpdate, 62 NotValid 63 }; 64 // ------------------------ End don't putback if not needed 65 66 private: 67 // 32-bit: 4K * 8 = 32KiB; 64-bit: 8K * 16 = 128KiB 68 #define QUEUE_SIZE (1 << NOT_LP64(12) LP64_ONLY(13)) 69 typedef OverflowTaskQueue<ObjArrayTask, mtGC, QUEUE_SIZE> ObjArrayTaskQueue; 70 typedef GenericTaskQueueSet<ObjArrayTaskQueue, mtGC> ObjArrayTaskQueueSet; 71 #undef QUEUE_SIZE 72 73 static ParCompactionManager** _manager_array; 74 static OopTaskQueueSet* _stack_array; 75 static ObjArrayTaskQueueSet* _objarray_queues; 76 static ObjectStartArray* _start_array; 77 static RegionTaskQueueSet* _region_array; 78 static PSOldGen* _old_gen; 79 80 private: 81 OverflowTaskQueue<oop, mtGC> _marking_stack; 82 ObjArrayTaskQueue _objarray_stack; 83 84 // Is there a way to reuse the _marking_stack for the 85 // saving empty regions? For now just create a different 86 // type of TaskQueue. 87 RegionTaskQueue* _region_stack; 88 89 static RegionTaskQueue** _region_list; 90 // Index in _region_list for current _region_stack. 91 uint _region_stack_index; 92 93 // Indexes of recycled region stacks/overflow stacks 94 // Stacks of regions to be compacted are embedded in the tasks doing 95 // the compaction. A thread that executes the task extracts the 96 // region stack and drains it. These threads keep these region 97 // stacks for use during compaction task stealing. If a thread 98 // gets a second draining task, it pushed its current region stack 99 // index into the array _recycled_stack_index and gets a new 100 // region stack from the task. A thread that is executing a 101 // compaction stealing task without ever having executing a 102 // draining task, will get a region stack from _recycled_stack_index. 103 // 104 // Array of indexes into the array of region stacks. 105 static uint* _recycled_stack_index; 106 // The index into _recycled_stack_index of the last region stack index 107 // pushed. If -1, there are no entries into _recycled_stack_index. 108 static int _recycled_top; 109 // The index into _recycled_stack_index of the last region stack index 110 // popped. If -1, there has not been any entry popped. 111 static int _recycled_bottom; 112 113 static ParMarkBitMap* _mark_bitmap; 114 115 Action _action; 116 117 static PSOldGen* old_gen() { return _old_gen; } 118 static ObjectStartArray* start_array() { return _start_array; } 119 static OopTaskQueueSet* stack_array() { return _stack_array; } 120 121 static void initialize(ParMarkBitMap* mbm); 122 123 protected: 124 // Array of tasks. Needed by the ParallelTaskTerminator. 125 static RegionTaskQueueSet* region_array() { return _region_array; } 126 OverflowTaskQueue<oop, mtGC>* marking_stack() { return &_marking_stack; } 127 128 // Pushes onto the marking stack. If the marking stack is full, 129 // pushes onto the overflow stack. 130 void stack_push(oop obj); 131 // Do not implement an equivalent stack_pop. Deal with the 132 // marking stack and overflow stack directly. 133 134 public: 135 Action action() { return _action; } 136 void set_action(Action v) { _action = v; } 137 138 RegionTaskQueue* region_stack() { return _region_stack; } 139 void set_region_stack(RegionTaskQueue* v) { _region_stack = v; } 140 141 inline static ParCompactionManager* manager_array(int index); 142 143 inline static RegionTaskQueue* region_list(int index) { 144 return _region_list[index]; 145 } 146 147 uint region_stack_index() { return _region_stack_index; } 148 void set_region_stack_index(uint v) { _region_stack_index = v; } 149 150 // Pop and push unique reusable stack index 151 static int pop_recycled_stack_index(); 152 static void push_recycled_stack_index(uint v); 153 static void reset_recycled_stack_index() { 154 _recycled_bottom = _recycled_top = -1; 155 } 156 157 ParCompactionManager(); 158 ~ParCompactionManager(); 159 160 // Pushes onto the region stack at the given index. If the 161 // region stack is full, 162 // pushes onto the region overflow stack. 163 static void region_list_push(uint stack_index, size_t region_index); 164 static void verify_region_list_empty(uint stack_index); 165 ParMarkBitMap* mark_bitmap() { return _mark_bitmap; } 166 167 // void drain_stacks(); 168 169 bool should_update(); 170 bool should_copy(); 171 172 // Save for later processing. Must not fail. 173 inline void push(oop obj); 174 inline void push_objarray(oop objarray, size_t index); 175 inline void push_region(size_t index); 176 177 template <typename T> 178 inline void mark_and_push(T* p); 179 180 inline void follow_klass(Klass* klass); 181 182 // Access function for compaction managers 183 static ParCompactionManager* gc_thread_compaction_manager(int index); 184 185 static bool steal(int queue_num, int* seed, oop& t) { 186 return stack_array()->steal(queue_num, seed, t); 187 } 188 189 static bool steal_objarray(int queue_num, int* seed, ObjArrayTask& t) { 190 return _objarray_queues->steal(queue_num, seed, t); 191 } 192 193 static bool steal(int queue_num, int* seed, size_t& region) { 194 return region_array()->steal(queue_num, seed, region); 195 } 196 197 // Process tasks remaining on any marking stack 198 void follow_marking_stacks(); 199 inline bool marking_stacks_empty() const; 200 201 // Process tasks remaining on any stack 202 void drain_region_stacks(); 203 204 void follow_contents(oop obj); 205 void follow_contents(objArrayOop array, int index); 206 207 void update_contents(oop obj); 208 209 class MarkAndPushClosure: public ExtendedOopClosure { 210 private: 211 ParCompactionManager* _compaction_manager; 212 public: 213 MarkAndPushClosure(ParCompactionManager* cm) : _compaction_manager(cm) { } 214 215 template <typename T> void do_oop_nv(T* p); 216 virtual void do_oop(oop* p); 217 virtual void do_oop(narrowOop* p); 218 219 // This closure provides its own oop verification code. 220 debug_only(virtual bool should_verify_oops() { return false; }) 221 }; 222 223 class FollowStackClosure: public VoidClosure { 224 private: 225 ParCompactionManager* _compaction_manager; 226 public: 227 FollowStackClosure(ParCompactionManager* cm) : _compaction_manager(cm) { } 228 virtual void do_void(); 229 }; 230 231 // The one and only place to start following the classes. 232 // Should only be applied to the ClassLoaderData klasses list. 233 class FollowKlassClosure : public KlassClosure { 234 private: 235 MarkAndPushClosure* _mark_and_push_closure; 236 public: 237 FollowKlassClosure(MarkAndPushClosure* mark_and_push_closure) : 238 _mark_and_push_closure(mark_and_push_closure) { } 239 void do_klass(Klass* klass); 240 }; 241 }; 242 243 inline ParCompactionManager* ParCompactionManager::manager_array(int index) { 244 assert(_manager_array != NULL, "access of NULL manager_array"); 245 assert(index >= 0 && index <= (int)ParallelGCThreads, 246 "out of range manager_array access"); 247 return _manager_array[index]; 248 } 249 250 bool ParCompactionManager::marking_stacks_empty() const { 251 return _marking_stack.is_empty() && _objarray_stack.is_empty(); 252 } 253 254 #endif // SHARE_VM_GC_IMPLEMENTATION_PARALLELSCAVENGE_PSCOMPACTIONMANAGER_HPP