1 /* 2 * Copyright (c) 2018, 2019, Oracle and/or its affiliates. All rights reserved. 3 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER. 4 * 5 * This code is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 only, as 7 * published by the Free Software Foundation. 8 * 9 * This code is distributed in the hope that it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License 12 * version 2 for more details (a copy is included in the LICENSE file that 13 * accompanied this code). 14 * 15 * You should have received a copy of the GNU General Public License version 16 * 2 along with this work; if not, write to the Free Software Foundation, 17 * Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA. 18 * 19 * Please contact Oracle, 500 Oracle Parkway, Redwood Shores, CA 94065 USA 20 * or visit www.oracle.com if you need additional information or have any 21 * questions. 22 * 23 */ 24 25 #ifndef SHARE_UTILITIES_CONCURRENTHASHTABLETASKS_INLINE_HPP 26 #define SHARE_UTILITIES_CONCURRENTHASHTABLETASKS_INLINE_HPP 27 28 #include "utilities/globalDefinitions.hpp" 29 #include "utilities/concurrentHashTable.inline.hpp" 30 31 // This inline file contains BulkDeleteTask and GrowTasks which are both bucket 32 // operations, which they are serialized with each other. 33 34 // Base class for pause and/or parallel bulk operations. 35 template <typename CONFIG, MEMFLAGS F> 36 class ConcurrentHashTable<CONFIG, F>::BucketsOperation { 37 protected: 38 ConcurrentHashTable<CONFIG, F>* _cht; 39 40 // Default size of _task_size_log2 41 static const size_t DEFAULT_TASK_SIZE_LOG2 = 12; 42 43 // The table is split into ranges, every increment is one range. 44 volatile size_t _next_to_claim; 45 size_t _task_size_log2; // Number of buckets. 46 size_t _stop_task; // Last task 47 size_t _size_log2; // Table size. 48 bool _is_mt; 49 50 BucketsOperation(ConcurrentHashTable<CONFIG, F>* cht, bool is_mt = false) 51 : _cht(cht), _next_to_claim(0), _task_size_log2(DEFAULT_TASK_SIZE_LOG2), 52 _stop_task(0), _size_log2(0), _is_mt(is_mt) {} 53 54 // Returns true if you succeeded to claim the range start -> (stop-1). 55 bool claim(size_t* start, size_t* stop) { 56 size_t claimed = Atomic::add((size_t)1, &_next_to_claim) - 1; 57 if (claimed >= _stop_task) { 58 return false; 59 } 60 *start = claimed * (((size_t)1) << _task_size_log2); 61 *stop = ((*start) + (((size_t)1) << _task_size_log2)); 62 return true; 63 } 64 65 // Calculate starting values. 66 void setup(Thread* thread) { 67 thread_owns_resize_lock(thread); 68 _size_log2 = _cht->_table->_log2_size; 69 _task_size_log2 = MIN2(_task_size_log2, _size_log2); 70 size_t tmp = _size_log2 > _task_size_log2 ? 71 _size_log2 - _task_size_log2 : 0; 72 _stop_task = (((size_t)1) << tmp); 73 } 74 75 // Returns false if all ranges are claimed. 76 bool have_more_work() { 77 return Atomic::load_acquire(&_next_to_claim) >= _stop_task; 78 } 79 80 void thread_owns_resize_lock(Thread* thread) { 81 assert(BucketsOperation::_cht->_resize_lock_owner == thread, 82 "Should be locked by me"); 83 assert(BucketsOperation::_cht->_resize_lock->owned_by_self(), 84 "Operations lock not held"); 85 } 86 void thread_owns_only_state_lock(Thread* thread) { 87 assert(BucketsOperation::_cht->_resize_lock_owner == thread, 88 "Should be locked by me"); 89 assert(!BucketsOperation::_cht->_resize_lock->owned_by_self(), 90 "Operations lock held"); 91 } 92 void thread_do_not_own_resize_lock(Thread* thread) { 93 assert(!BucketsOperation::_cht->_resize_lock->owned_by_self(), 94 "Operations lock held"); 95 assert(BucketsOperation::_cht->_resize_lock_owner != thread, 96 "Should not be locked by me"); 97 } 98 99 public: 100 // Pauses for safepoint 101 void pause(Thread* thread) { 102 // This leaves internal state locked. 103 this->thread_owns_resize_lock(thread); 104 BucketsOperation::_cht->_resize_lock->unlock(); 105 this->thread_owns_only_state_lock(thread); 106 } 107 108 // Continues after safepoint. 109 void cont(Thread* thread) { 110 this->thread_owns_only_state_lock(thread); 111 // If someone slips in here directly after safepoint. 112 while (!BucketsOperation::_cht->_resize_lock->try_lock()) 113 { /* for ever */ }; 114 this->thread_owns_resize_lock(thread); 115 } 116 }; 117 118 // For doing pausable/parallel bulk delete. 119 template <typename CONFIG, MEMFLAGS F> 120 class ConcurrentHashTable<CONFIG, F>::BulkDeleteTask : 121 public BucketsOperation 122 { 123 public: 124 BulkDeleteTask(ConcurrentHashTable<CONFIG, F>* cht, bool is_mt = false) 125 : BucketsOperation(cht, is_mt) { 126 } 127 // Before start prepare must be called. 128 bool prepare(Thread* thread) { 129 bool lock = BucketsOperation::_cht->try_resize_lock(thread); 130 if (!lock) { 131 return false; 132 } 133 this->setup(thread); 134 return true; 135 } 136 137 // Does one range destroying all matching EVALUATE_FUNC and 138 // DELETE_FUNC is called be destruction. Returns true if there is more work. 139 template <typename EVALUATE_FUNC, typename DELETE_FUNC> 140 bool do_task(Thread* thread, EVALUATE_FUNC& eval_f, DELETE_FUNC& del_f) { 141 size_t start, stop; 142 assert(BucketsOperation::_cht->_resize_lock_owner != NULL, 143 "Should be locked"); 144 if (!this->claim(&start, &stop)) { 145 return false; 146 } 147 BucketsOperation::_cht->do_bulk_delete_locked_for(thread, start, stop, 148 eval_f, del_f, 149 BucketsOperation::_is_mt); 150 assert(BucketsOperation::_cht->_resize_lock_owner != NULL, 151 "Should be locked"); 152 return true; 153 } 154 155 // Must be called after ranges are done. 156 void done(Thread* thread) { 157 this->thread_owns_resize_lock(thread); 158 BucketsOperation::_cht->unlock_resize_lock(thread); 159 this->thread_do_not_own_resize_lock(thread); 160 } 161 }; 162 163 template <typename CONFIG, MEMFLAGS F> 164 class ConcurrentHashTable<CONFIG, F>::GrowTask : 165 public BucketsOperation 166 { 167 public: 168 GrowTask(ConcurrentHashTable<CONFIG, F>* cht) : BucketsOperation(cht) { 169 } 170 // Before start prepare must be called. 171 bool prepare(Thread* thread) { 172 if (!BucketsOperation::_cht->internal_grow_prolog( 173 thread, BucketsOperation::_cht->_log2_size_limit)) { 174 return false; 175 } 176 this->setup(thread); 177 return true; 178 } 179 180 // Re-sizes a portion of the table. Returns true if there is more work. 181 bool do_task(Thread* thread) { 182 size_t start, stop; 183 assert(BucketsOperation::_cht->_resize_lock_owner != NULL, 184 "Should be locked"); 185 if (!this->claim(&start, &stop)) { 186 return false; 187 } 188 BucketsOperation::_cht->internal_grow_range(thread, start, stop); 189 assert(BucketsOperation::_cht->_resize_lock_owner != NULL, 190 "Should be locked"); 191 return true; 192 } 193 194 // Must be called after do_task returns false. 195 void done(Thread* thread) { 196 this->thread_owns_resize_lock(thread); 197 BucketsOperation::_cht->internal_grow_epilog(thread); 198 this->thread_do_not_own_resize_lock(thread); 199 } 200 }; 201 202 #endif // SHARE_UTILITIES_CONCURRENTHASHTABLETASKS_INLINE_HPP