Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 1 | /* |
| 2 | * Copyright (C) 2012 The Android Open Source Project |
| 3 | * |
| 4 | * Licensed under the Apache License, Version 2.0 (the "License"); |
| 5 | * you may not use this file except in compliance with the License. |
| 6 | * You may obtain a copy of the License at |
| 7 | * |
| 8 | * http://www.apache.org/licenses/LICENSE-2.0 |
| 9 | * |
| 10 | * Unless required by applicable law or agreed to in writing, software |
| 11 | * distributed under the License is distributed on an "AS IS" BASIS, |
| 12 | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| 13 | * See the License for the specific language governing permissions and |
| 14 | * limitations under the License. |
| 15 | */ |
| 16 | |
Brian Carlstrom | fc0e321 | 2013-07-17 14:40:12 -0700 | [diff] [blame] | 17 | #ifndef ART_RUNTIME_GC_ACCOUNTING_ATOMIC_STACK_H_ |
| 18 | #define ART_RUNTIME_GC_ACCOUNTING_ATOMIC_STACK_H_ |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 19 | |
Brian Carlstrom | a280655 | 2014-02-27 12:29:32 -0800 | [diff] [blame] | 20 | #include <algorithm> |
Ian Rogers | 700a402 | 2014-05-19 16:49:03 -0700 | [diff] [blame] | 21 | #include <memory> |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 22 | #include <string> |
| 23 | |
Ian Rogers | ef7d42f | 2014-01-06 12:55:46 -0800 | [diff] [blame] | 24 | #include "atomic.h" |
Elliott Hughes | 07ed66b | 2012-12-12 18:34:25 -0800 | [diff] [blame] | 25 | #include "base/logging.h" |
Elliott Hughes | 7616005 | 2012-12-12 16:31:20 -0800 | [diff] [blame] | 26 | #include "base/macros.h" |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 27 | #include "mem_map.h" |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 28 | #include "stack.h" |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 29 | #include "utils.h" |
| 30 | |
| 31 | namespace art { |
Ian Rogers | 1d54e73 | 2013-05-02 21:10:01 -0700 | [diff] [blame] | 32 | namespace gc { |
| 33 | namespace accounting { |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 34 | |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 35 | // Internal representation is StackReference<T>, so this only works with mirror::Object or it's |
| 36 | // subclasses. |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 37 | template <typename T> |
| 38 | class AtomicStack { |
| 39 | public: |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 40 | class ObjectComparator { |
| 41 | public: |
| 42 | // These two comparators are for std::binary_search. |
| 43 | bool operator()(const T* a, const StackReference<T>& b) const NO_THREAD_SAFETY_ANALYSIS { |
| 44 | return a < b.AsMirrorPtr(); |
| 45 | } |
| 46 | bool operator()(const StackReference<T>& a, const T* b) const NO_THREAD_SAFETY_ANALYSIS { |
| 47 | return a.AsMirrorPtr() < b; |
| 48 | } |
| 49 | // This comparator is for std::sort. |
| 50 | bool operator()(const StackReference<T>& a, const StackReference<T>& b) const |
| 51 | NO_THREAD_SAFETY_ANALYSIS { |
| 52 | return a.AsMirrorPtr() < b.AsMirrorPtr(); |
| 53 | } |
| 54 | }; |
| 55 | |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 56 | // Capacity is how many elements we can store in the stack. |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 57 | static AtomicStack* Create(const std::string& name, size_t growth_limit, size_t capacity) { |
| 58 | std::unique_ptr<AtomicStack> mark_stack(new AtomicStack(name, growth_limit, capacity)); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 59 | mark_stack->Init(); |
| 60 | return mark_stack.release(); |
| 61 | } |
| 62 | |
Ian Rogers | 1d54e73 | 2013-05-02 21:10:01 -0700 | [diff] [blame] | 63 | ~AtomicStack() {} |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 64 | |
| 65 | void Reset() { |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 66 | DCHECK(mem_map_.get() != nullptr); |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 67 | DCHECK(begin_ != nullptr); |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 68 | front_index_.StoreRelaxed(0); |
| 69 | back_index_.StoreRelaxed(0); |
Mathieu Chartier | f082d3c | 2013-07-29 17:04:07 -0700 | [diff] [blame] | 70 | debug_is_sorted_ = true; |
Ian Rogers | c5f1773 | 2014-06-05 20:48:42 -0700 | [diff] [blame] | 71 | mem_map_->MadviseDontNeedAndZero(); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 72 | } |
| 73 | |
| 74 | // Beware: Mixing atomic pushes and atomic pops will cause ABA problem. |
| 75 | |
| 76 | // Returns false if we overflowed the stack. |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 77 | bool AtomicPushBackIgnoreGrowthLimit(T* value) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 78 | return AtomicPushBackInternal(value, capacity_); |
| 79 | } |
| 80 | |
| 81 | // Returns false if we overflowed the stack. |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 82 | bool AtomicPushBack(T* value) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 83 | return AtomicPushBackInternal(value, growth_limit_); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 84 | } |
| 85 | |
Hiroshi Yamauchi | f5b0e20 | 2014-02-11 17:02:22 -0800 | [diff] [blame] | 86 | // Atomically bump the back index by the given number of |
| 87 | // slots. Returns false if we overflowed the stack. |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 88 | bool AtomicBumpBack(size_t num_slots, StackReference<T>** start_address, |
| 89 | StackReference<T>** end_address) |
| 90 | SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
Hiroshi Yamauchi | f5b0e20 | 2014-02-11 17:02:22 -0800 | [diff] [blame] | 91 | if (kIsDebugBuild) { |
| 92 | debug_is_sorted_ = false; |
| 93 | } |
| 94 | int32_t index; |
| 95 | int32_t new_index; |
| 96 | do { |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 97 | index = back_index_.LoadRelaxed(); |
Hiroshi Yamauchi | f5b0e20 | 2014-02-11 17:02:22 -0800 | [diff] [blame] | 98 | new_index = index + num_slots; |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 99 | if (UNLIKELY(static_cast<size_t>(new_index) >= growth_limit_)) { |
Hiroshi Yamauchi | f5b0e20 | 2014-02-11 17:02:22 -0800 | [diff] [blame] | 100 | // Stack overflow. |
| 101 | return false; |
| 102 | } |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 103 | } while (!back_index_.CompareExchangeWeakRelaxed(index, new_index)); |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 104 | *start_address = begin_ + index; |
| 105 | *end_address = begin_ + new_index; |
Hiroshi Yamauchi | f5b0e20 | 2014-02-11 17:02:22 -0800 | [diff] [blame] | 106 | if (kIsDebugBuild) { |
| 107 | // Sanity check that the memory is zero. |
| 108 | for (int32_t i = index; i < new_index; ++i) { |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 109 | DCHECK_EQ(begin_[i].AsMirrorPtr(), static_cast<T*>(nullptr)) |
Brian Carlstrom | a280655 | 2014-02-27 12:29:32 -0800 | [diff] [blame] | 110 | << "i=" << i << " index=" << index << " new_index=" << new_index; |
Hiroshi Yamauchi | f5b0e20 | 2014-02-11 17:02:22 -0800 | [diff] [blame] | 111 | } |
| 112 | } |
| 113 | return true; |
| 114 | } |
| 115 | |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 116 | void AssertAllZero() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
Hiroshi Yamauchi | f5b0e20 | 2014-02-11 17:02:22 -0800 | [diff] [blame] | 117 | if (kIsDebugBuild) { |
| 118 | for (size_t i = 0; i < capacity_; ++i) { |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 119 | DCHECK_EQ(begin_[i].AsMirrorPtr(), static_cast<T*>(nullptr)) << "i=" << i; |
Hiroshi Yamauchi | f5b0e20 | 2014-02-11 17:02:22 -0800 | [diff] [blame] | 120 | } |
| 121 | } |
| 122 | } |
| 123 | |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 124 | void PushBack(T* value) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
Mathieu Chartier | f082d3c | 2013-07-29 17:04:07 -0700 | [diff] [blame] | 125 | if (kIsDebugBuild) { |
| 126 | debug_is_sorted_ = false; |
| 127 | } |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 128 | const int32_t index = back_index_.LoadRelaxed(); |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 129 | DCHECK_LT(static_cast<size_t>(index), growth_limit_); |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 130 | back_index_.StoreRelaxed(index + 1); |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 131 | begin_[index].Assign(value); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 132 | } |
| 133 | |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 134 | T* PopBack() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 135 | DCHECK_GT(back_index_.LoadRelaxed(), front_index_.LoadRelaxed()); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 136 | // Decrement the back index non atomically. |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 137 | back_index_.StoreRelaxed(back_index_.LoadRelaxed() - 1); |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 138 | return begin_[back_index_.LoadRelaxed()].AsMirrorPtr(); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 139 | } |
| 140 | |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 141 | // Take an item from the front of the stack. |
| 142 | T PopFront() { |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 143 | int32_t index = front_index_.LoadRelaxed(); |
| 144 | DCHECK_LT(index, back_index_.LoadRelaxed()); |
| 145 | front_index_.StoreRelaxed(index + 1); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 146 | return begin_[index]; |
| 147 | } |
| 148 | |
Mathieu Chartier | 94c32c5 | 2013-08-09 11:14:04 -0700 | [diff] [blame] | 149 | // Pop a number of elements. |
| 150 | void PopBackCount(int32_t n) { |
| 151 | DCHECK_GE(Size(), static_cast<size_t>(n)); |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 152 | back_index_.FetchAndSubSequentiallyConsistent(n); |
Mathieu Chartier | 94c32c5 | 2013-08-09 11:14:04 -0700 | [diff] [blame] | 153 | } |
| 154 | |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 155 | bool IsEmpty() const { |
| 156 | return Size() == 0; |
| 157 | } |
| 158 | |
| 159 | size_t Size() const { |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 160 | DCHECK_LE(front_index_.LoadRelaxed(), back_index_.LoadRelaxed()); |
| 161 | return back_index_.LoadRelaxed() - front_index_.LoadRelaxed(); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 162 | } |
| 163 | |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 164 | StackReference<T>* Begin() const { |
| 165 | return begin_ + front_index_.LoadRelaxed(); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 166 | } |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 167 | StackReference<T>* End() const { |
| 168 | return begin_ + back_index_.LoadRelaxed(); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 169 | } |
| 170 | |
| 171 | size_t Capacity() const { |
| 172 | return capacity_; |
| 173 | } |
| 174 | |
| 175 | // Will clear the stack. |
| 176 | void Resize(size_t new_capacity) { |
| 177 | capacity_ = new_capacity; |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 178 | growth_limit_ = new_capacity; |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 179 | Init(); |
| 180 | } |
| 181 | |
Ian Rogers | 1d54e73 | 2013-05-02 21:10:01 -0700 | [diff] [blame] | 182 | void Sort() { |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 183 | int32_t start_back_index = back_index_.LoadRelaxed(); |
| 184 | int32_t start_front_index = front_index_.LoadRelaxed(); |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 185 | std::sort(Begin(), End(), ObjectComparator()); |
Ian Rogers | 3e5cf30 | 2014-05-20 16:40:37 -0700 | [diff] [blame] | 186 | CHECK_EQ(start_back_index, back_index_.LoadRelaxed()); |
| 187 | CHECK_EQ(start_front_index, front_index_.LoadRelaxed()); |
Mathieu Chartier | f082d3c | 2013-07-29 17:04:07 -0700 | [diff] [blame] | 188 | if (kIsDebugBuild) { |
| 189 | debug_is_sorted_ = true; |
Ian Rogers | 1d54e73 | 2013-05-02 21:10:01 -0700 | [diff] [blame] | 190 | } |
| 191 | } |
| 192 | |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 193 | bool ContainsSorted(const T* value) const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
Mathieu Chartier | f082d3c | 2013-07-29 17:04:07 -0700 | [diff] [blame] | 194 | DCHECK(debug_is_sorted_); |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 195 | return std::binary_search(Begin(), End(), value, ObjectComparator()); |
Mathieu Chartier | f082d3c | 2013-07-29 17:04:07 -0700 | [diff] [blame] | 196 | } |
| 197 | |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 198 | bool Contains(const T* value) const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
| 199 | for (auto cur = Begin(), end = End(); cur != end; ++cur) { |
| 200 | if (cur->AsMirrorPtr() == value) { |
| 201 | return true; |
| 202 | } |
| 203 | } |
| 204 | return false; |
Ian Rogers | 1d54e73 | 2013-05-02 21:10:01 -0700 | [diff] [blame] | 205 | } |
| 206 | |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 207 | private: |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 208 | AtomicStack(const std::string& name, size_t growth_limit, size_t capacity) |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 209 | : name_(name), |
| 210 | back_index_(0), |
| 211 | front_index_(0), |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 212 | begin_(nullptr), |
| 213 | growth_limit_(growth_limit), |
Ian Rogers | 1d54e73 | 2013-05-02 21:10:01 -0700 | [diff] [blame] | 214 | capacity_(capacity), |
Mathieu Chartier | f082d3c | 2013-07-29 17:04:07 -0700 | [diff] [blame] | 215 | debug_is_sorted_(true) { |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 216 | } |
| 217 | |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 218 | // Returns false if we overflowed the stack. |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 219 | bool AtomicPushBackInternal(T* value, size_t limit) ALWAYS_INLINE |
| 220 | SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 221 | if (kIsDebugBuild) { |
| 222 | debug_is_sorted_ = false; |
| 223 | } |
| 224 | int32_t index; |
| 225 | do { |
| 226 | index = back_index_.LoadRelaxed(); |
| 227 | if (UNLIKELY(static_cast<size_t>(index) >= limit)) { |
| 228 | // Stack overflow. |
| 229 | return false; |
| 230 | } |
| 231 | } while (!back_index_.CompareExchangeWeakRelaxed(index, index + 1)); |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 232 | begin_[index].Assign(value); |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 233 | return true; |
| 234 | } |
| 235 | |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 236 | // Size in number of elements. |
| 237 | void Init() { |
Ian Rogers | 8d31bbd | 2013-10-13 10:44:14 -0700 | [diff] [blame] | 238 | std::string error_msg; |
Vladimir Marko | 5c42c29 | 2015-02-25 12:02:49 +0000 | [diff] [blame] | 239 | mem_map_.reset(MemMap::MapAnonymous(name_.c_str(), nullptr, capacity_ * sizeof(begin_[0]), |
| 240 | PROT_READ | PROT_WRITE, false, false, &error_msg)); |
Ian Rogers | 8d31bbd | 2013-10-13 10:44:14 -0700 | [diff] [blame] | 241 | CHECK(mem_map_.get() != NULL) << "couldn't allocate mark stack.\n" << error_msg; |
Ian Rogers | 1373595 | 2014-10-08 12:43:28 -0700 | [diff] [blame] | 242 | uint8_t* addr = mem_map_->Begin(); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 243 | CHECK(addr != NULL); |
Mathieu Chartier | f082d3c | 2013-07-29 17:04:07 -0700 | [diff] [blame] | 244 | debug_is_sorted_ = true; |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 245 | begin_ = reinterpret_cast<StackReference<T>*>(addr); |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 246 | Reset(); |
| 247 | } |
| 248 | |
| 249 | // Name of the mark stack. |
| 250 | std::string name_; |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 251 | // Memory mapping of the atomic stack. |
Ian Rogers | 700a402 | 2014-05-19 16:49:03 -0700 | [diff] [blame] | 252 | std::unique_ptr<MemMap> mem_map_; |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 253 | // Back index (index after the last element pushed). |
| 254 | AtomicInteger back_index_; |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 255 | // Front index, used for implementing PopFront. |
| 256 | AtomicInteger front_index_; |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 257 | // Base of the atomic stack. |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 258 | StackReference<T>* begin_; |
Mathieu Chartier | c179016 | 2014-05-23 10:54:50 -0700 | [diff] [blame] | 259 | // Current maximum which we can push back to, must be <= capacity_. |
| 260 | size_t growth_limit_; |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 261 | // Maximum number of elements. |
| 262 | size_t capacity_; |
Mathieu Chartier | f082d3c | 2013-07-29 17:04:07 -0700 | [diff] [blame] | 263 | // Whether or not the stack is sorted, only updated in debug mode to avoid performance overhead. |
| 264 | bool debug_is_sorted_; |
Ian Rogers | 1d54e73 | 2013-05-02 21:10:01 -0700 | [diff] [blame] | 265 | |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 266 | DISALLOW_COPY_AND_ASSIGN(AtomicStack); |
| 267 | }; |
| 268 | |
Mathieu Chartier | cb535da | 2015-01-23 13:50:03 -0800 | [diff] [blame] | 269 | typedef AtomicStack<mirror::Object> ObjectStack; |
Ian Rogers | 2dd0e2c | 2013-01-24 12:42:14 -0800 | [diff] [blame] | 270 | |
Ian Rogers | 1d54e73 | 2013-05-02 21:10:01 -0700 | [diff] [blame] | 271 | } // namespace accounting |
| 272 | } // namespace gc |
Mathieu Chartier | d8195f1 | 2012-10-05 12:21:28 -0700 | [diff] [blame] | 273 | } // namespace art |
| 274 | |
Brian Carlstrom | fc0e321 | 2013-07-17 14:40:12 -0700 | [diff] [blame] | 275 | #endif // ART_RUNTIME_GC_ACCOUNTING_ATOMIC_STACK_H_ |