blob: 222bd9c8064d4362797c95ef27a4592b4ef18fb0 [file] [log] [blame]
Elliott Hughese6c57fc2014-05-23 20:06:03 -07001/*
2 * Copyright (C) 2014 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
Elliott Hughese6c57fc2014-05-23 20:06:03 -070017#include <stdatomic.h>
Hans Boehm00aaea32014-08-19 16:14:01 -070018#include <gtest/gtest.h>
19#include <pthread.h>
20#include <stdint.h>
Elliott Hughese6c57fc2014-05-23 20:06:03 -070021
22TEST(stdatomic, LOCK_FREE) {
23 ASSERT_TRUE(ATOMIC_BOOL_LOCK_FREE);
24 ASSERT_TRUE(ATOMIC_CHAR16_T_LOCK_FREE);
25 ASSERT_TRUE(ATOMIC_CHAR32_T_LOCK_FREE);
26 ASSERT_TRUE(ATOMIC_CHAR_LOCK_FREE);
27 ASSERT_TRUE(ATOMIC_INT_LOCK_FREE);
28 ASSERT_TRUE(ATOMIC_LLONG_LOCK_FREE);
29 ASSERT_TRUE(ATOMIC_LONG_LOCK_FREE);
30 ASSERT_TRUE(ATOMIC_POINTER_LOCK_FREE);
31 ASSERT_TRUE(ATOMIC_SHORT_LOCK_FREE);
32 ASSERT_TRUE(ATOMIC_WCHAR_T_LOCK_FREE);
33}
34
35TEST(stdatomic, init) {
36 atomic_int v = ATOMIC_VAR_INIT(123);
37 ASSERT_EQ(123, atomic_load(&v));
38
39 atomic_init(&v, 456);
40 ASSERT_EQ(456, atomic_load(&v));
41
42 atomic_flag f = ATOMIC_FLAG_INIT;
43 ASSERT_FALSE(atomic_flag_test_and_set(&f));
44}
45
46TEST(stdatomic, atomic_thread_fence) {
47 atomic_thread_fence(memory_order_relaxed);
48 atomic_thread_fence(memory_order_consume);
49 atomic_thread_fence(memory_order_acquire);
50 atomic_thread_fence(memory_order_release);
51 atomic_thread_fence(memory_order_acq_rel);
52 atomic_thread_fence(memory_order_seq_cst);
53}
54
55TEST(stdatomic, atomic_signal_fence) {
56 atomic_signal_fence(memory_order_relaxed);
57 atomic_signal_fence(memory_order_consume);
58 atomic_signal_fence(memory_order_acquire);
59 atomic_signal_fence(memory_order_release);
60 atomic_signal_fence(memory_order_acq_rel);
61 atomic_signal_fence(memory_order_seq_cst);
62}
63
64TEST(stdatomic, atomic_is_lock_free) {
65 atomic_char small;
66 atomic_intmax_t big;
67 ASSERT_TRUE(atomic_is_lock_free(&small));
Raghu Gandhamf1837372014-07-24 15:56:51 -070068 // atomic_intmax_t(size = 64) is not lock free on mips32.
69#if defined(__mips__) && !defined(__LP64__)
70 ASSERT_FALSE(atomic_is_lock_free(&big));
71#else
Elliott Hughese6c57fc2014-05-23 20:06:03 -070072 ASSERT_TRUE(atomic_is_lock_free(&big));
Raghu Gandhamf1837372014-07-24 15:56:51 -070073#endif
Elliott Hughese6c57fc2014-05-23 20:06:03 -070074}
75
76TEST(stdatomic, atomic_flag) {
77 atomic_flag f = ATOMIC_FLAG_INIT;
78 ASSERT_FALSE(atomic_flag_test_and_set(&f));
79 ASSERT_TRUE(atomic_flag_test_and_set(&f));
80
81 atomic_flag_clear(&f);
82
83 ASSERT_FALSE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed));
84 ASSERT_TRUE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed));
85
86 atomic_flag_clear_explicit(&f, memory_order_relaxed);
87 ASSERT_FALSE(atomic_flag_test_and_set_explicit(&f, memory_order_relaxed));
88}
89
90TEST(stdatomic, atomic_store) {
91 atomic_int i;
92 atomic_store(&i, 123);
93 ASSERT_EQ(123, atomic_load(&i));
94 atomic_store_explicit(&i, 123, memory_order_relaxed);
95 ASSERT_EQ(123, atomic_load_explicit(&i, memory_order_relaxed));
96}
97
98TEST(stdatomic, atomic_exchange) {
99 atomic_int i;
100 atomic_store(&i, 123);
101 ASSERT_EQ(123, atomic_exchange(&i, 456));
102 ASSERT_EQ(456, atomic_exchange_explicit(&i, 123, memory_order_relaxed));
103}
104
105TEST(stdatomic, atomic_compare_exchange) {
106 atomic_int i;
Dan Albert6b3beb22014-05-28 16:27:32 -0700107 int expected;
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700108
109 atomic_store(&i, 123);
Dan Albert6b3beb22014-05-28 16:27:32 -0700110 expected = 123;
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700111 ASSERT_TRUE(atomic_compare_exchange_strong(&i, &expected, 456));
112 ASSERT_FALSE(atomic_compare_exchange_strong(&i, &expected, 456));
Dan Albert6b3beb22014-05-28 16:27:32 -0700113 ASSERT_EQ(456, expected);
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700114
115 atomic_store(&i, 123);
Dan Albert6b3beb22014-05-28 16:27:32 -0700116 expected = 123;
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700117 ASSERT_TRUE(atomic_compare_exchange_strong_explicit(&i, &expected, 456, memory_order_relaxed, memory_order_relaxed));
118 ASSERT_FALSE(atomic_compare_exchange_strong_explicit(&i, &expected, 456, memory_order_relaxed, memory_order_relaxed));
Dan Albert6b3beb22014-05-28 16:27:32 -0700119 ASSERT_EQ(456, expected);
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700120
121 atomic_store(&i, 123);
Dan Albert6b3beb22014-05-28 16:27:32 -0700122 expected = 123;
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700123 ASSERT_TRUE(atomic_compare_exchange_weak(&i, &expected, 456));
124 ASSERT_FALSE(atomic_compare_exchange_weak(&i, &expected, 456));
Dan Albert6b3beb22014-05-28 16:27:32 -0700125 ASSERT_EQ(456, expected);
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700126
127 atomic_store(&i, 123);
Dan Albert6b3beb22014-05-28 16:27:32 -0700128 expected = 123;
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700129 ASSERT_TRUE(atomic_compare_exchange_weak_explicit(&i, &expected, 456, memory_order_relaxed, memory_order_relaxed));
130 ASSERT_FALSE(atomic_compare_exchange_weak_explicit(&i, &expected, 456, memory_order_relaxed, memory_order_relaxed));
Dan Albert6b3beb22014-05-28 16:27:32 -0700131 ASSERT_EQ(456, expected);
Elliott Hughese6c57fc2014-05-23 20:06:03 -0700132}
133
134TEST(stdatomic, atomic_fetch_add) {
135 atomic_int i = ATOMIC_VAR_INIT(123);
136 ASSERT_EQ(123, atomic_fetch_add(&i, 1));
137 ASSERT_EQ(124, atomic_fetch_add_explicit(&i, 1, memory_order_relaxed));
138 ASSERT_EQ(125, atomic_load(&i));
139}
140
141TEST(stdatomic, atomic_fetch_sub) {
142 atomic_int i = ATOMIC_VAR_INIT(123);
143 ASSERT_EQ(123, atomic_fetch_sub(&i, 1));
144 ASSERT_EQ(122, atomic_fetch_sub_explicit(&i, 1, memory_order_relaxed));
145 ASSERT_EQ(121, atomic_load(&i));
146}
147
148TEST(stdatomic, atomic_fetch_or) {
149 atomic_int i = ATOMIC_VAR_INIT(0x100);
150 ASSERT_EQ(0x100, atomic_fetch_or(&i, 0x020));
151 ASSERT_EQ(0x120, atomic_fetch_or_explicit(&i, 0x003, memory_order_relaxed));
152 ASSERT_EQ(0x123, atomic_load(&i));
153}
154
155TEST(stdatomic, atomic_fetch_xor) {
156 atomic_int i = ATOMIC_VAR_INIT(0x100);
157 ASSERT_EQ(0x100, atomic_fetch_xor(&i, 0x120));
158 ASSERT_EQ(0x020, atomic_fetch_xor_explicit(&i, 0x103, memory_order_relaxed));
159 ASSERT_EQ(0x123, atomic_load(&i));
160}
161
162TEST(stdatomic, atomic_fetch_and) {
163 atomic_int i = ATOMIC_VAR_INIT(0x123);
164 ASSERT_EQ(0x123, atomic_fetch_and(&i, 0x00f));
165 ASSERT_EQ(0x003, atomic_fetch_and_explicit(&i, 0x2, memory_order_relaxed));
166 ASSERT_EQ(0x002, atomic_load(&i));
167}
168
Hans Boehm00aaea32014-08-19 16:14:01 -0700169// And a rudimentary test of acquire-release memory ordering:
170
171constexpr static uint_least32_t BIG = 10000000ul; // Assumed even below.
172
173struct three_atomics {
174 atomic_uint_least32_t x;
175 char a[123]; // Everything in different cache lines,
176 // increase chance of compiler getting alignment wrong.
177 atomic_uint_least32_t y;
178 char b[4013];
179 atomic_uint_least32_t z;
180};
181
182// Very simple acquire/release memory ordering sanity check.
183static void* writer(void* arg) {
184 three_atomics* a = reinterpret_cast<three_atomics*>(arg);
185 for (uint_least32_t i = 0; i <= BIG; i+=2) {
186 atomic_store_explicit(&a->x, i, memory_order_relaxed);
187 atomic_store_explicit(&a->z, i, memory_order_relaxed);
188 atomic_store_explicit(&a->y, i, memory_order_release);
189 atomic_store_explicit(&a->x, i+1, memory_order_relaxed);
190 atomic_store_explicit(&a->z, i+1, memory_order_relaxed);
191 atomic_store_explicit(&a->y, i+1, memory_order_release);
192 }
193 return 0;
194}
195
196static void* reader(void* arg) {
197 three_atomics* a = reinterpret_cast<three_atomics*>(arg);
198 uint_least32_t xval = 0, yval = 0, zval = 0;
199 size_t repeat = 0;
200 size_t repeat_limit = 1000;
201 while (yval != BIG + 1) {
202 yval = atomic_load_explicit(&a->y, memory_order_acquire);
203 zval = atomic_load_explicit(&a->z, memory_order_relaxed);
204 xval = atomic_load_explicit(&a->x, memory_order_relaxed);
205 // If we see a given value of y, the immediately preceding
206 // stores to z and x, or later ones, should also be visible.
207 if (zval < yval) {
208 // Cant just ASSERT, since we are in a non-void function.
209 ADD_FAILURE() << "acquire-release ordering violation: "
210 << zval << " < " << yval << ", " << xval << "\n";
211 return 0; // Only report once.
212 }
213 if (xval < yval) {
214 // Cant just ASSERT, since we are in a non-void function.
215 ADD_FAILURE() << "acquire-release ordering violation: "
216 << xval << " < " << yval << ", " << zval << "\n";
217 return 0; // Only report once.
218 }
219 if (repeat < repeat_limit) ++repeat;
220 }
221 // The following assertion is not technically guaranteed to hold.
222 // But if it fails to hold, this test was useless, and we have a
223 // serious scheduling issue that we should probably know about.
224 EXPECT_EQ(repeat, repeat_limit);
225 return 0;
226}
227
228TEST(stdatomic, ordering) {
229 // Run a memory ordering sanity test.
230 void* result;
231 three_atomics a;
232 atomic_init(&a.x, 0ul);
233 atomic_init(&a.y, 0ul);
234 atomic_init(&a.z, 0ul);
235 pthread_t t1,t2;
236 ASSERT_EQ(0, pthread_create(&t1, 0, reader, &a));
237 ASSERT_EQ(0, pthread_create(&t2, 0, writer, &a));
238 ASSERT_EQ(0, pthread_join(t1, &result));
239 EXPECT_EQ(0, result);
240 ASSERT_EQ(0, pthread_join(t2, &result));
241 EXPECT_EQ(0, result);
242 EXPECT_EQ(atomic_load_explicit(&a.x, memory_order_consume), BIG + 1);
243 EXPECT_EQ(atomic_load_explicit(&a.y, memory_order_seq_cst), BIG + 1);
244 EXPECT_EQ(atomic_load(&a.z), BIG + 1);
245}