/* Copyright (c) 2006, Google Inc. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are * met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following disclaimer * in the documentation and/or other materials provided with the * distribution. * * Neither the name of Google Inc. nor the names of its * contributors may be used to endorse or promote products derived from * this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * * --- * Author: Sanjay Ghemawat */ #include #include "base/logging.h" #include "base/atomicops.h" #define GG_ULONGLONG(x) static_cast(x) template static void TestAtomicIncrement(AtomicType (*atomic_increment_func) (volatile AtomicType*, AtomicType)) { // For now, we just test single threaded execution // use a guard value to make sure the atomic_increment_func doesn't go // outside the expected address bounds. This is in particular to // test that some future change to the asm code doesn't cause the // 32-bit atomic_increment_func doesn't do the wrong thing on 64-bit // machines. struct { AtomicType prev_word; AtomicType count; AtomicType next_word; } s; AtomicType prev_word_value, next_word_value; memset(&prev_word_value, 0xFF, sizeof(AtomicType)); memset(&next_word_value, 0xEE, sizeof(AtomicType)); s.prev_word = prev_word_value; s.count = 0; s.next_word = next_word_value; ASSERT_EQ(1, (*atomic_increment_func)(&s.count, 1)); ASSERT_EQ(1, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); ASSERT_EQ(3, (*atomic_increment_func)(&s.count, 2)); ASSERT_EQ(3, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); ASSERT_EQ(6, (*atomic_increment_func)(&s.count, 3)); ASSERT_EQ(6, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); ASSERT_EQ(3, (*atomic_increment_func)(&s.count, -3)); ASSERT_EQ(3, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); ASSERT_EQ(1, (*atomic_increment_func)(&s.count, -2)); ASSERT_EQ(1, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); ASSERT_EQ(0, (*atomic_increment_func)(&s.count, -1)); ASSERT_EQ(0, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); ASSERT_EQ(-1, (*atomic_increment_func)(&s.count, -1)); ASSERT_EQ(-1, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); ASSERT_EQ(-5, (*atomic_increment_func)(&s.count, -4)); ASSERT_EQ(-5, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); ASSERT_EQ(0, (*atomic_increment_func)(&s.count, 5)); ASSERT_EQ(0, s.count); ASSERT_EQ(prev_word_value, s.prev_word); ASSERT_EQ(next_word_value, s.next_word); } #define NUM_BITS(T) (sizeof(T) * 8) template static void TestCompareAndSwap(AtomicType (*compare_and_swap_func) (volatile AtomicType*, AtomicType, AtomicType)) { AtomicType value = 0; AtomicType prev = (*compare_and_swap_func)(&value, 0, 1); ASSERT_EQ(1, value); ASSERT_EQ(0, prev); // Use test value that has non-zero bits in both halves, more for testing // 64-bit implementation on 32-bit platforms. const AtomicType k_test_val = (GG_ULONGLONG(1) << (NUM_BITS(AtomicType) - 2)) + 11; value = k_test_val; prev = (*compare_and_swap_func)(&value, 0, 5); ASSERT_EQ(k_test_val, value); ASSERT_EQ(k_test_val, prev); value = k_test_val; prev = (*compare_and_swap_func)(&value, k_test_val, 5); ASSERT_EQ(5, value); ASSERT_EQ(k_test_val, prev); } template static void TestAtomicExchange(AtomicType (*atomic_exchange_func) (volatile AtomicType*, AtomicType)) { AtomicType value = 0; AtomicType new_value = (*atomic_exchange_func)(&value, 1); ASSERT_EQ(1, value); ASSERT_EQ(0, new_value); // Use test value that has non-zero bits in both halves, more for testing // 64-bit implementation on 32-bit platforms. const AtomicType k_test_val = (GG_ULONGLONG(1) << (NUM_BITS(AtomicType) - 2)) + 11; value = k_test_val; new_value = (*atomic_exchange_func)(&value, k_test_val); ASSERT_EQ(k_test_val, value); ASSERT_EQ(k_test_val, new_value); value = k_test_val; new_value = (*atomic_exchange_func)(&value, 5); ASSERT_EQ(5, value); ASSERT_EQ(k_test_val, new_value); } template static void TestAtomicIncrementBounds(AtomicType (*atomic_increment_func) (volatile AtomicType*, AtomicType)) { // Test increment at the half-width boundary of the atomic type. // It is primarily for testing at the 32-bit boundary for 64-bit atomic type. AtomicType test_val = GG_ULONGLONG(1) << (NUM_BITS(AtomicType) / 2); AtomicType value = test_val - 1; AtomicType new_value = (*atomic_increment_func)(&value, 1); ASSERT_EQ(test_val, value); ASSERT_EQ(value, new_value); (*atomic_increment_func)(&value, -1); ASSERT_EQ(test_val - 1, value); } // This is a simple sanity check that values are correct. Not testing // atomicity template static void TestStore() { const AtomicType kVal1 = static_cast(0xa5a5a5a5a5a5a5a5LL); const AtomicType kVal2 = static_cast(-1); AtomicType value; base::subtle::NoBarrier_Store(&value, kVal1); ASSERT_EQ(kVal1, value); base::subtle::NoBarrier_Store(&value, kVal2); ASSERT_EQ(kVal2, value); base::subtle::Acquire_Store(&value, kVal1); ASSERT_EQ(kVal1, value); base::subtle::Acquire_Store(&value, kVal2); ASSERT_EQ(kVal2, value); base::subtle::Release_Store(&value, kVal1); ASSERT_EQ(kVal1, value); base::subtle::Release_Store(&value, kVal2); ASSERT_EQ(kVal2, value); } // This is a simple sanity check that values are correct. Not testing // atomicity template static void TestLoad() { const AtomicType kVal1 = static_cast(0xa5a5a5a5a5a5a5a5LL); const AtomicType kVal2 = static_cast(-1); AtomicType value; value = kVal1; ASSERT_EQ(kVal1, base::subtle::NoBarrier_Load(&value)); value = kVal2; ASSERT_EQ(kVal2, base::subtle::NoBarrier_Load(&value)); value = kVal1; ASSERT_EQ(kVal1, base::subtle::Acquire_Load(&value)); value = kVal2; ASSERT_EQ(kVal2, base::subtle::Acquire_Load(&value)); value = kVal1; ASSERT_EQ(kVal1, base::subtle::Release_Load(&value)); value = kVal2; ASSERT_EQ(kVal2, base::subtle::Release_Load(&value)); } template static void TestAtomicOps() { TestCompareAndSwap(base::subtle::NoBarrier_CompareAndSwap); TestCompareAndSwap(base::subtle::Acquire_CompareAndSwap); TestCompareAndSwap(base::subtle::Release_CompareAndSwap); TestAtomicExchange(base::subtle::NoBarrier_AtomicExchange); TestAtomicExchange(base::subtle::Acquire_AtomicExchange); TestAtomicExchange(base::subtle::Release_AtomicExchange); TestAtomicIncrementBounds( base::subtle::NoBarrier_AtomicIncrement); TestAtomicIncrementBounds( base::subtle::Barrier_AtomicIncrement); TestStore(); TestLoad(); } int main(int argc, char** argv) { TestAtomicIncrement(base::subtle::NoBarrier_AtomicIncrement); TestAtomicIncrement(base::subtle::Barrier_AtomicIncrement); TestAtomicIncrement(base::subtle::NoBarrier_AtomicIncrement); TestAtomicIncrement(base::subtle::Barrier_AtomicIncrement); TestAtomicOps(); TestAtomicOps(); // I've commented the Atomic64 tests out for now, because Atomic64 // doesn't work on x86 systems that are not compiled to support mmx // registers. Since I want this project to be as portable as // possible -- that is, not to assume we've compiled for mmx or even // that the processor supports it -- and we don't actually use // Atomic64 anywhere, I've commented it out of the test for now. // (Luckily, if we ever do use Atomic64 by accident, we'll get told // via a compiler error rather than some obscure runtime failure, so // this course of action is safe.) // If we ever *do* want to enable this, try adding -msse (or -mmmx?) // to the CXXFLAGS in Makefile.am. #if 0 and defined(BASE_HAS_ATOMIC64) TestAtomicIncrement( base::subtle::NoBarrier_AtomicIncrement); TestAtomicIncrement( base::subtle::Barrier_AtomicIncrement); #endif printf("PASS\n"); return 0; }