- 64db19b773134c6c8004e3b23c9ca892efbf8bae Move SpinLock's adaptive spin count computation from a st... by Derek Mauro <dmauro@google.com>
  - 6f9533fb44a52485a7c2bbb9b4efc7bf8d6c359a Import of CCTZ from GitHub. by Abseil Team <absl-team@google.com>
  - a211d7255c986e8dd4ceada362c0d054a6a1969a Cleanup exception flags by Abseil Team <absl-team@google.com>
  - babdb29c590126fe9bba5229fe91034b5b5c358a Release time benchmarks. by Alex Strelnikov <strel@google.com>
  - 5803b32a3ff123d1fb57a0c471d199c818357c9f Release memutil microbenchmarks. by Alex Strelnikov <strel@google.com>
  - 5357d4890d30e80c53beb05af32500fb20e9402b Add parens around expansion of ABSL_PREDICT_{FALSE,TRUE} ... by Abseil Team <absl-team@google.com>
  - 32023f61a239a5f6b1c59e577bfe81b179bbcd2d Reformat build rule tag. by Alex Strelnikov <strel@google.com>
  - 833758ecf2b0cf7a42bbd50b5b127e416425c168 Release uint128 microbenchmarks. by Alex Strelnikov <strel@google.com>
  - c115a9bca1f944b90fdc78a56b2de176466b124f Disambiguate bitwise-not of size_type by Abseil Team <absl-team@google.com>
  - f6905f5b5f6e425792de646edafde440548d9346 Updates ConstructorTracker and TrackedObjects with 1) a m... by Abseil Team <absl-team@google.com>
  - 147c553bdd5d2db20a38f75c4d1ef973d6c709c5 Changes the absl::Duration factory functions to disallow ... by Greg Miller <jgm@google.com>
  - dba2b96d11b5264546b283ba452f2de1303b0f07 White space fix by Alex Strelnikov <strel@google.com>
GitOrigin-RevId: abacaab4b11a69dd4db627bd183571d7cabb8def
Change-Id: I6fa34f20d0b2f898e7b5475a603111413bb80a67
		
	
			
		
			
				
	
	
		
			165 lines
		
	
	
	
		
			6.1 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
			
		
		
	
	
			165 lines
		
	
	
	
		
			6.1 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
| //
 | |
| // Copyright 2017 The Abseil Authors.
 | |
| //
 | |
| // Licensed under the Apache License, Version 2.0 (the "License");
 | |
| // you may not use this file except in compliance with the License.
 | |
| // You may obtain a copy of the License at
 | |
| //
 | |
| //      http://www.apache.org/licenses/LICENSE-2.0
 | |
| //
 | |
| // Unless required by applicable law or agreed to in writing, software
 | |
| // distributed under the License is distributed on an "AS IS" BASIS,
 | |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| // See the License for the specific language governing permissions and
 | |
| // limitations under the License.
 | |
| //
 | |
| // -----------------------------------------------------------------------------
 | |
| // File: optimization.h
 | |
| // -----------------------------------------------------------------------------
 | |
| //
 | |
| // This header file defines portable macros for performance optimization.
 | |
| 
 | |
| #ifndef ABSL_BASE_OPTIMIZATION_H_
 | |
| #define ABSL_BASE_OPTIMIZATION_H_
 | |
| 
 | |
| #include "absl/base/config.h"
 | |
| 
 | |
| // ABSL_BLOCK_TAIL_CALL_OPTIMIZATION
 | |
| //
 | |
| // Instructs the compiler to avoid optimizing tail-call recursion. Use of this
 | |
| // macro is useful when you wish to preserve the existing function order within
 | |
| // a stack trace for logging, debugging, or profiling purposes.
 | |
| //
 | |
| // Example:
 | |
| //
 | |
| //   int f() {
 | |
| //     int result = g();
 | |
| //     ABSL_BLOCK_TAIL_CALL_OPTIMIZATION();
 | |
| //     return result;
 | |
| //   }
 | |
| #if defined(__pnacl__)
 | |
| #define ABSL_BLOCK_TAIL_CALL_OPTIMIZATION() if (volatile int x = 0) { (void)x; }
 | |
| #elif defined(__clang__)
 | |
| // Clang will not tail call given inline volatile assembly.
 | |
| #define ABSL_BLOCK_TAIL_CALL_OPTIMIZATION() __asm__ __volatile__("")
 | |
| #elif defined(__GNUC__)
 | |
| // GCC will not tail call given inline volatile assembly.
 | |
| #define ABSL_BLOCK_TAIL_CALL_OPTIMIZATION() __asm__ __volatile__("")
 | |
| #elif defined(_MSC_VER)
 | |
| #include <intrin.h>
 | |
| // The __nop() intrinsic blocks the optimisation.
 | |
| #define ABSL_BLOCK_TAIL_CALL_OPTIMIZATION() __nop()
 | |
| #else
 | |
| #define ABSL_BLOCK_TAIL_CALL_OPTIMIZATION() if (volatile int x = 0) { (void)x; }
 | |
| #endif
 | |
| 
 | |
| // ABSL_CACHELINE_SIZE
 | |
| //
 | |
| // Explicitly defines the size of the L1 cache for purposes of alignment.
 | |
| // Setting the cacheline size allows you to specify that certain objects be
 | |
| // aligned on a cacheline boundary with `ABSL_CACHELINE_ALIGNED` declarations.
 | |
| // (See below.)
 | |
| //
 | |
| // NOTE: this macro should be replaced with the following C++17 features, when
 | |
| // those are generally available:
 | |
| //
 | |
| //   * `std::hardware_constructive_interference_size`
 | |
| //   * `std::hardware_destructive_interference_size`
 | |
| //
 | |
| // See http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2016/p0154r1.html
 | |
| // for more information.
 | |
| #if defined(__GNUC__)
 | |
| // Cache line alignment
 | |
| #if defined(__i386__) || defined(__x86_64__)
 | |
| #define ABSL_CACHELINE_SIZE 64
 | |
| #elif defined(__powerpc64__)
 | |
| #define ABSL_CACHELINE_SIZE 128
 | |
| #elif defined(__aarch64__)
 | |
| // We would need to read special register ctr_el0 to find out L1 dcache size.
 | |
| // This value is a good estimate based on a real aarch64 machine.
 | |
| #define ABSL_CACHELINE_SIZE 64
 | |
| #elif defined(__arm__)
 | |
| // Cache line sizes for ARM: These values are not strictly correct since
 | |
| // cache line sizes depend on implementations, not architectures.  There
 | |
| // are even implementations with cache line sizes configurable at boot
 | |
| // time.
 | |
| #if defined(__ARM_ARCH_5T__)
 | |
| #define ABSL_CACHELINE_SIZE 32
 | |
| #elif defined(__ARM_ARCH_7A__)
 | |
| #define ABSL_CACHELINE_SIZE 64
 | |
| #endif
 | |
| #endif
 | |
| 
 | |
| #ifndef ABSL_CACHELINE_SIZE
 | |
| // A reasonable default guess.  Note that overestimates tend to waste more
 | |
| // space, while underestimates tend to waste more time.
 | |
| #define ABSL_CACHELINE_SIZE 64
 | |
| #endif
 | |
| 
 | |
| // ABSL_CACHELINE_ALIGNED
 | |
| //
 | |
| // Indicates that the declared object be cache aligned using
 | |
| // `ABSL_CACHELINE_SIZE` (see above). Cacheline aligning objects allows you to
 | |
| // load a set of related objects in the L1 cache for performance improvements.
 | |
| // Cacheline aligning objects properly allows constructive memory sharing and
 | |
| // prevents destructive (or "false") memory sharing.
 | |
| //
 | |
| // NOTE: this macro should be replaced with usage of `alignas()` using
 | |
| // `std::hardware_constructive_interference_size` and/or
 | |
| // `std::hardware_destructive_interference_size` when available within C++17.
 | |
| //
 | |
| // See http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2016/p0154r1.html
 | |
| // for more information.
 | |
| //
 | |
| // On some compilers, `ABSL_CACHELINE_ALIGNED` expands to
 | |
| // `__attribute__((aligned(ABSL_CACHELINE_SIZE)))`. For compilers where this is
 | |
| // not known to work, the macro expands to nothing.
 | |
| //
 | |
| // No further guarantees are made here. The result of applying the macro
 | |
| // to variables and types is always implementation-defined.
 | |
| //
 | |
| // WARNING: It is easy to use this attribute incorrectly, even to the point
 | |
| // of causing bugs that are difficult to diagnose, crash, etc. It does not
 | |
| // of itself guarantee that objects are aligned to a cache line.
 | |
| //
 | |
| // Recommendations:
 | |
| //
 | |
| // 1) Consult compiler documentation; this comment is not kept in sync as
 | |
| //    toolchains evolve.
 | |
| // 2) Verify your use has the intended effect. This often requires inspecting
 | |
| //    the generated machine code.
 | |
| // 3) Prefer applying this attribute to individual variables. Avoid
 | |
| //    applying it to types. This tends to localize the effect.
 | |
| #define ABSL_CACHELINE_ALIGNED __attribute__((aligned(ABSL_CACHELINE_SIZE)))
 | |
| 
 | |
| #else  // not GCC
 | |
| #define ABSL_CACHELINE_SIZE 64
 | |
| #define ABSL_CACHELINE_ALIGNED
 | |
| #endif
 | |
| 
 | |
| // ABSL_PREDICT_TRUE, ABSL_PREDICT_FALSE
 | |
| //
 | |
| // Enables the compiler to prioritize compilation using static analysis for
 | |
| // likely paths within a boolean branch.
 | |
| //
 | |
| // Example:
 | |
| //
 | |
| //   if (ABSL_PREDICT_TRUE(expression)) {
 | |
| //     return result;                        // Faster if more likely
 | |
| //   } else {
 | |
| //     return 0;
 | |
| //   }
 | |
| //
 | |
| // Compilers can use the information that a certain branch is not likely to be
 | |
| // taken (for instance, a CHECK failure) to optimize for the common case in
 | |
| // the absence of better information (ie. compiling gcc with `-fprofile-arcs`).
 | |
| #if ABSL_HAVE_BUILTIN(__builtin_expect) || \
 | |
|     (defined(__GNUC__) && !defined(__clang__))
 | |
| #define ABSL_PREDICT_FALSE(x) (__builtin_expect(x, 0))
 | |
| #define ABSL_PREDICT_TRUE(x) (__builtin_expect(!!(x), 1))
 | |
| #else
 | |
| #define ABSL_PREDICT_FALSE(x) (x)
 | |
| #define ABSL_PREDICT_TRUE(x) (x)
 | |
| #endif
 | |
| 
 | |
| #endif  // ABSL_BASE_OPTIMIZATION_H_
 |