-- 049ac45508e335c6f010f2d28d71016b9fa65b4e by Derek Mauro <dmauro@google.com>: Fix librt detection PiperOrigin-RevId: 280207723 -- 6382c3a9fb2643af9dc031f92ca846c4a78e249c by Andy Getzendanner <durandal@google.com>: Fix Conan builds Import of https://github.com/abseil/abseil-cpp/pull/400 PiperOrigin-RevId: 280025424 -- aebcd52b1686ac82663a8d0193b60d0122a43372 by Samuel Benzaquen <sbenza@google.com>: Enable the assertion in the iterator's operator== and operator!= PiperOrigin-RevId: 279998951 -- 5b61d909e2159ac6fd45e0e456818db1e725ecd1 by Derek Mauro <dmauro@google.com>: Add best effort support for compiling much of Abseil with MinGW. This involves disabling ABSL_ATTRIBUTE_WEAK and adding link flags. A change to CCTZ is still necessary. Tests were not run yet, but most of them now build. PiperOrigin-RevId: 279966541 -- 4336f8b10cff906e2defdd7d1d449cde4907da5d by Abseil Team <absl-team@google.com>: Add comments and relax memory orders in base_internal::CallOnceImpl. Add a comment to document the memory order guarantee if base_internal::SpinLockWait() is called and returns kOnceDone. Add a comment for the load/store sequence in base_internal::CallOnceImpl based on Mike Burrows' explanation. The atomic load of 'control' in the #ifndef NDEBUG block does not need std::memory_order_acquire. It can use std::memory_order_relaxed. The atomic compare_exchange_strong of 'control' does not need std::memory_order_acquire in the success case. It can use std::memory_order_relaxed. PiperOrigin-RevId: 279814155 -- 407de3a5e9af957cded54a136ca0468bde620d4d by Abseil Team <absl-team@google.com>: Added a script to generate abseil.podspec from all BUILD.bazel files automatically. PiperOrigin-RevId: 279811441 -- 26139497d4a363d6c7bc989c554da593e8819a07 by Derek Mauro <dmauro@google.com>: Add missing copyright and Apache License to //absl/functional/BUILD.bazel PiperOrigin-RevId: 279795227 -- 98ed625b02af6e5834edf52a920d8ca2dab4cd90 by Matt Kulukundis <kfm@google.com>: Switch the implementation of hashtablez to *only* work on platforms that have a PER_THREAD_TLS. The old case is very slow (global mutex) and nobody collects data from that configuration anyway. PiperOrigin-RevId: 279775149 -- 07225900ef672c005c38f467ad3f92f38d0922b3 by Derek Mauro <dmauro@google.com>: Remove the minumum glibc version check PiperOrigin-RevId: 279750412 -- ec09956a951b4f52228ecc81968b8db7ae19ed15 by Derek Mauro <dmauro@google.com>: CMake only: link with -lrt to support older glibc versions PiperOrigin-RevId: 279741661 -- 97b113fb2e8246f6152c36330ba13793b37154b6 by Xiaoyi Zhang <zhangxy@google.com>: Internal change. PiperOrigin-RevId: 279390188 -- ca8f72f2721546cc9b01bd01b2ea144962e6e0c5 by Andy Getzendanner <durandal@google.com>: Expose PutTwoDigits for internal use within Abseil. PiperOrigin-RevId: 279374239 -- 14c6384cc03bbdfdefd2e4b635f104af5dd7e026 by Derek Mauro <dmauro@google.com>: Remove log_severity sources from the base target. They are already compiled as part of a separate library. PiperOrigin-RevId: 279372619 -- 3c5d926c718f8bf394e3bee87b6ba8d94601e0d3 by Abseil Team <absl-team@google.com>: s/indepdent/independent/g in SimpleAtof's documentation. PiperOrigin-RevId: 279350836 -- de2c44be8a8edf9efa1fe2007cba3564f3e5b0b8 by Abseil Team <absl-team@google.com>: Internal change PiperOrigin-RevId: 279346990 -- 2ba078341423fcf6d0ba5ca1831f86570a26e615 by Samuel Benzaquen <sbenza@google.com>: Add hash support for std::wstring, std::u16string and std::u32string. PiperOrigin-RevId: 279320672 -- 3272d3ffcfa55283a04f90e5868701912da95ef7 by Andy Soffer <asoffer@google.com>: Removing a bunch of __restricts that amount to no performance differences. One of these is the cause of https://github.com/abseil/abseil-cpp/issues/396. In particular, in one of the Vector128Store functions, restricts on two pointers that were indeed aliased seems to be the root cause of the issues. Closes #396 PiperOrigin-RevId: 279318999 -- 342f338ab31cc24344d5de8f28cf455bbb629a17 by Jorg Brown <jorg@google.com>: Support uint128 in SimpleAtoi PiperOrigin-RevId: 279234038 -- 81cb0a04cf2dc4515d303679fc60968712191571 by Derek Mauro <dmauro@google.com>: Change the check for futex availability to support older Linux systems PiperOrigin-RevId: 279147079 -- cb4ca4aa4c8d2d710a5d483c56c4ce4f979e14b1 by Abseil Team <absl-team@google.com>: Add IWYU pragma: export for int128 .inc files. PiperOrigin-RevId: 279107098 -- b8df86ef610c366729f07326c726f3e34817b4dd by Abseil Team <absl-team@google.com>: An optimization for Waiter::Post() in the SEM waiter mode. Like the FUTEX waiter mode, Waiter::Post() only needs to call Poke() if it incremented the atomic variable from 0. PiperOrigin-RevId: 279086133 GitOrigin-RevId: 049ac45508e335c6f010f2d28d71016b9fa65b4e Change-Id: I4c1a4073fff62cb6a1fcb1c104aa7d62dad588c2
		
			
				
	
	
		
			286 lines
		
	
	
	
		
			10 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
			
		
		
	
	
			286 lines
		
	
	
	
		
			10 KiB
		
	
	
	
		
			C++
		
	
	
	
	
	
| // Copyright 2018 The Abseil Authors.
 | |
| //
 | |
| // Licensed under the Apache License, Version 2.0 (the "License");
 | |
| // you may not use this file except in compliance with the License.
 | |
| // You may obtain a copy of the License at
 | |
| //
 | |
| //      https://www.apache.org/licenses/LICENSE-2.0
 | |
| //
 | |
| // Unless required by applicable law or agreed to in writing, software
 | |
| // distributed under the License is distributed on an "AS IS" BASIS,
 | |
| // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | |
| // See the License for the specific language governing permissions and
 | |
| // limitations under the License.
 | |
| //
 | |
| // -----------------------------------------------------------------------------
 | |
| // File: hashtablez_sampler.h
 | |
| // -----------------------------------------------------------------------------
 | |
| //
 | |
| // This header file defines the API for a low level library to sample hashtables
 | |
| // and collect runtime statistics about them.
 | |
| //
 | |
| // `HashtablezSampler` controls the lifecycle of `HashtablezInfo` objects which
 | |
| // store information about a single sample.
 | |
| //
 | |
| // `Record*` methods store information into samples.
 | |
| // `Sample()` and `Unsample()` make use of a single global sampler with
 | |
| // properties controlled by the flags hashtablez_enabled,
 | |
| // hashtablez_sample_rate, and hashtablez_max_samples.
 | |
| //
 | |
| // WARNING
 | |
| //
 | |
| // Using this sampling API may cause sampled Swiss tables to use the global
 | |
| // allocator (operator `new`) in addition to any custom allocator.  If you
 | |
| // are using a table in an unusual circumstance where allocation or calling a
 | |
| // linux syscall is unacceptable, this could interfere.
 | |
| //
 | |
| // This utility is internal-only. Use at your own risk.
 | |
| 
 | |
| #ifndef ABSL_CONTAINER_INTERNAL_HASHTABLEZ_SAMPLER_H_
 | |
| #define ABSL_CONTAINER_INTERNAL_HASHTABLEZ_SAMPLER_H_
 | |
| 
 | |
| #include <atomic>
 | |
| #include <functional>
 | |
| #include <memory>
 | |
| #include <vector>
 | |
| 
 | |
| #include "absl/base/internal/per_thread_tls.h"
 | |
| #include "absl/base/optimization.h"
 | |
| #include "absl/container/internal/have_sse.h"
 | |
| #include "absl/synchronization/mutex.h"
 | |
| #include "absl/utility/utility.h"
 | |
| 
 | |
| namespace absl {
 | |
| namespace container_internal {
 | |
| 
 | |
| // Stores information about a sampled hashtable.  All mutations to this *must*
 | |
| // be made through `Record*` functions below.  All reads from this *must* only
 | |
| // occur in the callback to `HashtablezSampler::Iterate`.
 | |
| struct HashtablezInfo {
 | |
|   // Constructs the object but does not fill in any fields.
 | |
|   HashtablezInfo();
 | |
|   ~HashtablezInfo();
 | |
|   HashtablezInfo(const HashtablezInfo&) = delete;
 | |
|   HashtablezInfo& operator=(const HashtablezInfo&) = delete;
 | |
| 
 | |
|   // Puts the object into a clean state, fills in the logically `const` members,
 | |
|   // blocking for any readers that are currently sampling the object.
 | |
|   void PrepareForSampling() ABSL_EXCLUSIVE_LOCKS_REQUIRED(init_mu);
 | |
| 
 | |
|   // These fields are mutated by the various Record* APIs and need to be
 | |
|   // thread-safe.
 | |
|   std::atomic<size_t> capacity;
 | |
|   std::atomic<size_t> size;
 | |
|   std::atomic<size_t> num_erases;
 | |
|   std::atomic<size_t> max_probe_length;
 | |
|   std::atomic<size_t> total_probe_length;
 | |
|   std::atomic<size_t> hashes_bitwise_or;
 | |
|   std::atomic<size_t> hashes_bitwise_and;
 | |
| 
 | |
|   // `HashtablezSampler` maintains intrusive linked lists for all samples.  See
 | |
|   // comments on `HashtablezSampler::all_` for details on these.  `init_mu`
 | |
|   // guards the ability to restore the sample to a pristine state.  This
 | |
|   // prevents races with sampling and resurrecting an object.
 | |
|   absl::Mutex init_mu;
 | |
|   HashtablezInfo* next;
 | |
|   HashtablezInfo* dead ABSL_GUARDED_BY(init_mu);
 | |
| 
 | |
|   // All of the fields below are set by `PrepareForSampling`, they must not be
 | |
|   // mutated in `Record*` functions.  They are logically `const` in that sense.
 | |
|   // These are guarded by init_mu, but that is not externalized to clients, who
 | |
|   // can only read them during `HashtablezSampler::Iterate` which will hold the
 | |
|   // lock.
 | |
|   static constexpr int kMaxStackDepth = 64;
 | |
|   absl::Time create_time;
 | |
|   int32_t depth;
 | |
|   void* stack[kMaxStackDepth];
 | |
| };
 | |
| 
 | |
| inline void RecordRehashSlow(HashtablezInfo* info, size_t total_probe_length) {
 | |
| #if SWISSTABLE_HAVE_SSE2
 | |
|   total_probe_length /= 16;
 | |
| #else
 | |
|   total_probe_length /= 8;
 | |
| #endif
 | |
|   info->total_probe_length.store(total_probe_length, std::memory_order_relaxed);
 | |
|   info->num_erases.store(0, std::memory_order_relaxed);
 | |
| }
 | |
| 
 | |
| inline void RecordStorageChangedSlow(HashtablezInfo* info, size_t size,
 | |
|                                      size_t capacity) {
 | |
|   info->size.store(size, std::memory_order_relaxed);
 | |
|   info->capacity.store(capacity, std::memory_order_relaxed);
 | |
|   if (size == 0) {
 | |
|     // This is a clear, reset the total/num_erases too.
 | |
|     RecordRehashSlow(info, 0);
 | |
|   }
 | |
| }
 | |
| 
 | |
| void RecordInsertSlow(HashtablezInfo* info, size_t hash,
 | |
|                       size_t distance_from_desired);
 | |
| 
 | |
| inline void RecordEraseSlow(HashtablezInfo* info) {
 | |
|   info->size.fetch_sub(1, std::memory_order_relaxed);
 | |
|   info->num_erases.fetch_add(1, std::memory_order_relaxed);
 | |
| }
 | |
| 
 | |
| HashtablezInfo* SampleSlow(int64_t* next_sample);
 | |
| void UnsampleSlow(HashtablezInfo* info);
 | |
| 
 | |
| class HashtablezInfoHandle {
 | |
|  public:
 | |
|   explicit HashtablezInfoHandle() : info_(nullptr) {}
 | |
|   explicit HashtablezInfoHandle(HashtablezInfo* info) : info_(info) {}
 | |
|   ~HashtablezInfoHandle() {
 | |
|     if (ABSL_PREDICT_TRUE(info_ == nullptr)) return;
 | |
|     UnsampleSlow(info_);
 | |
|   }
 | |
| 
 | |
|   HashtablezInfoHandle(const HashtablezInfoHandle&) = delete;
 | |
|   HashtablezInfoHandle& operator=(const HashtablezInfoHandle&) = delete;
 | |
| 
 | |
|   HashtablezInfoHandle(HashtablezInfoHandle&& o) noexcept
 | |
|       : info_(absl::exchange(o.info_, nullptr)) {}
 | |
|   HashtablezInfoHandle& operator=(HashtablezInfoHandle&& o) noexcept {
 | |
|     if (ABSL_PREDICT_FALSE(info_ != nullptr)) {
 | |
|       UnsampleSlow(info_);
 | |
|     }
 | |
|     info_ = absl::exchange(o.info_, nullptr);
 | |
|     return *this;
 | |
|   }
 | |
| 
 | |
|   inline void RecordStorageChanged(size_t size, size_t capacity) {
 | |
|     if (ABSL_PREDICT_TRUE(info_ == nullptr)) return;
 | |
|     RecordStorageChangedSlow(info_, size, capacity);
 | |
|   }
 | |
| 
 | |
|   inline void RecordRehash(size_t total_probe_length) {
 | |
|     if (ABSL_PREDICT_TRUE(info_ == nullptr)) return;
 | |
|     RecordRehashSlow(info_, total_probe_length);
 | |
|   }
 | |
| 
 | |
|   inline void RecordInsert(size_t hash, size_t distance_from_desired) {
 | |
|     if (ABSL_PREDICT_TRUE(info_ == nullptr)) return;
 | |
|     RecordInsertSlow(info_, hash, distance_from_desired);
 | |
|   }
 | |
| 
 | |
|   inline void RecordErase() {
 | |
|     if (ABSL_PREDICT_TRUE(info_ == nullptr)) return;
 | |
|     RecordEraseSlow(info_);
 | |
|   }
 | |
| 
 | |
|   friend inline void swap(HashtablezInfoHandle& lhs,
 | |
|                           HashtablezInfoHandle& rhs) {
 | |
|     std::swap(lhs.info_, rhs.info_);
 | |
|   }
 | |
| 
 | |
|  private:
 | |
|   friend class HashtablezInfoHandlePeer;
 | |
|   HashtablezInfo* info_;
 | |
| };
 | |
| 
 | |
| #if ABSL_PER_THREAD_TLS == 1
 | |
| extern ABSL_PER_THREAD_TLS_KEYWORD int64_t global_next_sample;
 | |
| #endif  // ABSL_PER_THREAD_TLS
 | |
| 
 | |
| // Returns an RAII sampling handle that manages registration and unregistation
 | |
| // with the global sampler.
 | |
| inline HashtablezInfoHandle Sample() {
 | |
| #if ABSL_PER_THREAD_TLS == 1
 | |
|   if (ABSL_PREDICT_TRUE(--global_next_sample > 0)) {
 | |
|     return HashtablezInfoHandle(nullptr);
 | |
|   }
 | |
|   return HashtablezInfoHandle(SampleSlow(&global_next_sample));
 | |
| #else
 | |
|   return HashtablezInfoHandle(nullptr);
 | |
| #endif  // !ABSL_PER_THREAD_TLS
 | |
| }
 | |
| 
 | |
| // Holds samples and their associated stack traces with a soft limit of
 | |
| // `SetHashtablezMaxSamples()`.
 | |
| //
 | |
| // Thread safe.
 | |
| class HashtablezSampler {
 | |
|  public:
 | |
|   // Returns a global Sampler.
 | |
|   static HashtablezSampler& Global();
 | |
| 
 | |
|   HashtablezSampler();
 | |
|   ~HashtablezSampler();
 | |
| 
 | |
|   // Registers for sampling.  Returns an opaque registration info.
 | |
|   HashtablezInfo* Register();
 | |
| 
 | |
|   // Unregisters the sample.
 | |
|   void Unregister(HashtablezInfo* sample);
 | |
| 
 | |
|   // The dispose callback will be called on all samples the moment they are
 | |
|   // being unregistered. Only affects samples that are unregistered after the
 | |
|   // callback has been set.
 | |
|   // Returns the previous callback.
 | |
|   using DisposeCallback = void (*)(const HashtablezInfo&);
 | |
|   DisposeCallback SetDisposeCallback(DisposeCallback f);
 | |
| 
 | |
|   // Iterates over all the registered `StackInfo`s.  Returning the number of
 | |
|   // samples that have been dropped.
 | |
|   int64_t Iterate(const std::function<void(const HashtablezInfo& stack)>& f);
 | |
| 
 | |
|  private:
 | |
|   void PushNew(HashtablezInfo* sample);
 | |
|   void PushDead(HashtablezInfo* sample);
 | |
|   HashtablezInfo* PopDead();
 | |
| 
 | |
|   std::atomic<size_t> dropped_samples_;
 | |
|   std::atomic<size_t> size_estimate_;
 | |
| 
 | |
|   // Intrusive lock free linked lists for tracking samples.
 | |
|   //
 | |
|   // `all_` records all samples (they are never removed from this list) and is
 | |
|   // terminated with a `nullptr`.
 | |
|   //
 | |
|   // `graveyard_.dead` is a circular linked list.  When it is empty,
 | |
|   // `graveyard_.dead == &graveyard`.  The list is circular so that
 | |
|   // every item on it (even the last) has a non-null dead pointer.  This allows
 | |
|   // `Iterate` to determine if a given sample is live or dead using only
 | |
|   // information on the sample itself.
 | |
|   //
 | |
|   // For example, nodes [A, B, C, D, E] with [A, C, E] alive and [B, D] dead
 | |
|   // looks like this (G is the Graveyard):
 | |
|   //
 | |
|   //           +---+    +---+    +---+    +---+    +---+
 | |
|   //    all -->| A |--->| B |--->| C |--->| D |--->| E |
 | |
|   //           |   |    |   |    |   |    |   |    |   |
 | |
|   //   +---+   |   | +->|   |-+  |   | +->|   |-+  |   |
 | |
|   //   | G |   +---+ |  +---+ |  +---+ |  +---+ |  +---+
 | |
|   //   |   |         |        |        |        |
 | |
|   //   |   | --------+        +--------+        |
 | |
|   //   +---+                                    |
 | |
|   //     ^                                      |
 | |
|   //     +--------------------------------------+
 | |
|   //
 | |
|   std::atomic<HashtablezInfo*> all_;
 | |
|   HashtablezInfo graveyard_;
 | |
| 
 | |
|   std::atomic<DisposeCallback> dispose_;
 | |
| };
 | |
| 
 | |
| // Enables or disables sampling for Swiss tables.
 | |
| void SetHashtablezEnabled(bool enabled);
 | |
| 
 | |
| // Sets the rate at which Swiss tables will be sampled.
 | |
| void SetHashtablezSampleParameter(int32_t rate);
 | |
| 
 | |
| // Sets a soft max for the number of samples that will be kept.
 | |
| void SetHashtablezMaxSamples(int32_t max);
 | |
| 
 | |
| // Configuration override.
 | |
| // This allows process-wide sampling without depending on order of
 | |
| // initialization of static storage duration objects.
 | |
| // The definition of this constant is weak, which allows us to inject a
 | |
| // different value for it at link time.
 | |
| extern "C" bool AbslContainerInternalSampleEverything();
 | |
| 
 | |
| }  // namespace container_internal
 | |
| }  // namespace absl
 | |
| 
 | |
| #endif  // ABSL_CONTAINER_INTERNAL_HASHTABLEZ_SAMPLER_H_
 |