ClickHouse/src/Common/MemoryTracker.cpp

430 lines
14 KiB
C++
Raw Normal View History

#include "MemoryTracker.h"
2020-01-16 12:37:29 +00:00
#include <IO/WriteHelpers.h>
#include "Common/TraceCollector.h"
2021-10-22 12:56:09 +00:00
#include "Common/VariableContext.h"
#include <Common/Exception.h>
#include <Common/formatReadable.h>
2021-10-02 07:13:14 +00:00
#include <base/logger_useful.h>
2020-09-09 16:59:38 +00:00
#include <Common/ProfileEvents.h>
2021-01-27 00:54:57 +00:00
#include <Common/thread_local_rng.h>
2021-10-22 12:56:09 +00:00
#include <Common/OvercommitTracker.h>
2020-01-16 12:37:29 +00:00
2020-01-21 13:53:30 +00:00
#include <atomic>
2020-01-22 15:20:19 +00:00
#include <cmath>
2020-04-30 13:25:17 +00:00
#include <random>
2020-01-16 12:37:29 +00:00
#include <cstdlib>
2021-01-27 00:54:57 +00:00
2021-01-12 14:34:50 +00:00
#ifdef MEMORY_TRACKER_DEBUG_CHECKS
2021-08-27 08:21:19 +00:00
thread_local bool memory_tracker_always_throw_logical_error_on_allocation = false;
2021-01-12 14:34:50 +00:00
#endif
namespace
{
/// MemoryTracker cannot throw MEMORY_LIMIT_EXCEEDED (either configured memory
/// limit reached or fault injected), in the following cases:
///
/// - when it is explicitly blocked with LockExceptionInThread
///
/// - when there are uncaught exceptions objects in the current thread
/// (to avoid std::terminate())
///
/// NOTE: that since C++11 destructor marked with noexcept by default, and
/// this means that any throw from destructor (that is not marked with
/// noexcept(false)) will cause std::terminate()
bool inline memoryTrackerCanThrow(VariableContext level, bool fault_injection)
{
return !MemoryTracker::LockExceptionInThread::isBlocked(level, fault_injection) && !std::uncaught_exceptions();
}
}
namespace DB
{
namespace ErrorCodes
{
extern const int MEMORY_LIMIT_EXCEEDED;
2020-08-07 19:36:04 +00:00
extern const int LOGICAL_ERROR;
}
}
2020-09-09 16:59:38 +00:00
namespace ProfileEvents
{
extern const Event QueryMemoryLimitExceeded;
}
static constexpr size_t log_peak_memory_usage_every = 1ULL << 30;
// BlockerInThread
thread_local uint64_t MemoryTracker::BlockerInThread::counter = 0;
thread_local VariableContext MemoryTracker::BlockerInThread::level = VariableContext::Global;
MemoryTracker::BlockerInThread::BlockerInThread(VariableContext level_)
: previous_level(level)
{
++counter;
level = level_;
}
MemoryTracker::BlockerInThread::~BlockerInThread()
{
--counter;
level = previous_level;
}
/// LockExceptionInThread
thread_local uint64_t MemoryTracker::LockExceptionInThread::counter = 0;
thread_local VariableContext MemoryTracker::LockExceptionInThread::level = VariableContext::Global;
thread_local bool MemoryTracker::LockExceptionInThread::block_fault_injections = false;
MemoryTracker::LockExceptionInThread::LockExceptionInThread(VariableContext level_, bool block_fault_injections_)
: previous_level(level)
, previous_block_fault_injections(block_fault_injections)
{
++counter;
level = level_;
block_fault_injections = block_fault_injections_;
}
MemoryTracker::LockExceptionInThread::~LockExceptionInThread()
{
--counter;
level = previous_level;
block_fault_injections = previous_block_fault_injections;
}
2020-04-19 21:43:06 +00:00
MemoryTracker total_memory_tracker(nullptr, VariableContext::Global);
MemoryTracker::MemoryTracker(VariableContext level_) : parent(&total_memory_tracker), level(level_) {}
MemoryTracker::MemoryTracker(MemoryTracker * parent_, VariableContext level_) : parent(parent_), level(level_) {}
MemoryTracker::~MemoryTracker()
{
2021-10-22 12:56:09 +00:00
if (level == VariableContext::Process)
{
auto * loaded_next = getParent();
while (loaded_next != nullptr)
{
if (auto * next_overcommit_tracker = loaded_next->overcommit_tracker)
next_overcommit_tracker->unsubscribe(this);
loaded_next = loaded_next->getParent();
}
}
2020-04-19 21:43:06 +00:00
if ((level == VariableContext::Process || level == VariableContext::User) && peak)
{
try
{
logPeakMemoryUsage();
}
catch (...)
{
/// Exception in Logger, intentionally swallow.
}
}
}
void MemoryTracker::logPeakMemoryUsage() const
{
const auto * description = description_ptr.load(std::memory_order_relaxed);
LOG_DEBUG(&Poco::Logger::get("MemoryTracker"),
"Peak memory usage{}: {}.", (description ? " " + std::string(description) : ""), ReadableSize(peak));
}
2020-04-19 21:43:06 +00:00
void MemoryTracker::logMemoryUsage(Int64 current) const
{
const auto * description = description_ptr.load(std::memory_order_relaxed);
LOG_DEBUG(&Poco::Logger::get("MemoryTracker"),
"Current memory usage{}: {}.", (description ? " " + std::string(description) : ""), ReadableSize(current));
}
2021-05-25 09:50:55 +00:00
void MemoryTracker::allocImpl(Int64 size, bool throw_if_memory_exceeded)
{
2020-08-07 19:36:04 +00:00
if (size < 0)
throw DB::Exception(DB::ErrorCodes::LOGICAL_ERROR, "Negative size ({}) is passed to MemoryTracker. It is a bug.", size);
if (BlockerInThread::isBlocked(level))
{
/// Since the BlockerInThread should respect the level, we should go to the next parent.
if (auto * loaded_next = parent.load(std::memory_order_relaxed))
2021-05-25 09:50:55 +00:00
loaded_next->allocImpl(size, throw_if_memory_exceeded);
return;
}
2017-09-17 00:13:26 +00:00
/** Using memory_order_relaxed means that if allocations are done simultaneously,
* we allow exception about memory limit exceeded to be thrown only on next allocation.
2017-09-17 00:13:26 +00:00
* So, we allow over-allocations.
*/
Int64 will_be = size + amount.fetch_add(size, std::memory_order_relaxed);
2021-01-17 12:40:14 +00:00
auto metric_loaded = metric.load(std::memory_order_relaxed);
if (metric_loaded != CurrentMetrics::end())
CurrentMetrics::add(metric_loaded, size);
2020-01-21 13:53:30 +00:00
Int64 current_hard_limit = hard_limit.load(std::memory_order_relaxed);
2020-01-22 15:20:19 +00:00
Int64 current_profiler_limit = profiler_limit.load(std::memory_order_relaxed);
/// Cap the limit to the total_memory_tracker, since it may include some drift
/// for user-level memory tracker.
///
2020-08-08 00:47:03 +00:00
/// And since total_memory_tracker is reset to the process resident
/// memory peridically (in AsynchronousMetrics::update()), any limit can be
/// capped to it, to avoid possible drift.
if (unlikely(current_hard_limit
&& will_be > current_hard_limit
&& level == VariableContext::User))
{
Int64 total_amount = total_memory_tracker.get();
if (amount > total_amount)
{
set(total_amount);
will_be = size + total_amount;
}
}
2021-05-25 09:50:55 +00:00
#ifdef MEMORY_TRACKER_DEBUG_CHECKS
2021-08-27 08:21:19 +00:00
if (unlikely(memory_tracker_always_throw_logical_error_on_allocation))
2021-05-25 20:54:48 +00:00
{
2021-08-27 08:21:19 +00:00
memory_tracker_always_throw_logical_error_on_allocation = false;
2021-05-25 20:54:48 +00:00
throw DB::Exception(DB::ErrorCodes::LOGICAL_ERROR, "Memory tracker: allocations not allowed.");
}
2021-01-12 14:34:50 +00:00
#endif
2021-05-25 20:54:48 +00:00
std::bernoulli_distribution fault(fault_probability);
if (unlikely(fault_probability && fault(thread_local_rng)) && memoryTrackerCanThrow(level, true) && throw_if_memory_exceeded)
{
/// Prevent recursion. Exception::ctor -> std::string -> new[] -> MemoryTracker::alloc
BlockerInThread untrack_lock(VariableContext::Global);
2020-01-21 13:53:30 +00:00
2021-05-25 20:54:48 +00:00
ProfileEvents::increment(ProfileEvents::QueryMemoryLimitExceeded);
const auto * description = description_ptr.load(std::memory_order_relaxed);
amount.fetch_sub(size, std::memory_order_relaxed);
throw DB::Exception(
DB::ErrorCodes::MEMORY_LIMIT_EXCEEDED,
"Memory tracker{}{}: fault injected. Would use {} (attempt to allocate chunk of {} bytes), maximum: {}",
description ? " " : "",
description ? description : "",
formatReadableSizeWithBinarySuffix(will_be),
size,
formatReadableSizeWithBinarySuffix(current_hard_limit));
}
2020-04-30 13:25:17 +00:00
2021-05-25 20:54:48 +00:00
bool allocation_traced = false;
2021-05-25 20:54:48 +00:00
if (unlikely(current_profiler_limit && will_be > current_profiler_limit))
{
BlockerInThread untrack_lock(VariableContext::Global);
DB::TraceCollector::collect(DB::TraceType::Memory, StackTrace(), size);
setOrRaiseProfilerLimit((will_be + profiler_step - 1) / profiler_step * profiler_step);
allocation_traced = true;
}
2021-05-25 20:54:48 +00:00
std::bernoulli_distribution sample(sample_probability);
if (unlikely(sample_probability && sample(thread_local_rng)))
{
BlockerInThread untrack_lock(VariableContext::Global);
DB::TraceCollector::collect(DB::TraceType::MemorySample, StackTrace(), size);
allocation_traced = true;
2021-05-25 20:54:48 +00:00
}
if (unlikely(current_hard_limit && will_be > current_hard_limit) && memoryTrackerCanThrow(level, false) && throw_if_memory_exceeded)
{
2021-10-22 12:56:09 +00:00
bool need_to_throw = true;
if (overcommit_tracker)
need_to_throw = overcommit_tracker->needToStopQuery(this);
if (need_to_throw)
{
/// Prevent recursion. Exception::ctor -> std::string -> new[] -> MemoryTracker::alloc
BlockerInThread untrack_lock(VariableContext::Global);
ProfileEvents::increment(ProfileEvents::QueryMemoryLimitExceeded);
const auto * description = description_ptr.load(std::memory_order_relaxed);
throw DB::Exception(
DB::ErrorCodes::MEMORY_LIMIT_EXCEEDED,
"Memory limit{}{} exceeded: would use {} (attempt to allocate chunk of {} bytes), maximum: {}",
description ? " " : "",
description ? description : "",
formatReadableSizeWithBinarySuffix(will_be),
size,
formatReadableSizeWithBinarySuffix(current_hard_limit));
}
else
{
will_be = amount.load(std::memory_order_relaxed);
}
2021-05-25 20:54:48 +00:00
}
bool peak_updated;
2021-05-26 20:31:35 +00:00
if (throw_if_memory_exceeded)
{
/// Prevent recursion. Exception::ctor -> std::string -> new[] -> MemoryTracker::alloc
BlockerInThread untrack_lock(VariableContext::Global);
bool log_memory_usage = true;
peak_updated = updatePeak(will_be, log_memory_usage);
2021-05-26 20:31:35 +00:00
}
else
{
bool log_memory_usage = false;
peak_updated = updatePeak(will_be, log_memory_usage);
}
if (peak_updated && allocation_traced)
{
BlockerInThread untrack_lock(VariableContext::Global);
DB::TraceCollector::collect(DB::TraceType::MemoryPeak, StackTrace(), will_be);
2021-05-26 20:31:35 +00:00
}
2020-04-19 21:43:06 +00:00
2020-04-22 00:29:38 +00:00
if (auto * loaded_next = parent.load(std::memory_order_relaxed))
2021-05-25 09:50:55 +00:00
loaded_next->allocImpl(size, throw_if_memory_exceeded);
2020-04-19 21:43:06 +00:00
}
2021-05-25 09:50:55 +00:00
void MemoryTracker::alloc(Int64 size)
{
bool throw_if_memory_exceeded = true;
allocImpl(size, throw_if_memory_exceeded);
}
void MemoryTracker::allocNoThrow(Int64 size)
{
bool throw_if_memory_exceeded = false;
allocImpl(size, throw_if_memory_exceeded);
}
2020-04-19 21:43:06 +00:00
bool MemoryTracker::updatePeak(Int64 will_be, bool log_memory_usage)
2020-04-19 21:43:06 +00:00
{
auto peak_old = peak.load(std::memory_order_relaxed);
if (will_be > peak_old) /// Races doesn't matter. Could rewrite with CAS, but not worth.
{
peak.store(will_be, std::memory_order_relaxed);
2021-05-25 20:54:48 +00:00
if (log_memory_usage && (level == VariableContext::Process || level == VariableContext::Global)
2020-04-19 21:43:06 +00:00
&& will_be / log_peak_memory_usage_every > peak_old / log_peak_memory_usage_every)
logMemoryUsage(will_be);
return true;
}
return false;
}
void MemoryTracker::free(Int64 size)
{
if (BlockerInThread::isBlocked(level))
{
/// Since the BlockerInThread should respect the level, we should go to the next parent.
if (auto * loaded_next = parent.load(std::memory_order_relaxed))
loaded_next->free(size);
return;
}
2020-04-30 13:25:17 +00:00
std::bernoulli_distribution sample(sample_probability);
if (unlikely(sample_probability && sample(thread_local_rng)))
{
BlockerInThread untrack_lock(VariableContext::Global);
2020-04-30 13:25:17 +00:00
DB::TraceCollector::collect(DB::TraceType::MemorySample, StackTrace(), -size);
}
Int64 accounted_size = size;
if (level == VariableContext::Thread)
2017-09-17 00:13:26 +00:00
{
/// Could become negative if memory allocated in this thread is freed in another one
amount.fetch_sub(accounted_size, std::memory_order_relaxed);
}
else
{
Int64 new_amount = amount.fetch_sub(accounted_size, std::memory_order_relaxed) - accounted_size;
/** Sometimes, query could free some data, that was allocated outside of query context.
* Example: cache eviction.
* To avoid negative memory usage, we "saturate" amount.
* Memory usage will be calculated with some error.
* NOTE: The code is not atomic. Not worth to fix.
*/
if (unlikely(new_amount < 0))
{
amount.fetch_sub(new_amount);
accounted_size += new_amount;
}
2017-09-17 00:13:26 +00:00
}
2020-04-22 00:29:38 +00:00
if (auto * loaded_next = parent.load(std::memory_order_relaxed))
loaded_next->free(size);
2018-10-08 05:30:03 +00:00
2021-01-17 12:40:14 +00:00
auto metric_loaded = metric.load(std::memory_order_relaxed);
if (metric_loaded != CurrentMetrics::end())
CurrentMetrics::sub(metric_loaded, accounted_size);
}
2021-10-22 12:56:09 +00:00
OvercommitRatio MemoryTracker::getOvercommitRatio()
{
return { amount.load(std::memory_order_relaxed), soft_limit.load(std::memory_order_relaxed) };
}
OvercommitRatio MemoryTracker::getOvercommitRatio(Int64 limit)
{
return { amount.load(std::memory_order_relaxed), limit };
}
void MemoryTracker::resetCounters()
{
amount.store(0, std::memory_order_relaxed);
peak.store(0, std::memory_order_relaxed);
2021-10-22 12:56:09 +00:00
soft_limit.store(0, std::memory_order_relaxed);
2020-01-21 13:53:30 +00:00
hard_limit.store(0, std::memory_order_relaxed);
2020-01-22 15:20:19 +00:00
profiler_limit.store(0, std::memory_order_relaxed);
}
void MemoryTracker::reset()
{
2021-01-17 12:40:14 +00:00
auto metric_loaded = metric.load(std::memory_order_relaxed);
if (metric_loaded != CurrentMetrics::end())
CurrentMetrics::sub(metric_loaded, amount.load(std::memory_order_relaxed));
resetCounters();
}
2020-04-19 21:43:06 +00:00
void MemoryTracker::set(Int64 to)
{
amount.store(to, std::memory_order_relaxed);
2021-05-25 20:54:48 +00:00
bool log_memory_usage = true;
updatePeak(to, log_memory_usage);
2020-04-19 21:43:06 +00:00
}
2021-10-22 12:56:09 +00:00
void MemoryTracker::setSoftLimit(Int64 value)
{
soft_limit.store(value, std::memory_order_relaxed);
}
void MemoryTracker::setHardLimit(Int64 value)
{
hard_limit.store(value, std::memory_order_relaxed);
}
2020-01-21 13:53:30 +00:00
void MemoryTracker::setOrRaiseHardLimit(Int64 value)
{
/// This is just atomic set to maximum.
Int64 old_value = hard_limit.load(std::memory_order_relaxed);
2021-08-13 10:23:54 +00:00
while ((value == 0 || old_value < value) && !hard_limit.compare_exchange_weak(old_value, value))
2020-01-21 13:53:30 +00:00
;
}
2020-01-22 15:20:19 +00:00
void MemoryTracker::setOrRaiseProfilerLimit(Int64 value)
{
2020-01-22 15:20:19 +00:00
Int64 old_value = profiler_limit.load(std::memory_order_relaxed);
2021-08-13 10:23:54 +00:00
while ((value == 0 || old_value < value) && !profiler_limit.compare_exchange_weak(old_value, value))
;
}