ClickHouse/src/Common/MemoryTracker.cpp

343 lines
11 KiB
C++
Raw Normal View History

#include "MemoryTracker.h"
2020-01-16 12:37:29 +00:00
#include <IO/WriteHelpers.h>
#include "Common/TraceCollector.h"
#include <Common/CurrentThread.h>
#include <Common/Exception.h>
#include <Common/formatReadable.h>
2020-01-16 12:37:29 +00:00
#include <common/logger_useful.h>
2020-09-09 16:59:38 +00:00
#include <Common/ProfileEvents.h>
2020-01-16 12:37:29 +00:00
2020-01-21 13:53:30 +00:00
#include <atomic>
2020-01-22 15:20:19 +00:00
#include <cmath>
2020-04-30 13:25:17 +00:00
#include <random>
2020-01-16 12:37:29 +00:00
#include <cstdlib>
namespace
{
MemoryTracker * getMemoryTracker()
{
if (auto * thread_memory_tracker = DB::CurrentThread::getMemoryTracker())
return thread_memory_tracker;
/// Once the main thread is initialized,
/// total_memory_tracker is initialized too.
/// And can be used, since MainThreadStatus is required for profiling.
if (DB::MainThreadStatus::get())
return &total_memory_tracker;
return nullptr;
}
}
namespace DB
{
namespace ErrorCodes
{
extern const int MEMORY_LIMIT_EXCEEDED;
2020-08-07 19:36:04 +00:00
extern const int LOGICAL_ERROR;
}
}
2020-09-09 16:59:38 +00:00
namespace ProfileEvents
{
extern const Event QueryMemoryLimitExceeded;
}
static constexpr size_t log_peak_memory_usage_every = 1ULL << 30;
thread_local bool MemoryTracker::BlockerInThread::is_blocked = false;
2020-04-19 21:43:06 +00:00
MemoryTracker total_memory_tracker(nullptr, VariableContext::Global);
MemoryTracker::MemoryTracker(VariableContext level_) : parent(&total_memory_tracker), level(level_) {}
MemoryTracker::MemoryTracker(MemoryTracker * parent_, VariableContext level_) : parent(parent_), level(level_) {}
MemoryTracker::~MemoryTracker()
{
2020-04-19 21:43:06 +00:00
if ((level == VariableContext::Process || level == VariableContext::User) && peak)
{
try
{
logPeakMemoryUsage();
}
catch (...)
{
/// Exception in Logger, intentionally swallow.
}
}
}
void MemoryTracker::logPeakMemoryUsage() const
{
const auto * description = description_ptr.load(std::memory_order_relaxed);
LOG_DEBUG(&Poco::Logger::get("MemoryTracker"),
"Peak memory usage{}: {}.", (description ? " " + std::string(description) : ""), ReadableSize(peak));
}
2020-04-19 21:43:06 +00:00
void MemoryTracker::logMemoryUsage(Int64 current) const
{
const auto * description = description_ptr.load(std::memory_order_relaxed);
LOG_DEBUG(&Poco::Logger::get("MemoryTracker"),
"Current memory usage{}: {}.", (description ? " " + std::string(description) : ""), ReadableSize(current));
}
void MemoryTracker::alloc(Int64 size)
{
2020-08-07 19:36:04 +00:00
if (size < 0)
throw DB::Exception(DB::ErrorCodes::LOGICAL_ERROR, "Negative size ({}) is passed to MemoryTracker. It is a bug.", size);
if (BlockerInThread::isBlocked())
return;
2017-09-17 00:13:26 +00:00
/** Using memory_order_relaxed means that if allocations are done simultaneously,
* we allow exception about memory limit exceeded to be thrown only on next allocation.
2017-09-17 00:13:26 +00:00
* So, we allow over-allocations.
*/
Int64 will_be = size + amount.fetch_add(size, std::memory_order_relaxed);
2018-10-08 05:30:03 +00:00
if (metric != CurrentMetrics::end())
CurrentMetrics::add(metric, size);
2020-01-21 13:53:30 +00:00
Int64 current_hard_limit = hard_limit.load(std::memory_order_relaxed);
2020-01-22 15:20:19 +00:00
Int64 current_profiler_limit = profiler_limit.load(std::memory_order_relaxed);
/// Cap the limit to the total_memory_tracker, since it may include some drift
/// for user-level memory tracker.
///
2020-08-08 00:47:03 +00:00
/// And since total_memory_tracker is reset to the process resident
/// memory peridically (in AsynchronousMetrics::update()), any limit can be
/// capped to it, to avoid possible drift.
if (unlikely(current_hard_limit
&& will_be > current_hard_limit
&& level == VariableContext::User))
{
Int64 total_amount = total_memory_tracker.get();
if (amount > total_amount)
{
set(total_amount);
will_be = size + total_amount;
}
}
2020-04-30 13:25:17 +00:00
std::bernoulli_distribution fault(fault_probability);
if (unlikely(fault_probability && fault(thread_local_rng)))
{
/// Prevent recursion. Exception::ctor -> std::string -> new[] -> MemoryTracker::alloc
BlockerInThread untrack_lock;
2020-09-09 16:59:38 +00:00
ProfileEvents::increment(ProfileEvents::QueryMemoryLimitExceeded);
std::stringstream message;
2020-11-07 00:14:53 +00:00
message.exceptions(std::ios::failbit);
message << "Memory tracker";
if (const auto * description = description_ptr.load(std::memory_order_relaxed))
message << " " << description;
message << ": fault injected. Would use " << formatReadableSizeWithBinarySuffix(will_be)
<< " (attempt to allocate chunk of " << size << " bytes)"
2020-01-21 13:53:30 +00:00
<< ", maximum: " << formatReadableSizeWithBinarySuffix(current_hard_limit);
amount.fetch_sub(size, std::memory_order_relaxed);
throw DB::Exception(message.str(), DB::ErrorCodes::MEMORY_LIMIT_EXCEEDED);
}
2020-01-22 15:20:19 +00:00
if (unlikely(current_profiler_limit && will_be > current_profiler_limit))
2020-01-21 13:53:30 +00:00
{
BlockerInThread untrack_lock;
DB::TraceCollector::collect(DB::TraceType::Memory, StackTrace(), size);
setOrRaiseProfilerLimit((will_be + profiler_step - 1) / profiler_step * profiler_step);
2020-01-21 13:53:30 +00:00
}
2020-04-30 13:25:17 +00:00
std::bernoulli_distribution sample(sample_probability);
if (unlikely(sample_probability && sample(thread_local_rng)))
{
BlockerInThread untrack_lock;
2020-04-30 13:25:17 +00:00
DB::TraceCollector::collect(DB::TraceType::MemorySample, StackTrace(), size);
}
2020-01-21 13:53:30 +00:00
if (unlikely(current_hard_limit && will_be > current_hard_limit))
{
/// Prevent recursion. Exception::ctor -> std::string -> new[] -> MemoryTracker::alloc
BlockerInThread untrack_lock;
2020-09-09 16:59:38 +00:00
ProfileEvents::increment(ProfileEvents::QueryMemoryLimitExceeded);
std::stringstream message;
2020-11-07 00:14:53 +00:00
message.exceptions(std::ios::failbit);
message << "Memory limit";
if (const auto * description = description_ptr.load(std::memory_order_relaxed))
message << " " << description;
message << " exceeded: would use " << formatReadableSizeWithBinarySuffix(will_be)
<< " (attempt to allocate chunk of " << size << " bytes)"
2020-01-21 13:53:30 +00:00
<< ", maximum: " << formatReadableSizeWithBinarySuffix(current_hard_limit);
2019-07-17 15:16:28 +00:00
amount.fetch_sub(size, std::memory_order_relaxed);
throw DB::Exception(message.str(), DB::ErrorCodes::MEMORY_LIMIT_EXCEEDED);
}
2020-04-19 21:43:06 +00:00
updatePeak(will_be);
2020-04-22 00:29:38 +00:00
if (auto * loaded_next = parent.load(std::memory_order_relaxed))
2020-04-19 21:43:06 +00:00
loaded_next->alloc(size);
}
void MemoryTracker::updatePeak(Int64 will_be)
{
auto peak_old = peak.load(std::memory_order_relaxed);
if (will_be > peak_old) /// Races doesn't matter. Could rewrite with CAS, but not worth.
{
peak.store(will_be, std::memory_order_relaxed);
2020-04-19 21:43:06 +00:00
if ((level == VariableContext::Process || level == VariableContext::Global)
&& will_be / log_peak_memory_usage_every > peak_old / log_peak_memory_usage_every)
logMemoryUsage(will_be);
}
}
void MemoryTracker::free(Int64 size)
{
if (BlockerInThread::isBlocked())
return;
2020-04-30 13:25:17 +00:00
std::bernoulli_distribution sample(sample_probability);
if (unlikely(sample_probability && sample(thread_local_rng)))
{
BlockerInThread untrack_lock;
2020-04-30 13:25:17 +00:00
DB::TraceCollector::collect(DB::TraceType::MemorySample, StackTrace(), -size);
}
Int64 accounted_size = size;
if (level == VariableContext::Thread)
2017-09-17 00:13:26 +00:00
{
/// Could become negative if memory allocated in this thread is freed in another one
amount.fetch_sub(accounted_size, std::memory_order_relaxed);
}
else
{
Int64 new_amount = amount.fetch_sub(accounted_size, std::memory_order_relaxed) - accounted_size;
/** Sometimes, query could free some data, that was allocated outside of query context.
* Example: cache eviction.
* To avoid negative memory usage, we "saturate" amount.
* Memory usage will be calculated with some error.
* NOTE: The code is not atomic. Not worth to fix.
*/
if (unlikely(new_amount < 0))
{
amount.fetch_sub(new_amount);
accounted_size += new_amount;
}
2017-09-17 00:13:26 +00:00
}
2020-04-22 00:29:38 +00:00
if (auto * loaded_next = parent.load(std::memory_order_relaxed))
loaded_next->free(size);
2018-10-08 05:30:03 +00:00
if (metric != CurrentMetrics::end())
CurrentMetrics::sub(metric, accounted_size);
}
void MemoryTracker::resetCounters()
{
amount.store(0, std::memory_order_relaxed);
peak.store(0, std::memory_order_relaxed);
2020-01-21 13:53:30 +00:00
hard_limit.store(0, std::memory_order_relaxed);
2020-01-22 15:20:19 +00:00
profiler_limit.store(0, std::memory_order_relaxed);
}
void MemoryTracker::reset()
{
if (metric != CurrentMetrics::end())
CurrentMetrics::sub(metric, amount.load(std::memory_order_relaxed));
resetCounters();
}
2020-04-19 21:43:06 +00:00
void MemoryTracker::set(Int64 to)
{
amount.store(to, std::memory_order_relaxed);
updatePeak(to);
}
2020-01-21 13:53:30 +00:00
void MemoryTracker::setOrRaiseHardLimit(Int64 value)
{
/// This is just atomic set to maximum.
Int64 old_value = hard_limit.load(std::memory_order_relaxed);
while (old_value < value && !hard_limit.compare_exchange_weak(old_value, value))
;
}
2020-01-22 15:20:19 +00:00
void MemoryTracker::setOrRaiseProfilerLimit(Int64 value)
{
2020-01-22 15:20:19 +00:00
Int64 old_value = profiler_limit.load(std::memory_order_relaxed);
while (old_value < value && !profiler_limit.compare_exchange_weak(old_value, value))
;
}
namespace CurrentMemoryTracker
{
2020-04-30 13:25:17 +00:00
using DB::current_thread;
void alloc(Int64 size)
{
if (auto * memory_tracker = getMemoryTracker())
2019-07-10 18:12:50 +00:00
{
if (current_thread)
{
current_thread->untracked_memory += size;
if (current_thread->untracked_memory > current_thread->untracked_memory_limit)
{
/// Zero untracked before track. If tracker throws out-of-limit we would be able to alloc up to untracked_memory_limit bytes
/// more. It could be useful to enlarge Exception message in rethrow logic.
Int64 tmp = current_thread->untracked_memory;
current_thread->untracked_memory = 0;
memory_tracker->alloc(tmp);
}
}
/// total_memory_tracker only, ignore untracked_memory
else
2019-07-10 18:12:50 +00:00
{
memory_tracker->alloc(size);
2019-07-10 18:12:50 +00:00
}
}
}
void realloc(Int64 old_size, Int64 new_size)
{
2019-07-10 18:12:50 +00:00
Int64 addition = new_size - old_size;
2020-01-16 12:37:29 +00:00
addition > 0 ? alloc(addition) : free(-addition);
}
void free(Int64 size)
{
if (auto * memory_tracker = getMemoryTracker())
2019-07-10 18:12:50 +00:00
{
if (current_thread)
{
current_thread->untracked_memory -= size;
if (current_thread->untracked_memory < -current_thread->untracked_memory_limit)
{
memory_tracker->free(-current_thread->untracked_memory);
current_thread->untracked_memory = 0;
}
}
/// total_memory_tracker only, ignore untracked_memory
else
2019-07-10 18:12:50 +00:00
{
memory_tracker->free(size);
2019-07-10 18:12:50 +00:00
}
}
}
}