- Add applesauce library reference documentation - Add rate limiting test report for Badger - Add memory monitoring for rate limiter (platform-specific implementations) - Enhance PID-controlled adaptive rate limiting - Update Neo4j and Badger monitors with improved load metrics - Add docker-compose configuration - Update README and configuration options 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
339 lines
9.5 KiB
Go
339 lines
9.5 KiB
Go
//go:build !(js && wasm)
|
|
|
|
package ratelimit
|
|
|
|
import (
|
|
"sync"
|
|
"sync/atomic"
|
|
"time"
|
|
|
|
"github.com/dgraph-io/badger/v4"
|
|
"lol.mleku.dev/log"
|
|
"next.orly.dev/pkg/interfaces/loadmonitor"
|
|
)
|
|
|
|
// BadgerMonitor implements loadmonitor.Monitor for the Badger database.
|
|
// It collects metrics from Badger's LSM tree, caches, and actual process memory.
|
|
// It also implements CompactableMonitor and EmergencyModeMonitor interfaces.
|
|
type BadgerMonitor struct {
|
|
db *badger.DB
|
|
|
|
// Target memory for pressure calculation
|
|
targetMemoryBytes atomic.Uint64
|
|
|
|
// Emergency mode configuration
|
|
emergencyThreshold atomic.Uint64 // stored as threshold * 1000 (e.g., 1500 = 1.5)
|
|
emergencyModeUntil atomic.Int64 // Unix nano when forced emergency mode ends
|
|
inEmergencyMode atomic.Bool
|
|
|
|
// Compaction state
|
|
isCompacting atomic.Bool
|
|
|
|
// Latency tracking with exponential moving average
|
|
queryLatencyNs atomic.Int64
|
|
writeLatencyNs atomic.Int64
|
|
latencyAlpha float64 // EMA coefficient (default 0.1)
|
|
|
|
// Cached metrics (updated by background goroutine)
|
|
metricsLock sync.RWMutex
|
|
cachedMetrics loadmonitor.Metrics
|
|
lastL0Tables int
|
|
lastL0Score float64
|
|
|
|
// Background collection
|
|
stopChan chan struct{}
|
|
stopped chan struct{}
|
|
interval time.Duration
|
|
}
|
|
|
|
// Compile-time checks for interface implementation
|
|
var _ loadmonitor.Monitor = (*BadgerMonitor)(nil)
|
|
var _ loadmonitor.CompactableMonitor = (*BadgerMonitor)(nil)
|
|
var _ loadmonitor.EmergencyModeMonitor = (*BadgerMonitor)(nil)
|
|
|
|
// NewBadgerMonitor creates a new Badger load monitor.
|
|
// The updateInterval controls how often metrics are collected (default 100ms).
|
|
func NewBadgerMonitor(db *badger.DB, updateInterval time.Duration) *BadgerMonitor {
|
|
if updateInterval <= 0 {
|
|
updateInterval = 100 * time.Millisecond
|
|
}
|
|
|
|
m := &BadgerMonitor{
|
|
db: db,
|
|
latencyAlpha: 0.1, // 10% new, 90% old for smooth EMA
|
|
stopChan: make(chan struct{}),
|
|
stopped: make(chan struct{}),
|
|
interval: updateInterval,
|
|
}
|
|
|
|
// Set a default target (1.5GB)
|
|
m.targetMemoryBytes.Store(1500 * 1024 * 1024)
|
|
|
|
// Default emergency threshold: 150% of target
|
|
m.emergencyThreshold.Store(1500)
|
|
|
|
return m
|
|
}
|
|
|
|
// SetEmergencyThreshold sets the memory threshold above which emergency mode is triggered.
|
|
// threshold is a fraction, e.g., 1.5 = 150% of target memory.
|
|
func (m *BadgerMonitor) SetEmergencyThreshold(threshold float64) {
|
|
m.emergencyThreshold.Store(uint64(threshold * 1000))
|
|
}
|
|
|
|
// GetEmergencyThreshold returns the current emergency threshold as a fraction.
|
|
func (m *BadgerMonitor) GetEmergencyThreshold() float64 {
|
|
return float64(m.emergencyThreshold.Load()) / 1000.0
|
|
}
|
|
|
|
// ForceEmergencyMode manually triggers emergency mode for a duration.
|
|
func (m *BadgerMonitor) ForceEmergencyMode(duration time.Duration) {
|
|
m.emergencyModeUntil.Store(time.Now().Add(duration).UnixNano())
|
|
m.inEmergencyMode.Store(true)
|
|
log.W.F("⚠️ emergency mode forced for %v", duration)
|
|
}
|
|
|
|
// TriggerCompaction initiates a Badger Flatten operation to compact all levels.
|
|
// This should be called when memory pressure is high and the database needs to
|
|
// reclaim space. It runs synchronously and may take significant time.
|
|
func (m *BadgerMonitor) TriggerCompaction() error {
|
|
if m.db == nil || m.db.IsClosed() {
|
|
return nil
|
|
}
|
|
|
|
if m.isCompacting.Load() {
|
|
log.D.Ln("compaction already in progress, skipping")
|
|
return nil
|
|
}
|
|
|
|
m.isCompacting.Store(true)
|
|
defer m.isCompacting.Store(false)
|
|
|
|
log.I.Ln("🗜️ triggering Badger compaction (Flatten)")
|
|
start := time.Now()
|
|
|
|
// Flatten with 4 workers (matches NumCompactors default)
|
|
err := m.db.Flatten(4)
|
|
if err != nil {
|
|
log.E.F("compaction failed: %v", err)
|
|
return err
|
|
}
|
|
|
|
// Also run value log GC to reclaim space
|
|
for {
|
|
err := m.db.RunValueLogGC(0.5)
|
|
if err != nil {
|
|
break // No more GC needed
|
|
}
|
|
}
|
|
|
|
log.I.F("🗜️ compaction completed in %v", time.Since(start))
|
|
return nil
|
|
}
|
|
|
|
// IsCompacting returns true if a compaction is currently in progress.
|
|
func (m *BadgerMonitor) IsCompacting() bool {
|
|
return m.isCompacting.Load()
|
|
}
|
|
|
|
// GetMetrics returns the current load metrics.
|
|
func (m *BadgerMonitor) GetMetrics() loadmonitor.Metrics {
|
|
m.metricsLock.RLock()
|
|
defer m.metricsLock.RUnlock()
|
|
return m.cachedMetrics
|
|
}
|
|
|
|
// RecordQueryLatency records a query latency sample using exponential moving average.
|
|
func (m *BadgerMonitor) RecordQueryLatency(latency time.Duration) {
|
|
ns := latency.Nanoseconds()
|
|
for {
|
|
old := m.queryLatencyNs.Load()
|
|
if old == 0 {
|
|
if m.queryLatencyNs.CompareAndSwap(0, ns) {
|
|
return
|
|
}
|
|
continue
|
|
}
|
|
// EMA: new = alpha * sample + (1-alpha) * old
|
|
newVal := int64(m.latencyAlpha*float64(ns) + (1-m.latencyAlpha)*float64(old))
|
|
if m.queryLatencyNs.CompareAndSwap(old, newVal) {
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// RecordWriteLatency records a write latency sample using exponential moving average.
|
|
func (m *BadgerMonitor) RecordWriteLatency(latency time.Duration) {
|
|
ns := latency.Nanoseconds()
|
|
for {
|
|
old := m.writeLatencyNs.Load()
|
|
if old == 0 {
|
|
if m.writeLatencyNs.CompareAndSwap(0, ns) {
|
|
return
|
|
}
|
|
continue
|
|
}
|
|
// EMA: new = alpha * sample + (1-alpha) * old
|
|
newVal := int64(m.latencyAlpha*float64(ns) + (1-m.latencyAlpha)*float64(old))
|
|
if m.writeLatencyNs.CompareAndSwap(old, newVal) {
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// SetMemoryTarget sets the target memory limit in bytes.
|
|
func (m *BadgerMonitor) SetMemoryTarget(bytes uint64) {
|
|
m.targetMemoryBytes.Store(bytes)
|
|
}
|
|
|
|
// Start begins background metric collection.
|
|
func (m *BadgerMonitor) Start() <-chan struct{} {
|
|
go m.collectLoop()
|
|
return m.stopped
|
|
}
|
|
|
|
// Stop halts background metric collection.
|
|
func (m *BadgerMonitor) Stop() {
|
|
close(m.stopChan)
|
|
<-m.stopped
|
|
}
|
|
|
|
// collectLoop periodically collects metrics from Badger.
|
|
func (m *BadgerMonitor) collectLoop() {
|
|
defer close(m.stopped)
|
|
|
|
ticker := time.NewTicker(m.interval)
|
|
defer ticker.Stop()
|
|
|
|
for {
|
|
select {
|
|
case <-m.stopChan:
|
|
return
|
|
case <-ticker.C:
|
|
m.updateMetrics()
|
|
}
|
|
}
|
|
}
|
|
|
|
// updateMetrics collects current metrics from Badger and actual process memory.
|
|
func (m *BadgerMonitor) updateMetrics() {
|
|
if m.db == nil || m.db.IsClosed() {
|
|
return
|
|
}
|
|
|
|
metrics := loadmonitor.Metrics{
|
|
Timestamp: time.Now(),
|
|
}
|
|
|
|
// Use RSS-based memory pressure (actual physical memory, not Go runtime)
|
|
procMem := ReadProcessMemoryStats()
|
|
physicalMemBytes := procMem.PhysicalMemoryBytes()
|
|
metrics.PhysicalMemoryMB = physicalMemBytes / (1024 * 1024)
|
|
|
|
targetBytes := m.targetMemoryBytes.Load()
|
|
if targetBytes > 0 {
|
|
// Use actual physical memory (RSS - shared) for pressure calculation
|
|
metrics.MemoryPressure = float64(physicalMemBytes) / float64(targetBytes)
|
|
}
|
|
|
|
// Check emergency mode
|
|
emergencyThreshold := float64(m.emergencyThreshold.Load()) / 1000.0
|
|
forcedUntil := m.emergencyModeUntil.Load()
|
|
now := time.Now().UnixNano()
|
|
|
|
if forcedUntil > now {
|
|
// Still in forced emergency mode
|
|
metrics.InEmergencyMode = true
|
|
} else if metrics.MemoryPressure >= emergencyThreshold {
|
|
// Memory pressure exceeds emergency threshold
|
|
metrics.InEmergencyMode = true
|
|
if !m.inEmergencyMode.Load() {
|
|
log.W.F("⚠️ entering emergency mode: memory pressure %.1f%% >= threshold %.1f%%",
|
|
metrics.MemoryPressure*100, emergencyThreshold*100)
|
|
}
|
|
} else {
|
|
if m.inEmergencyMode.Load() {
|
|
log.I.F("✅ exiting emergency mode: memory pressure %.1f%% < threshold %.1f%%",
|
|
metrics.MemoryPressure*100, emergencyThreshold*100)
|
|
}
|
|
}
|
|
m.inEmergencyMode.Store(metrics.InEmergencyMode)
|
|
|
|
// Get Badger LSM tree information for write load
|
|
levels := m.db.Levels()
|
|
var l0Tables int
|
|
var maxScore float64
|
|
|
|
for _, level := range levels {
|
|
if level.Level == 0 {
|
|
l0Tables = level.NumTables
|
|
}
|
|
if level.Score > maxScore {
|
|
maxScore = level.Score
|
|
}
|
|
}
|
|
|
|
// Calculate write load based on L0 tables and compaction score
|
|
// L0 tables stall at NumLevelZeroTablesStall (default 16)
|
|
// We consider write pressure high when approaching that limit
|
|
const l0StallThreshold = 16
|
|
l0Load := float64(l0Tables) / float64(l0StallThreshold)
|
|
if l0Load > 1.0 {
|
|
l0Load = 1.0
|
|
}
|
|
|
|
// Compaction score > 1.0 means compaction is needed
|
|
// We blend L0 tables and compaction score for write load
|
|
compactionLoad := maxScore / 2.0 // Score of 2.0 = fully loaded
|
|
if compactionLoad > 1.0 {
|
|
compactionLoad = 1.0
|
|
}
|
|
|
|
// Mark compaction as pending if score is high
|
|
metrics.CompactionPending = maxScore > 1.5 || l0Tables > 10
|
|
|
|
// Blend: 60% L0 (immediate backpressure), 40% compaction score
|
|
metrics.WriteLoad = 0.6*l0Load + 0.4*compactionLoad
|
|
|
|
// Calculate read load from cache metrics
|
|
blockMetrics := m.db.BlockCacheMetrics()
|
|
indexMetrics := m.db.IndexCacheMetrics()
|
|
|
|
var blockHitRatio, indexHitRatio float64
|
|
if blockMetrics != nil {
|
|
blockHitRatio = blockMetrics.Ratio()
|
|
}
|
|
if indexMetrics != nil {
|
|
indexHitRatio = indexMetrics.Ratio()
|
|
}
|
|
|
|
// Average cache hit ratio (0 = no hits = high load, 1 = all hits = low load)
|
|
avgHitRatio := (blockHitRatio + indexHitRatio) / 2.0
|
|
|
|
// Invert: low hit ratio = high read load
|
|
// Use 0.5 as the threshold (below 50% hit ratio is concerning)
|
|
if avgHitRatio < 0.5 {
|
|
metrics.ReadLoad = 1.0 - avgHitRatio*2 // 0% hits = 1.0 load, 50% hits = 0.0 load
|
|
} else {
|
|
metrics.ReadLoad = 0 // Above 50% hit ratio = minimal load
|
|
}
|
|
|
|
// Store latencies
|
|
metrics.QueryLatency = time.Duration(m.queryLatencyNs.Load())
|
|
metrics.WriteLatency = time.Duration(m.writeLatencyNs.Load())
|
|
|
|
// Update cached metrics
|
|
m.metricsLock.Lock()
|
|
m.cachedMetrics = metrics
|
|
m.lastL0Tables = l0Tables
|
|
m.lastL0Score = maxScore
|
|
m.metricsLock.Unlock()
|
|
}
|
|
|
|
// GetL0Stats returns L0-specific statistics for debugging.
|
|
func (m *BadgerMonitor) GetL0Stats() (tables int, score float64) {
|
|
m.metricsLock.RLock()
|
|
defer m.metricsLock.RUnlock()
|
|
return m.lastL0Tables, m.lastL0Score
|
|
}
|