feat(cache): 实现 RoutingCache 和 StatsBuffer 优化数据库写入
- 新增 RoutingCache 组件,使用 sync.Map 缓存 Provider 和 Model - 新增 StatsBuffer 组件,使用 sync.Map + atomic.Int64 缓冲统计数据 - 扩展 StatsRepository.BatchUpdate 支持批量增量更新 - 改造 RoutingService/StatsService/ProviderService/ModelService 集成缓存 - 更新 usage-statistics spec,新增 routing-cache 和 stats-buffer spec - 新增单元测试覆盖缓存命中/失效/并发场景
This commit is contained in:
251
backend/internal/service/stats_buffer_test.go
Normal file
251
backend/internal/service/stats_buffer_test.go
Normal file
@@ -0,0 +1,251 @@
|
||||
package service
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"sync"
|
||||
"sync/atomic"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/stretchr/testify/assert"
|
||||
"go.uber.org/zap"
|
||||
|
||||
"nex/backend/internal/domain"
|
||||
)
|
||||
|
||||
type mockStatsRepo struct {
|
||||
records []struct {
|
||||
providerID string
|
||||
modelName string
|
||||
date time.Time
|
||||
delta int
|
||||
}
|
||||
fail bool
|
||||
mu sync.Mutex
|
||||
}
|
||||
|
||||
func (m *mockStatsRepo) Record(providerID, modelName string) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *mockStatsRepo) BatchUpdate(providerID, modelName string, date time.Time, delta int) error {
|
||||
m.mu.Lock()
|
||||
defer m.mu.Unlock()
|
||||
if m.fail {
|
||||
return errors.New("db error")
|
||||
}
|
||||
m.records = append(m.records, struct {
|
||||
providerID string
|
||||
modelName string
|
||||
date time.Time
|
||||
delta int
|
||||
}{providerID, modelName, date, delta})
|
||||
return nil
|
||||
}
|
||||
|
||||
func (m *mockStatsRepo) Query(providerID, modelName string, startDate, endDate *time.Time) ([]domain.UsageStats, error) {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
func TestStatsBuffer_Increment(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{}
|
||||
buffer := NewStatsBuffer(statsRepo, logger)
|
||||
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
buffer.Increment("openai", "gpt-3.5")
|
||||
|
||||
var count int64
|
||||
buffer.counters.Range(func(key, value interface{}) bool {
|
||||
counter := value.(*int64)
|
||||
count += atomic.LoadInt64(counter)
|
||||
return true
|
||||
})
|
||||
assert.Equal(t, int64(3), count)
|
||||
}
|
||||
|
||||
func TestStatsBuffer_ConcurrentIncrement(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{}
|
||||
buffer := NewStatsBuffer(statsRepo, logger)
|
||||
|
||||
var wg sync.WaitGroup
|
||||
for i := 0; i < 100; i++ {
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
}()
|
||||
}
|
||||
wg.Wait()
|
||||
|
||||
var count int64
|
||||
buffer.counters.Range(func(key, value interface{}) bool {
|
||||
counter := value.(*int64)
|
||||
count = atomic.LoadInt64(counter)
|
||||
return true
|
||||
})
|
||||
assert.Equal(t, int64(100), count)
|
||||
}
|
||||
|
||||
func TestStatsBuffer_LoadOrStore(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{}
|
||||
buffer := NewStatsBuffer(statsRepo, logger)
|
||||
|
||||
var wg sync.WaitGroup
|
||||
for i := 0; i < 10; i++ {
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
}()
|
||||
}
|
||||
wg.Wait()
|
||||
|
||||
var counterCount int
|
||||
buffer.counters.Range(func(key, value interface{}) bool {
|
||||
counterCount++
|
||||
return true
|
||||
})
|
||||
assert.Equal(t, 1, counterCount)
|
||||
}
|
||||
|
||||
func TestStatsBuffer_FlushByInterval(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{}
|
||||
buffer := NewStatsBuffer(statsRepo, logger,
|
||||
WithFlushInterval(100*time.Millisecond))
|
||||
|
||||
buffer.Start()
|
||||
defer buffer.Stop()
|
||||
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
|
||||
time.Sleep(200 * time.Millisecond)
|
||||
|
||||
statsRepo.mu.Lock()
|
||||
assert.GreaterOrEqual(t, len(statsRepo.records), 1)
|
||||
statsRepo.mu.Unlock()
|
||||
}
|
||||
|
||||
func TestStatsBuffer_FlushByThreshold(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{}
|
||||
buffer := NewStatsBuffer(statsRepo, logger,
|
||||
WithFlushThreshold(10))
|
||||
|
||||
buffer.Start()
|
||||
defer buffer.Stop()
|
||||
|
||||
for i := 0; i < 10; i++ {
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
}
|
||||
|
||||
time.Sleep(50 * time.Millisecond)
|
||||
|
||||
statsRepo.mu.Lock()
|
||||
assert.GreaterOrEqual(t, len(statsRepo.records), 1)
|
||||
statsRepo.mu.Unlock()
|
||||
}
|
||||
|
||||
func TestStatsBuffer_SwapInt64(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{}
|
||||
buffer := NewStatsBuffer(statsRepo, logger)
|
||||
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
|
||||
var beforeCount int64
|
||||
buffer.counters.Range(func(key, value interface{}) bool {
|
||||
counter := value.(*int64)
|
||||
beforeCount = atomic.LoadInt64(counter)
|
||||
return true
|
||||
})
|
||||
assert.Equal(t, int64(2), beforeCount)
|
||||
|
||||
buffer.flush()
|
||||
|
||||
var afterCount int64
|
||||
buffer.counters.Range(func(key, value interface{}) bool {
|
||||
counter := value.(*int64)
|
||||
afterCount = atomic.LoadInt64(counter)
|
||||
return true
|
||||
})
|
||||
assert.Equal(t, int64(0), afterCount)
|
||||
}
|
||||
|
||||
func TestStatsBuffer_FailRetry(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{fail: true}
|
||||
buffer := NewStatsBuffer(statsRepo, logger)
|
||||
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
|
||||
buffer.flush()
|
||||
|
||||
var count int64
|
||||
buffer.counters.Range(func(key, value interface{}) bool {
|
||||
counter := value.(*int64)
|
||||
count = atomic.LoadInt64(counter)
|
||||
return true
|
||||
})
|
||||
assert.Equal(t, int64(2), count)
|
||||
}
|
||||
|
||||
func TestStatsBuffer_Stop(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{}
|
||||
buffer := NewStatsBuffer(statsRepo, logger,
|
||||
WithFlushInterval(10*time.Second))
|
||||
|
||||
buffer.Start()
|
||||
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
|
||||
start := time.Now()
|
||||
buffer.Stop()
|
||||
elapsed := time.Since(start)
|
||||
|
||||
assert.Less(t, elapsed, 1*time.Second)
|
||||
|
||||
statsRepo.mu.Lock()
|
||||
assert.GreaterOrEqual(t, len(statsRepo.records), 1)
|
||||
statsRepo.mu.Unlock()
|
||||
}
|
||||
|
||||
func TestStatsBuffer_ConcurrentIncrementAndFlush(t *testing.T) {
|
||||
logger := zap.NewNop()
|
||||
statsRepo := &mockStatsRepo{}
|
||||
buffer := NewStatsBuffer(statsRepo, logger,
|
||||
WithFlushInterval(50*time.Millisecond))
|
||||
|
||||
buffer.Start()
|
||||
defer buffer.Stop()
|
||||
|
||||
var wg sync.WaitGroup
|
||||
for i := 0; i < 100; i++ {
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
defer wg.Done()
|
||||
buffer.Increment("openai", "gpt-4")
|
||||
}()
|
||||
}
|
||||
wg.Wait()
|
||||
|
||||
time.Sleep(100 * time.Millisecond)
|
||||
|
||||
statsRepo.mu.Lock()
|
||||
totalDelta := 0
|
||||
for _, r := range statsRepo.records {
|
||||
totalDelta += r.delta
|
||||
}
|
||||
statsRepo.mu.Unlock()
|
||||
|
||||
assert.Equal(t, 100, totalDelta)
|
||||
}
|
||||
Reference in New Issue
Block a user