- 移除 OpenAIGatewayHandler 未使用字段 - 删除并发缓存中未使用的 Redis 脚本常量 - 将仅供 unit 测试使用的 parseIntegralNumber 移入 unit build tag 文件
429 lines
13 KiB
Go
429 lines
13 KiB
Go
package repository
|
||
|
||
import (
|
||
"context"
|
||
"errors"
|
||
"fmt"
|
||
"strconv"
|
||
|
||
"github.com/Wei-Shaw/sub2api/internal/service"
|
||
"github.com/redis/go-redis/v9"
|
||
)
|
||
|
||
// 并发控制缓存常量定义
|
||
//
|
||
// 性能优化说明:
|
||
// 原实现使用 SCAN 命令遍历独立的槽位键(concurrency:account:{id}:{requestID}),
|
||
// 在高并发场景下 SCAN 需要多次往返,且遍历大量键时性能下降明显。
|
||
//
|
||
// 新实现改用 Redis 有序集合(Sorted Set):
|
||
// 1. 每个账号/用户只有一个键,成员为 requestID,分数为时间戳
|
||
// 2. 使用 ZCARD 原子获取并发数,时间复杂度 O(1)
|
||
// 3. 使用 ZREMRANGEBYSCORE 清理过期槽位,避免手动管理 TTL
|
||
// 4. 单次 Redis 调用完成计数,减少网络往返
|
||
const (
|
||
// 并发槽位键前缀(有序集合)
|
||
// 格式: concurrency:account:{accountID}
|
||
accountSlotKeyPrefix = "concurrency:account:"
|
||
// 格式: concurrency:user:{userID}
|
||
userSlotKeyPrefix = "concurrency:user:"
|
||
// 等待队列计数器格式: concurrency:wait:{userID}
|
||
waitQueueKeyPrefix = "concurrency:wait:"
|
||
// 账号级等待队列计数器格式: wait:account:{accountID}
|
||
accountWaitKeyPrefix = "wait:account:"
|
||
|
||
// 默认槽位过期时间(分钟),可通过配置覆盖
|
||
defaultSlotTTLMinutes = 15
|
||
)
|
||
|
||
var (
|
||
// acquireScript 使用有序集合计数并在未达上限时添加槽位
|
||
// 使用 Redis TIME 命令获取服务器时间,避免多实例时钟不同步问题
|
||
// KEYS[1] = 有序集合键 (concurrency:account:{id} / concurrency:user:{id})
|
||
// ARGV[1] = maxConcurrency
|
||
// ARGV[2] = TTL(秒)
|
||
// ARGV[3] = requestID
|
||
acquireScript = redis.NewScript(`
|
||
local key = KEYS[1]
|
||
local maxConcurrency = tonumber(ARGV[1])
|
||
local ttl = tonumber(ARGV[2])
|
||
local requestID = ARGV[3]
|
||
|
||
-- 使用 Redis 服务器时间,确保多实例时钟一致
|
||
local timeResult = redis.call('TIME')
|
||
local now = tonumber(timeResult[1])
|
||
local expireBefore = now - ttl
|
||
|
||
-- 清理过期槽位
|
||
redis.call('ZREMRANGEBYSCORE', key, '-inf', expireBefore)
|
||
|
||
-- 检查是否已存在(支持重试场景刷新时间戳)
|
||
local exists = redis.call('ZSCORE', key, requestID)
|
||
if exists ~= false then
|
||
redis.call('ZADD', key, now, requestID)
|
||
redis.call('EXPIRE', key, ttl)
|
||
return 1
|
||
end
|
||
|
||
-- 检查是否达到并发上限
|
||
local count = redis.call('ZCARD', key)
|
||
if count < maxConcurrency then
|
||
redis.call('ZADD', key, now, requestID)
|
||
redis.call('EXPIRE', key, ttl)
|
||
return 1
|
||
end
|
||
|
||
return 0
|
||
`)
|
||
|
||
// getCountScript 统计有序集合中的槽位数量并清理过期条目
|
||
// 使用 Redis TIME 命令获取服务器时间
|
||
// KEYS[1] = 有序集合键
|
||
// ARGV[1] = TTL(秒)
|
||
getCountScript = redis.NewScript(`
|
||
local key = KEYS[1]
|
||
local ttl = tonumber(ARGV[1])
|
||
|
||
-- 使用 Redis 服务器时间
|
||
local timeResult = redis.call('TIME')
|
||
local now = tonumber(timeResult[1])
|
||
local expireBefore = now - ttl
|
||
|
||
redis.call('ZREMRANGEBYSCORE', key, '-inf', expireBefore)
|
||
return redis.call('ZCARD', key)
|
||
`)
|
||
|
||
// incrementWaitScript - refreshes TTL on each increment to keep queue depth accurate
|
||
// KEYS[1] = wait queue key
|
||
// ARGV[1] = maxWait
|
||
// ARGV[2] = TTL in seconds
|
||
incrementWaitScript = redis.NewScript(`
|
||
local current = redis.call('GET', KEYS[1])
|
||
if current == false then
|
||
current = 0
|
||
else
|
||
current = tonumber(current)
|
||
end
|
||
|
||
if current >= tonumber(ARGV[1]) then
|
||
return 0
|
||
end
|
||
|
||
local newVal = redis.call('INCR', KEYS[1])
|
||
|
||
-- Refresh TTL so long-running traffic doesn't expire active queue counters.
|
||
redis.call('EXPIRE', KEYS[1], ARGV[2])
|
||
|
||
return 1
|
||
`)
|
||
|
||
// incrementAccountWaitScript - account-level wait queue count (refresh TTL on each increment)
|
||
incrementAccountWaitScript = redis.NewScript(`
|
||
local current = redis.call('GET', KEYS[1])
|
||
if current == false then
|
||
current = 0
|
||
else
|
||
current = tonumber(current)
|
||
end
|
||
|
||
if current >= tonumber(ARGV[1]) then
|
||
return 0
|
||
end
|
||
|
||
local newVal = redis.call('INCR', KEYS[1])
|
||
|
||
-- Refresh TTL so long-running traffic doesn't expire active queue counters.
|
||
redis.call('EXPIRE', KEYS[1], ARGV[2])
|
||
|
||
return 1
|
||
`)
|
||
|
||
// decrementWaitScript - same as before
|
||
decrementWaitScript = redis.NewScript(`
|
||
local current = redis.call('GET', KEYS[1])
|
||
if current ~= false and tonumber(current) > 0 then
|
||
redis.call('DECR', KEYS[1])
|
||
end
|
||
return 1
|
||
`)
|
||
|
||
// cleanupExpiredSlotsScript - remove expired slots
|
||
// KEYS[1] = concurrency:account:{accountID}
|
||
// ARGV[1] = TTL (seconds)
|
||
cleanupExpiredSlotsScript = redis.NewScript(`
|
||
local key = KEYS[1]
|
||
local ttl = tonumber(ARGV[1])
|
||
local timeResult = redis.call('TIME')
|
||
local now = tonumber(timeResult[1])
|
||
local expireBefore = now - ttl
|
||
return redis.call('ZREMRANGEBYSCORE', key, '-inf', expireBefore)
|
||
`)
|
||
)
|
||
|
||
type concurrencyCache struct {
|
||
rdb *redis.Client
|
||
slotTTLSeconds int // 槽位过期时间(秒)
|
||
waitQueueTTLSeconds int // 等待队列过期时间(秒)
|
||
}
|
||
|
||
// NewConcurrencyCache 创建并发控制缓存
|
||
// slotTTLMinutes: 槽位过期时间(分钟),0 或负数使用默认值 15 分钟
|
||
// waitQueueTTLSeconds: 等待队列过期时间(秒),0 或负数使用 slot TTL
|
||
func NewConcurrencyCache(rdb *redis.Client, slotTTLMinutes int, waitQueueTTLSeconds int) service.ConcurrencyCache {
|
||
if slotTTLMinutes <= 0 {
|
||
slotTTLMinutes = defaultSlotTTLMinutes
|
||
}
|
||
if waitQueueTTLSeconds <= 0 {
|
||
waitQueueTTLSeconds = slotTTLMinutes * 60
|
||
}
|
||
return &concurrencyCache{
|
||
rdb: rdb,
|
||
slotTTLSeconds: slotTTLMinutes * 60,
|
||
waitQueueTTLSeconds: waitQueueTTLSeconds,
|
||
}
|
||
}
|
||
|
||
// Helper functions for key generation
|
||
func accountSlotKey(accountID int64) string {
|
||
return fmt.Sprintf("%s%d", accountSlotKeyPrefix, accountID)
|
||
}
|
||
|
||
func userSlotKey(userID int64) string {
|
||
return fmt.Sprintf("%s%d", userSlotKeyPrefix, userID)
|
||
}
|
||
|
||
func waitQueueKey(userID int64) string {
|
||
return fmt.Sprintf("%s%d", waitQueueKeyPrefix, userID)
|
||
}
|
||
|
||
func accountWaitKey(accountID int64) string {
|
||
return fmt.Sprintf("%s%d", accountWaitKeyPrefix, accountID)
|
||
}
|
||
|
||
// Account slot operations
|
||
|
||
func (c *concurrencyCache) AcquireAccountSlot(ctx context.Context, accountID int64, maxConcurrency int, requestID string) (bool, error) {
|
||
key := accountSlotKey(accountID)
|
||
// 时间戳在 Lua 脚本内使用 Redis TIME 命令获取,确保多实例时钟一致
|
||
result, err := acquireScript.Run(ctx, c.rdb, []string{key}, maxConcurrency, c.slotTTLSeconds, requestID).Int()
|
||
if err != nil {
|
||
return false, err
|
||
}
|
||
return result == 1, nil
|
||
}
|
||
|
||
func (c *concurrencyCache) ReleaseAccountSlot(ctx context.Context, accountID int64, requestID string) error {
|
||
key := accountSlotKey(accountID)
|
||
return c.rdb.ZRem(ctx, key, requestID).Err()
|
||
}
|
||
|
||
func (c *concurrencyCache) GetAccountConcurrency(ctx context.Context, accountID int64) (int, error) {
|
||
key := accountSlotKey(accountID)
|
||
// 时间戳在 Lua 脚本内使用 Redis TIME 命令获取
|
||
result, err := getCountScript.Run(ctx, c.rdb, []string{key}, c.slotTTLSeconds).Int()
|
||
if err != nil {
|
||
return 0, err
|
||
}
|
||
return result, nil
|
||
}
|
||
|
||
// User slot operations
|
||
|
||
func (c *concurrencyCache) AcquireUserSlot(ctx context.Context, userID int64, maxConcurrency int, requestID string) (bool, error) {
|
||
key := userSlotKey(userID)
|
||
// 时间戳在 Lua 脚本内使用 Redis TIME 命令获取,确保多实例时钟一致
|
||
result, err := acquireScript.Run(ctx, c.rdb, []string{key}, maxConcurrency, c.slotTTLSeconds, requestID).Int()
|
||
if err != nil {
|
||
return false, err
|
||
}
|
||
return result == 1, nil
|
||
}
|
||
|
||
func (c *concurrencyCache) ReleaseUserSlot(ctx context.Context, userID int64, requestID string) error {
|
||
key := userSlotKey(userID)
|
||
return c.rdb.ZRem(ctx, key, requestID).Err()
|
||
}
|
||
|
||
func (c *concurrencyCache) GetUserConcurrency(ctx context.Context, userID int64) (int, error) {
|
||
key := userSlotKey(userID)
|
||
// 时间戳在 Lua 脚本内使用 Redis TIME 命令获取
|
||
result, err := getCountScript.Run(ctx, c.rdb, []string{key}, c.slotTTLSeconds).Int()
|
||
if err != nil {
|
||
return 0, err
|
||
}
|
||
return result, nil
|
||
}
|
||
|
||
// Wait queue operations
|
||
|
||
func (c *concurrencyCache) IncrementWaitCount(ctx context.Context, userID int64, maxWait int) (bool, error) {
|
||
key := waitQueueKey(userID)
|
||
result, err := incrementWaitScript.Run(ctx, c.rdb, []string{key}, maxWait, c.waitQueueTTLSeconds).Int()
|
||
if err != nil {
|
||
return false, err
|
||
}
|
||
return result == 1, nil
|
||
}
|
||
|
||
func (c *concurrencyCache) DecrementWaitCount(ctx context.Context, userID int64) error {
|
||
key := waitQueueKey(userID)
|
||
_, err := decrementWaitScript.Run(ctx, c.rdb, []string{key}).Result()
|
||
return err
|
||
}
|
||
|
||
// Account wait queue operations
|
||
|
||
func (c *concurrencyCache) IncrementAccountWaitCount(ctx context.Context, accountID int64, maxWait int) (bool, error) {
|
||
key := accountWaitKey(accountID)
|
||
result, err := incrementAccountWaitScript.Run(ctx, c.rdb, []string{key}, maxWait, c.waitQueueTTLSeconds).Int()
|
||
if err != nil {
|
||
return false, err
|
||
}
|
||
return result == 1, nil
|
||
}
|
||
|
||
func (c *concurrencyCache) DecrementAccountWaitCount(ctx context.Context, accountID int64) error {
|
||
key := accountWaitKey(accountID)
|
||
_, err := decrementWaitScript.Run(ctx, c.rdb, []string{key}).Result()
|
||
return err
|
||
}
|
||
|
||
func (c *concurrencyCache) GetAccountWaitingCount(ctx context.Context, accountID int64) (int, error) {
|
||
key := accountWaitKey(accountID)
|
||
val, err := c.rdb.Get(ctx, key).Int()
|
||
if err != nil && !errors.Is(err, redis.Nil) {
|
||
return 0, err
|
||
}
|
||
if errors.Is(err, redis.Nil) {
|
||
return 0, nil
|
||
}
|
||
return val, nil
|
||
}
|
||
|
||
func (c *concurrencyCache) GetAccountsLoadBatch(ctx context.Context, accounts []service.AccountWithConcurrency) (map[int64]*service.AccountLoadInfo, error) {
|
||
if len(accounts) == 0 {
|
||
return map[int64]*service.AccountLoadInfo{}, nil
|
||
}
|
||
|
||
// 使用 Pipeline 替代 Lua 脚本,兼容 Redis Cluster(Lua 内动态拼 key 会 CROSSSLOT)。
|
||
// 每个账号执行 3 个命令:ZREMRANGEBYSCORE(清理过期)、ZCARD(并发数)、GET(等待数)。
|
||
now, err := c.rdb.Time(ctx).Result()
|
||
if err != nil {
|
||
return nil, fmt.Errorf("redis TIME: %w", err)
|
||
}
|
||
cutoffTime := now.Unix() - int64(c.slotTTLSeconds)
|
||
|
||
pipe := c.rdb.Pipeline()
|
||
|
||
type accountCmds struct {
|
||
id int64
|
||
maxConcurrency int
|
||
zcardCmd *redis.IntCmd
|
||
getCmd *redis.StringCmd
|
||
}
|
||
cmds := make([]accountCmds, 0, len(accounts))
|
||
for _, acc := range accounts {
|
||
slotKey := accountSlotKeyPrefix + strconv.FormatInt(acc.ID, 10)
|
||
waitKey := accountWaitKeyPrefix + strconv.FormatInt(acc.ID, 10)
|
||
pipe.ZRemRangeByScore(ctx, slotKey, "-inf", strconv.FormatInt(cutoffTime, 10))
|
||
ac := accountCmds{
|
||
id: acc.ID,
|
||
maxConcurrency: acc.MaxConcurrency,
|
||
zcardCmd: pipe.ZCard(ctx, slotKey),
|
||
getCmd: pipe.Get(ctx, waitKey),
|
||
}
|
||
cmds = append(cmds, ac)
|
||
}
|
||
|
||
if _, err := pipe.Exec(ctx); err != nil && !errors.Is(err, redis.Nil) {
|
||
return nil, fmt.Errorf("pipeline exec: %w", err)
|
||
}
|
||
|
||
loadMap := make(map[int64]*service.AccountLoadInfo, len(accounts))
|
||
for _, ac := range cmds {
|
||
currentConcurrency := int(ac.zcardCmd.Val())
|
||
waitingCount := 0
|
||
if v, err := ac.getCmd.Int(); err == nil {
|
||
waitingCount = v
|
||
}
|
||
loadRate := 0
|
||
if ac.maxConcurrency > 0 {
|
||
loadRate = (currentConcurrency + waitingCount) * 100 / ac.maxConcurrency
|
||
}
|
||
loadMap[ac.id] = &service.AccountLoadInfo{
|
||
AccountID: ac.id,
|
||
CurrentConcurrency: currentConcurrency,
|
||
WaitingCount: waitingCount,
|
||
LoadRate: loadRate,
|
||
}
|
||
}
|
||
|
||
return loadMap, nil
|
||
}
|
||
|
||
func (c *concurrencyCache) GetUsersLoadBatch(ctx context.Context, users []service.UserWithConcurrency) (map[int64]*service.UserLoadInfo, error) {
|
||
if len(users) == 0 {
|
||
return map[int64]*service.UserLoadInfo{}, nil
|
||
}
|
||
|
||
// 使用 Pipeline 替代 Lua 脚本,兼容 Redis Cluster。
|
||
now, err := c.rdb.Time(ctx).Result()
|
||
if err != nil {
|
||
return nil, fmt.Errorf("redis TIME: %w", err)
|
||
}
|
||
cutoffTime := now.Unix() - int64(c.slotTTLSeconds)
|
||
|
||
pipe := c.rdb.Pipeline()
|
||
|
||
type userCmds struct {
|
||
id int64
|
||
maxConcurrency int
|
||
zcardCmd *redis.IntCmd
|
||
getCmd *redis.StringCmd
|
||
}
|
||
cmds := make([]userCmds, 0, len(users))
|
||
for _, u := range users {
|
||
slotKey := userSlotKeyPrefix + strconv.FormatInt(u.ID, 10)
|
||
waitKey := waitQueueKeyPrefix + strconv.FormatInt(u.ID, 10)
|
||
pipe.ZRemRangeByScore(ctx, slotKey, "-inf", strconv.FormatInt(cutoffTime, 10))
|
||
uc := userCmds{
|
||
id: u.ID,
|
||
maxConcurrency: u.MaxConcurrency,
|
||
zcardCmd: pipe.ZCard(ctx, slotKey),
|
||
getCmd: pipe.Get(ctx, waitKey),
|
||
}
|
||
cmds = append(cmds, uc)
|
||
}
|
||
|
||
if _, err := pipe.Exec(ctx); err != nil && !errors.Is(err, redis.Nil) {
|
||
return nil, fmt.Errorf("pipeline exec: %w", err)
|
||
}
|
||
|
||
loadMap := make(map[int64]*service.UserLoadInfo, len(users))
|
||
for _, uc := range cmds {
|
||
currentConcurrency := int(uc.zcardCmd.Val())
|
||
waitingCount := 0
|
||
if v, err := uc.getCmd.Int(); err == nil {
|
||
waitingCount = v
|
||
}
|
||
loadRate := 0
|
||
if uc.maxConcurrency > 0 {
|
||
loadRate = (currentConcurrency + waitingCount) * 100 / uc.maxConcurrency
|
||
}
|
||
loadMap[uc.id] = &service.UserLoadInfo{
|
||
UserID: uc.id,
|
||
CurrentConcurrency: currentConcurrency,
|
||
WaitingCount: waitingCount,
|
||
LoadRate: loadRate,
|
||
}
|
||
}
|
||
|
||
return loadMap, nil
|
||
}
|
||
|
||
func (c *concurrencyCache) CleanupExpiredAccountSlots(ctx context.Context, accountID int64) error {
|
||
key := accountSlotKey(accountID)
|
||
_, err := cleanupExpiredSlotsScript.Run(ctx, c.rdb, []string{key}, c.slotTTLSeconds).Result()
|
||
return err
|
||
}
|