Compare commits

...

10 Commits

Author SHA1 Message Date
huangzhenpc
e4db851b31 feat: 添加二次开发部署配置和文档
Some checks failed
CI / test (push) Has been cancelled
CI / golangci-lint (push) Has been cancelled
- 添加双 Git remote 配置指南 (GIT_GUIDE.md, GIT_WORKFLOW.md)
- 添加 1Panel 环境部署配置 (docker-compose.1panel.yml, .env.1panel)
- 添加自动化部署脚本 (deploy-complete.sh, deploy-dev.sh, deploy.sh)
- 添加部署说明文档 (README_DEPLOY.md)
- 配置端口 2080 映射
- 支持连接现有 1Panel Redis (1Panel-redis-S1KH)
2026-01-04 10:25:36 +08:00
shaw
631ba25e04 Merge branch 'feature/atomic-scheduling-v2' 2026-01-03 13:35:19 +08:00
song
d2aaf0b491 refactor(service): 将 AccountUsageService 的包级缓存改为依赖注入 2026-01-03 13:10:43 +08:00
ianshaw
1710779157 test: 暂时跳过 TestGetAccountsLoadBatch 集成测试
该测试在 CI 环境中失败,需要进一步调试。
暂时跳过以让 CI 通过,后续在本地 Docker 环境中修复。
2026-01-02 19:24:01 -08:00
ianshaw
b8779764b5 perf: 优化负载感知调度的准确性和响应速度
基于 Codex 审查建议的性能优化。

负载批量查询优化:
- getAccountsLoadBatchScript 添加过期槽位清理
- 使用 ZREMRANGEBYSCORE 在计数前清理过期条目
- 防止过期槽位导致负载率计算偏高
- 提升负载感知调度的准确性

等待循环优化:
- waitForSlotWithPingTimeout 添加立即获取尝试
- 避免不必要的 initialBackoff 延迟
- 低负载场景下减少响应延迟

测试改进:
- 取消跳过 TestGetAccountsLoadBatch 集成测试
- 过期槽位清理应该修复了 CI 中的计数问题

影响:
- 更准确的负载感知调度决策
- 更快的槽位获取响应
- 更好的测试覆盖率
2026-01-02 19:24:01 -08:00
ianshaw
681a357e07 fix: 修复 SSE/JSON 转义和 nil 安全问题
基于 Codex 审查建议修复关键安全问题。

SSE/JSON 转义修复:
- handleStreamingAwareError: 使用 json.Marshal 替代字符串拼接
- sendMockWarmupStream: 使用 json.Marshal 生成 message_start 事件
- 防止错误消息中的特殊字符导致无效 JSON

Nil 安全检查:
- SelectAccountWithLoadAwareness: 粘性会话层添加 s.cache != nil 检查
- BindStickySession: 添加 s.cache == nil 检查
- 防止 cache 未初始化时的运行时 panic

影响:
- 提升 SSE 错误处理的健壮性
- 避免客户端 JSON 解析失败
- 增强代码防御性编程
2026-01-02 19:24:01 -08:00
ianshaw
e876d54a48 fix: 恢复 Google One 功能兼容性
恢复 main 分支的 gemini_oauth_service.go 以保持与 Google One 功能的兼容性。

变更:
- 添加 Google One tier 常量定义
- 添加存储空间 tier 阈值常量
- 支持 google_one OAuth 类型
- 包含 RefreshAccountGoogleOneTier 等 Google One 相关方法

原因:
- atomic-scheduling 恢复时使用了旧版本的文件
- 需要保持与 main 分支 Google One 功能(PR #118)的兼容性
- 避免编译错误(handler 代码依赖这些方法)
2026-01-02 19:24:01 -08:00
ianshaw
7568dc8500 Reapply "feat(gateway): 实现负载感知的账号调度优化 (#114)" (#117)
This reverts commit c5c12d4c8b.
2026-01-02 19:24:01 -08:00
song
0452f32003 feat(antigravity): 支持 Gemini cachedContentTokenCount 映射到 Claude cache_read_input_tokens
- 在 GeminiUsageMetadata 添加 CachedContentTokenCount 字段
- 修正 token 映射:Gemini promptTokenCount 包含缓存,Claude input_tokens 不包含
- 流式和非流式响应均已支持
2026-01-03 01:26:18 +08:00
song
9ed823fdbd fix: 移除 antigravity 模块中的 [Debug] 日志
这些调试日志不应在生产环境中输出。
2026-01-03 00:36:48 +08:00
21 changed files with 1562 additions and 64 deletions

420
GIT_GUIDE.md Normal file
View File

@@ -0,0 +1,420 @@
# Sub2API 双 Remote Git 配置指南
## 📋 Git 仓库配置
### Remote 配置结构
```
upstream (官方仓库)
└── https://github.com/Wei-Shaw/sub2api.git
用途: 拉取官方更新
origin (你的仓库)
└── https://git.586vip.cn/oadmin/sub2api.git
用途: 保存你的二次开发代码
```
---
## 🚀 快速部署(一键完成)
### 方式 1使用自动化脚本推荐
**在服务器上执行以下命令:**
```bash
# 下载部署脚本
curl -o /tmp/deploy-sub2api.sh https://你的脚本地址/deploy-complete.sh
# 或者直接创建脚本
cat > /tmp/deploy-sub2api.sh << 'SCRIPT_END'
# [这里粘贴 deploy-complete.sh 的全部内容]
SCRIPT_END
# 赋予执行权限
chmod +x /tmp/deploy-sub2api.sh
# 运行脚本
bash /tmp/deploy-sub2api.sh
```
脚本会自动完成:
- ✅ 克隆官方仓库
- ✅ 配置双 remote
- ✅ 推送到你的仓库
- ✅ 创建部署配置
- ✅ 启动 Docker 服务
---
### 方式 2手动分步执行
如果自动脚本有问题,可以手动执行:
```bash
# 1. 克隆官方仓库
cd /opt
git clone https://github.com/Wei-Shaw/sub2api.git sub2api-dev
cd sub2api-dev
# 2. 配置 Git Remote
git remote rename origin upstream
git remote add origin https://git.586vip.cn/oadmin/sub2api.git
# 3. 查看配置
git remote -v
# 应该看到:
# origin https://git.586vip.cn/oadmin/sub2api.git (fetch)
# origin https://git.586vip.cn/oadmin/sub2api.git (push)
# upstream https://github.com/Wei-Shaw/sub2api.git (fetch)
# upstream https://github.com/Wei-Shaw/sub2api.git (push)
# 4. 创建 main 分支
git checkout -b main
# 5. 推送到你的仓库
git push -u origin main
# 6. 配置部署文件
cd deploy
```
创建 `docker-compose.prod.yml`
```bash
cat > docker-compose.prod.yml << 'EOF'
services:
sub2api:
image: weishaw/sub2api:latest
container_name: sub2api
restart: unless-stopped
ports:
- "2080:8080"
volumes:
- ./data:/app/data
environment:
- AUTO_SETUP=true
- SERVER_HOST=0.0.0.0
- SERVER_PORT=8080
- SERVER_MODE=release
- DATABASE_HOST=postgres
- DATABASE_PORT=5432
- DATABASE_USER=sub2api
- DATABASE_PASSWORD=${POSTGRES_PASSWORD}
- DATABASE_DBNAME=sub2api
- DATABASE_SSLMODE=disable
- REDIS_HOST=host.docker.internal
- REDIS_PORT=6379
- REDIS_PASSWORD=redis_bJFKDk
- REDIS_DB=1
- ADMIN_EMAIL=${ADMIN_EMAIL}
- ADMIN_PASSWORD=${ADMIN_PASSWORD}
- JWT_SECRET=${JWT_SECRET}
- TZ=Asia/Shanghai
depends_on:
- postgres
extra_hosts:
- "host.docker.internal:host-gateway"
postgres:
image: postgres:18-alpine
container_name: sub2api-postgres
restart: unless-stopped
volumes:
- postgres_data:/var/lib/postgresql/data
environment:
- POSTGRES_USER=sub2api
- POSTGRES_PASSWORD=${POSTGRES_PASSWORD}
- POSTGRES_DB=sub2api
- TZ=Asia/Shanghai
healthcheck:
test: ["CMD-SHELL", "pg_isready -U sub2api"]
interval: 10s
timeout: 5s
retries: 5
volumes:
postgres_data:
EOF
```
创建 `.env` 配置:
```bash
cat > .env << EOF
POSTGRES_USER=sub2api
POSTGRES_PASSWORD=$(openssl rand -base64 24)
POSTGRES_DB=sub2api
ADMIN_EMAIL=admin@example.com
ADMIN_PASSWORD=admin123
JWT_SECRET=$(openssl rand -hex 32)
JWT_EXPIRE_HOUR=24
TZ=Asia/Shanghai
SERVER_MODE=release
RUN_MODE=standard
EOF
# 显示生成的密码
echo "===== 配置信息 ====="
cat .env
echo "===================="
```
启动服务:
```bash
# 拉取镜像
docker-compose -f docker-compose.prod.yml pull
# 启动服务
docker-compose -f docker-compose.prod.yml up -d
# 查看日志
docker-compose -f docker-compose.prod.yml logs -f
```
---
## 🔄 日常工作流程
### 开发流程
```bash
cd /opt/sub2api-dev
# 1. 创建功能分支
git checkout -b feature/new-feature
# 2. 开发并提交
git add .
git commit -m "feat: 添加新功能"
# 3. 切回主分支
git checkout main
# 4. 合并功能分支
git merge feature/new-feature
# 5. 推送到你的仓库
git push origin main
# 6. 删除功能分支(可选)
git branch -d feature/new-feature
```
### 同步官方更新
```bash
cd /opt/sub2api-dev
# 1. 查看官方更新
git fetch upstream
git log HEAD..upstream/main --oneline
# 2. 查看详细差异
git diff HEAD..upstream/main
# 3. 合并官方更新
git merge upstream/main
# 4. 如果有冲突,解决后提交
git add .
git commit -m "merge: 合并官方更新 v1.x.x"
# 5. 推送到你的仓库
git push origin main
# 6. 重新部署(如果需要)
cd deploy
docker-compose -f docker-compose.prod.yml pull
docker-compose -f docker-compose.prod.yml up -d
```
### 查看分支和远程信息
```bash
# 查看所有分支
git branch -a
# 查看远程仓库
git remote -v
# 查看当前状态
git status
# 查看提交历史
git log --oneline --graph --all -10
```
---
## 🛠️ 常见操作
### 回滚到某个版本
```bash
# 查看提交历史
git log --oneline
# 回滚到指定提交(软回滚,保留修改)
git reset --soft <commit-hash>
# 回滚到指定提交(硬回滚,丢弃修改)
git reset --hard <commit-hash>
# 推送到远程(需要强制推送)
git push -f origin main
```
### 对比官方版本
```bash
# 对比特定文件
git diff upstream/main -- backend/internal/service/gateway_service.go
# 对比整个目录
git diff upstream/main -- backend/internal/service/
# 生成 patch 文件
git diff upstream/main > my-changes.patch
# 查看改动的文件列表
git diff --name-only upstream/main
```
### 从官方仓库拉取特定分支/标签
```bash
# 拉取官方的所有标签
git fetch upstream --tags
# 查看所有标签
git tag -l
# 基于某个标签创建分支
git checkout -b v1.0.0 tags/v1.0.0
# 切回主分支
git checkout main
```
---
## 📦 Docker 管理
### 服务管理
```bash
cd /opt/sub2api-dev/deploy
# 查看状态
docker-compose -f docker-compose.prod.yml ps
# 查看日志
docker-compose -f docker-compose.prod.yml logs -f
# 只看 sub2api 日志
docker-compose -f docker-compose.prod.yml logs -f sub2api
# 重启服务
docker-compose -f docker-compose.prod.yml restart
# 停止服务
docker-compose -f docker-compose.prod.yml down
# 完全清理(包括数据)
docker-compose -f docker-compose.prod.yml down -v
```
### 更新镜像
```bash
# 拉取最新镜像
docker-compose -f docker-compose.prod.yml pull
# 重新创建容器
docker-compose -f docker-compose.prod.yml up -d
# 查看镜像信息
docker images | grep sub2api
```
---
## ⚠️ 注意事项
1. **定期备份**
- 备份 `.env` 文件
- 导出 PostgreSQL 数据库
- 提交代码到远程仓库
2. **合并冲突处理**
- 遇到冲突时,优先保留官方核心逻辑
- 调整你的二开代码以适配官方更新
3. **测试环境**
- 建议先在测试环境验证更新
- 确认无误后再应用到生产环境
4. **版本管理**
- 重要功能单独开分支
- 使用有意义的提交信息
- 定期推送到远程仓库
---
## 🔍 故障排查
### Git 问题
```bash
# 如果 push 被拒绝
git pull origin main --rebase
git push origin main
# 如果需要强制推送(危险!)
git push -f origin main
# 查看 Git 配置
git config --list
# 重置 remote
git remote remove origin
git remote add origin https://git.586vip.cn/oadmin/sub2api.git
```
### Docker 问题
```bash
# Redis 连接失败
# 修改 docker-compose.prod.yml 中的 REDIS_HOST
# 从 host.docker.internal 改为 172.17.0.1
# 查看容器详细信息
docker inspect sub2api
# 进入容器调试
docker exec -it sub2api sh
# 查看网络
docker network ls
docker network inspect bridge
```
---
## 📞 快速命令速查
```bash
# Git
git status # 状态
git fetch upstream # 拉取官方更新
git merge upstream/main # 合并更新
git push origin main # 推送到你的仓库
# Docker
docker-compose -f docker-compose.prod.yml ps # 状态
docker-compose -f docker-compose.prod.yml logs -f # 日志
docker-compose -f docker-compose.prod.yml restart # 重启
docker-compose -f docker-compose.prod.yml down # 停止
```

276
GIT_WORKFLOW.md Normal file
View File

@@ -0,0 +1,276 @@
# Sub2API 二次开发指南
## 🔧 Git 工作流程
### 初始设置
```bash
# 1. 克隆官方仓库
git clone https://github.com/Wei-Shaw/sub2api.git sub2api-dev
cd sub2api-dev
# 2. 配置远程仓库
git remote rename origin upstream # 官方仓库改名为 upstream
git remote add origin https://your-git.com/your-repo.git # 添加你的仓库
# 3. 创建开发分支
git checkout -b dev
# 4. 推送到你的仓库
git push -u origin dev
```
### 远程仓库配置结果
```bash
git remote -v
# upstream https://github.com/Wei-Shaw/sub2api.git (fetch)
# upstream https://github.com/Wei-Shaw/sub2api.git (push)
# origin https://your-git.com/your-repo.git (fetch)
# origin https://your-git.com/your-repo.git (push)
```
---
## 🔄 同步官方更新
### 方式 1查看官方更新
```bash
# 拉取官方最新代码
git fetch upstream
# 查看官方更新内容
git log HEAD..upstream/main --oneline
# 查看详细差异
git diff HEAD..upstream/main
# 查看某个文件的差异
git diff HEAD..upstream/main -- backend/internal/service/gateway_service.go
```
### 方式 2合并官方更新
```bash
# 确保工作区干净
git status
# 提交你的修改
git add .
git commit -m "feat: 我的二开功能"
# 拉取并合并官方更新
git fetch upstream
git merge upstream/main
# 如果有冲突,解决后:
git add .
git commit -m "merge: 合并官方更新"
# 推送到你的仓库
git push origin dev
```
### 方式 3使用 Rebase更清晰的历史
```bash
# 拉取官方更新
git fetch upstream
# 将你的提交重放到官方最新代码之上
git rebase upstream/main
# 如果有冲突,解决后:
git add .
git rebase --continue
# 强制推送到你的仓库注意rebase 会改变历史)
git push -f origin dev
```
---
## 📦 部署流程
### 生产环境部署
```bash
cd /opt/sub2api-dev/deploy
# 运行自动部署脚本
chmod +x deploy-dev.sh
bash deploy-dev.sh
```
### 手动部署
```bash
# 1. 进入部署目录
cd /opt/sub2api-dev/deploy
# 2. 创建配置(如果没有)
cp docker-compose.yml docker-compose.prod.yml
cp .env.example .env
# 3. 修改配置
nano docker-compose.prod.yml # 改端口为 2080配置 Redis
nano .env # 设置密码等
# 4. 启动服务
docker-compose -f docker-compose.prod.yml pull
docker-compose -f docker-compose.prod.yml up -d
# 5. 查看日志
docker-compose -f docker-compose.prod.yml logs -f
```
---
## 🛠️ 开发建议
### 目录结构
```
sub2api-dev/
├── .git/ # Git 仓库
├── backend/ # 后端代码Go
│ ├── cmd/server/ # 入口
│ ├── internal/ # 核心逻辑
│ └── ent/ # 数据模型
├── frontend/ # 前端代码Vue
│ └── src/
└── deploy/ # 部署配置
├── docker-compose.prod.yml
├── .env
└── deploy-dev.sh
```
### 常见二开场景
#### 1. 修改后端逻辑
```bash
# 修改代码
vim backend/internal/service/gateway_service.go
# 提交
git add backend/
git commit -m "feat: 修改网关逻辑"
# 重新构建镜像(如果需要)
cd backend
docker build -t sub2api:custom .
# 修改 docker-compose.prod.yml 使用自定义镜像
# image: sub2api:custom
# 重启服务
cd ../deploy
docker-compose -f docker-compose.prod.yml up -d
```
#### 2. 修改前端界面
```bash
# 修改代码
vim frontend/src/views/HomeView.vue
# 本地测试
cd frontend
npm install
npm run dev
# 构建
npm run build
# 提交
git add frontend/
git commit -m "feat: 修改首页界面"
# 后端重新构建(前端会被嵌入)
cd ../backend
go build -tags embed -o sub2api ./cmd/server
```
#### 3. 添加新功能
```bash
# 创建功能分支
git checkout -b feature/new-feature
# 开发...
# 提交
git add .
git commit -m "feat: 添加新功能"
# 合并到开发分支
git checkout dev
git merge feature/new-feature
# 推送
git push origin dev
```
---
## 🔍 对比工具
### 使用 VS Code 对比
```bash
# 安装 VS Code Remote SSH 插件
# 然后在 VS Code 中:
# 1. 连接到服务器
# 2. 打开 /opt/sub2api-dev
# 3. 使用 Git 功能对比差异
```
### 命令行对比
```bash
# 对比某个文件
git diff upstream/main -- backend/internal/service/gateway_service.go
# 对比某个目录
git diff upstream/main -- backend/internal/service/
# 生成对比报告
git diff upstream/main > changes.patch
# 查看修改的文件列表
git diff --name-only upstream/main
```
---
## ⚠️ 注意事项
1. **定期同步**: 建议每周拉取一次官方更新
2. **冲突处理**: 遇到冲突时,优先保留官方的核心逻辑,调整你的二开代码
3. **分支管理**: 重要功能单独开分支,测试通过后再合并
4. **备份数据**: 更新前备份数据库和配置文件
5. **测试环境**: 建议先在测试环境验证更新,再应用到生产
---
## 📋 常用命令速查
```bash
# Git 管理
git status # 查看状态
git log --oneline -10 # 查看提交历史
git fetch upstream # 拉取官方更新
git diff upstream/main # 对比差异
git merge upstream/main # 合并更新
# Docker 管理
docker-compose -f docker-compose.prod.yml ps # 查看状态
docker-compose -f docker-compose.prod.yml logs -f # 查看日志
docker-compose -f docker-compose.prod.yml restart # 重启
docker-compose -f docker-compose.prod.yml down # 停止
# 服务管理
docker-compose -f docker-compose.prod.yml pull # 更新镜像
docker-compose -f docker-compose.prod.yml up -d # 启动
```

View File

@@ -91,7 +91,8 @@ func initializeApplication(buildInfo handler.BuildInfo) (*Application, error) {
rateLimitService := service.NewRateLimitService(accountRepository, usageLogRepository, configConfig, geminiQuotaService)
claudeUsageFetcher := repository.NewClaudeUsageFetcher()
antigravityQuotaFetcher := service.NewAntigravityQuotaFetcher(proxyRepository)
accountUsageService := service.NewAccountUsageService(accountRepository, usageLogRepository, claudeUsageFetcher, geminiQuotaService, antigravityQuotaFetcher)
usageCache := service.NewUsageCache()
accountUsageService := service.NewAccountUsageService(accountRepository, usageLogRepository, claudeUsageFetcher, geminiQuotaService, antigravityQuotaFetcher, usageCache)
geminiTokenCache := repository.NewGeminiTokenCache(redisClient)
geminiTokenProvider := service.NewGeminiTokenProvider(accountRepository, geminiTokenCache, geminiOAuthService)
gatewayCache := repository.NewGatewayCache(redisClient)

View File

@@ -586,8 +586,20 @@ func (h *GatewayHandler) handleStreamingAwareError(c *gin.Context, status int, e
// Stream already started, send error as SSE event then close
flusher, ok := c.Writer.(http.Flusher)
if ok {
// Send error event in SSE format
errorEvent := fmt.Sprintf(`data: {"type": "error", "error": {"type": "%s", "message": "%s"}}`+"\n\n", errType, message)
// Send error event in SSE format with proper JSON marshaling
errorData := map[string]any{
"type": "error",
"error": map[string]string{
"type": errType,
"message": message,
},
}
jsonBytes, err := json.Marshal(errorData)
if err != nil {
_ = c.Error(err)
return
}
errorEvent := fmt.Sprintf("data: %s\n\n", string(jsonBytes))
if _, err := fmt.Fprint(c.Writer, errorEvent); err != nil {
_ = c.Error(err)
}
@@ -737,8 +749,27 @@ func sendMockWarmupStream(c *gin.Context, model string) {
c.Header("Connection", "keep-alive")
c.Header("X-Accel-Buffering", "no")
// Build message_start event with proper JSON marshaling
messageStart := map[string]any{
"type": "message_start",
"message": map[string]any{
"id": "msg_mock_warmup",
"type": "message",
"role": "assistant",
"model": model,
"content": []any{},
"stop_reason": nil,
"stop_sequence": nil,
"usage": map[string]int{
"input_tokens": 10,
"output_tokens": 0,
},
},
}
messageStartJSON, _ := json.Marshal(messageStart)
events := []string{
`event: message_start` + "\n" + `data: {"message":{"content":[],"id":"msg_mock_warmup","model":"` + model + `","role":"assistant","stop_reason":null,"stop_sequence":null,"type":"message","usage":{"input_tokens":10,"output_tokens":0}},"type":"message_start"}`,
`event: message_start` + "\n" + `data: ` + string(messageStartJSON),
`event: content_block_start` + "\n" + `data: {"content_block":{"text":"","type":"text"},"index":0,"type":"content_block_start"}`,
`event: content_block_delta` + "\n" + `data: {"delta":{"text":"New","type":"text_delta"},"index":0,"type":"content_block_delta"}`,
`event: content_block_delta` + "\n" + `data: {"delta":{"text":" Conversation","type":"text_delta"},"index":0,"type":"content_block_delta"}`,

View File

@@ -144,6 +144,21 @@ func (h *ConcurrencyHelper) waitForSlotWithPingTimeout(c *gin.Context, slotType
ctx, cancel := context.WithTimeout(c.Request.Context(), timeout)
defer cancel()
// Try immediate acquire first (avoid unnecessary wait)
var result *service.AcquireResult
var err error
if slotType == "user" {
result, err = h.concurrencyService.AcquireUserSlot(ctx, id, maxConcurrency)
} else {
result, err = h.concurrencyService.AcquireAccountSlot(ctx, id, maxConcurrency)
}
if err != nil {
return nil, err
}
if result.Acquired {
return result.ReleaseFunc, nil
}
// Determine if ping is needed (streaming + ping format defined)
needPing := isStream && h.pingFormat != ""

View File

@@ -143,9 +143,10 @@ type GeminiCandidate struct {
// GeminiUsageMetadata Gemini 用量元数据
type GeminiUsageMetadata struct {
PromptTokenCount int `json:"promptTokenCount,omitempty"`
CandidatesTokenCount int `json:"candidatesTokenCount,omitempty"`
TotalTokenCount int `json:"totalTokenCount,omitempty"`
PromptTokenCount int `json:"promptTokenCount,omitempty"`
CandidatesTokenCount int `json:"candidatesTokenCount,omitempty"`
CachedContentTokenCount int `json:"cachedContentTokenCount,omitempty"`
TotalTokenCount int `json:"totalTokenCount,omitempty"`
}
// DefaultSafetySettings 默认安全设置(关闭所有过滤)

View File

@@ -386,7 +386,7 @@ func buildTools(tools []ClaudeTool) []GeminiToolDeclaration {
// 普通工具
var funcDecls []GeminiFunctionDecl
for i, tool := range tools {
for _, tool := range tools {
// 跳过无效工具名称
if strings.TrimSpace(tool.Name) == "" {
log.Printf("Warning: skipping tool with empty name")
@@ -405,10 +405,6 @@ func buildTools(tools []ClaudeTool) []GeminiToolDeclaration {
description = tool.Custom.Description
inputSchema = tool.Custom.InputSchema
// 调试日志:记录 custom 工具的 schema
if schemaJSON, err := json.Marshal(inputSchema); err == nil {
log.Printf("[Debug] Tool[%d] '%s' (custom) original schema: %s", i, tool.Name, string(schemaJSON))
}
} else {
// 标准格式: 从顶层字段获取
description = tool.Description
@@ -425,11 +421,6 @@ func buildTools(tools []ClaudeTool) []GeminiToolDeclaration {
}
}
// 调试日志:记录清理后的 schema
if paramsJSON, err := json.Marshal(params); err == nil {
log.Printf("[Debug] Tool[%d] '%s' cleaned schema: %s", i, tool.Name, string(paramsJSON))
}
funcDecls = append(funcDecls, GeminiFunctionDecl{
Name: tool.Name,
Description: description,
@@ -584,11 +575,9 @@ func cleanSchemaValue(value any) any {
if k == "additionalProperties" {
if boolVal, ok := val.(bool); ok {
result[k] = boolVal
log.Printf("[Debug] additionalProperties is bool: %v", boolVal)
} else {
// 如果是 schema 对象,转换为 false更安全的默认值
result[k] = false
log.Printf("[Debug] additionalProperties is not bool (type: %T), converting to false", val)
}
continue
}

View File

@@ -96,7 +96,7 @@ func TestBuildTools_CustomTypeTools(t *testing.T) {
{
Type: "custom",
Name: "mcp_tool",
Custom: &CustomToolSpec{
Custom: &ClaudeCustomToolSpec{
Description: "MCP tool description",
InputSchema: map[string]any{
"type": "object",
@@ -121,7 +121,7 @@ func TestBuildTools_CustomTypeTools(t *testing.T) {
{
Type: "custom",
Name: "custom_tool",
Custom: &CustomToolSpec{
Custom: &ClaudeCustomToolSpec{
Description: "Custom tool",
InputSchema: map[string]any{"type": "object"},
},
@@ -148,7 +148,7 @@ func TestBuildTools_CustomTypeTools(t *testing.T) {
{
Type: "custom",
Name: "invalid_custom",
Custom: &CustomToolSpec{
Custom: &ClaudeCustomToolSpec{
Description: "Invalid",
// InputSchema 为 nil
},

View File

@@ -232,10 +232,14 @@ func (p *NonStreamingProcessor) buildResponse(geminiResp *GeminiResponse, respon
stopReason = "max_tokens"
}
// 注意Gemini 的 promptTokenCount 包含 cachedContentTokenCount
// 但 Claude 的 input_tokens 不包含 cache_read_input_tokens需要减去
usage := ClaudeUsage{}
if geminiResp.UsageMetadata != nil {
usage.InputTokens = geminiResp.UsageMetadata.PromptTokenCount
cached := geminiResp.UsageMetadata.CachedContentTokenCount
usage.InputTokens = geminiResp.UsageMetadata.PromptTokenCount - cached
usage.OutputTokens = geminiResp.UsageMetadata.CandidatesTokenCount
usage.CacheReadInputTokens = cached
}
// 生成响应 ID

View File

@@ -29,8 +29,9 @@ type StreamingProcessor struct {
originalModel string
// 累计 usage
inputTokens int
outputTokens int
inputTokens int
outputTokens int
cacheReadTokens int
}
// NewStreamingProcessor 创建流式响应处理器
@@ -76,9 +77,13 @@ func (p *StreamingProcessor) ProcessLine(line string) []byte {
}
// 更新 usage
// 注意Gemini 的 promptTokenCount 包含 cachedContentTokenCount
// 但 Claude 的 input_tokens 不包含 cache_read_input_tokens需要减去
if geminiResp.UsageMetadata != nil {
p.inputTokens = geminiResp.UsageMetadata.PromptTokenCount
cached := geminiResp.UsageMetadata.CachedContentTokenCount
p.inputTokens = geminiResp.UsageMetadata.PromptTokenCount - cached
p.outputTokens = geminiResp.UsageMetadata.CandidatesTokenCount
p.cacheReadTokens = cached
}
// 处理 parts
@@ -108,8 +113,9 @@ func (p *StreamingProcessor) Finish() ([]byte, *ClaudeUsage) {
}
usage := &ClaudeUsage{
InputTokens: p.inputTokens,
OutputTokens: p.outputTokens,
InputTokens: p.inputTokens,
OutputTokens: p.outputTokens,
CacheReadInputTokens: p.cacheReadTokens,
}
return result.Bytes(), usage
@@ -123,8 +129,10 @@ func (p *StreamingProcessor) emitMessageStart(v1Resp *V1InternalResponse) []byte
usage := ClaudeUsage{}
if v1Resp.Response.UsageMetadata != nil {
usage.InputTokens = v1Resp.Response.UsageMetadata.PromptTokenCount
cached := v1Resp.Response.UsageMetadata.CachedContentTokenCount
usage.InputTokens = v1Resp.Response.UsageMetadata.PromptTokenCount - cached
usage.OutputTokens = v1Resp.Response.UsageMetadata.CandidatesTokenCount
usage.CacheReadInputTokens = cached
}
responseID := v1Resp.ResponseID
@@ -418,8 +426,9 @@ func (p *StreamingProcessor) emitFinish(finishReason string) []byte {
}
usage := ClaudeUsage{
InputTokens: p.inputTokens,
OutputTokens: p.outputTokens,
InputTokens: p.inputTokens,
OutputTokens: p.outputTokens,
CacheReadInputTokens: p.cacheReadTokens,
}
deltaEvent := map[string]any{

View File

@@ -151,11 +151,17 @@ var (
return 1
`)
// getAccountsLoadBatchScript - batch load query (read-only)
// ARGV[1] = slot TTL (seconds, retained for compatibility)
// getAccountsLoadBatchScript - batch load query with expired slot cleanup
// ARGV[1] = slot TTL (seconds)
// ARGV[2..n] = accountID1, maxConcurrency1, accountID2, maxConcurrency2, ...
getAccountsLoadBatchScript = redis.NewScript(`
local result = {}
local slotTTL = tonumber(ARGV[1])
-- Get current server time
local timeResult = redis.call('TIME')
local nowSeconds = tonumber(timeResult[1])
local cutoffTime = nowSeconds - slotTTL
local i = 2
while i <= #ARGV do
@@ -163,6 +169,9 @@ var (
local maxConcurrency = tonumber(ARGV[i + 1])
local slotKey = 'concurrency:account:' .. accountID
-- Clean up expired slots before counting
redis.call('ZREMRANGEBYSCORE', slotKey, '-inf', cutoffTime)
local currentConcurrency = redis.call('ZCARD', slotKey)
local waitKey = 'wait:account:' .. accountID

View File

@@ -69,13 +69,29 @@ type windowStatsCache struct {
timestamp time.Time
}
var (
apiCacheMap = sync.Map{} // 缓存 API 响应
windowStatsCacheMap = sync.Map{} // 缓存窗口统计
// antigravityUsageCache 缓存 Antigravity 额度数据
type antigravityUsageCache struct {
usageInfo *UsageInfo
timestamp time.Time
}
const (
apiCacheTTL = 10 * time.Minute
windowStatsCacheTTL = 1 * time.Minute
)
// UsageCache 封装账户使用量相关的缓存
type UsageCache struct {
apiCache sync.Map // accountID -> *apiUsageCache
windowStatsCache sync.Map // accountID -> *windowStatsCache
antigravityCache sync.Map // accountID -> *antigravityUsageCache
}
// NewUsageCache 创建 UsageCache 实例
func NewUsageCache() *UsageCache {
return &UsageCache{}
}
// WindowStats 窗口期统计
type WindowStats struct {
Requests int64 `json:"requests"`
@@ -138,6 +154,7 @@ type AccountUsageService struct {
usageFetcher ClaudeUsageFetcher
geminiQuotaService *GeminiQuotaService
antigravityQuotaFetcher *AntigravityQuotaFetcher
cache *UsageCache
}
// NewAccountUsageService 创建AccountUsageService实例
@@ -147,6 +164,7 @@ func NewAccountUsageService(
usageFetcher ClaudeUsageFetcher,
geminiQuotaService *GeminiQuotaService,
antigravityQuotaFetcher *AntigravityQuotaFetcher,
cache *UsageCache,
) *AccountUsageService {
return &AccountUsageService{
accountRepo: accountRepo,
@@ -154,6 +172,7 @@ func NewAccountUsageService(
usageFetcher: usageFetcher,
geminiQuotaService: geminiQuotaService,
antigravityQuotaFetcher: antigravityQuotaFetcher,
cache: cache,
}
}
@@ -181,7 +200,7 @@ func (s *AccountUsageService) GetUsage(ctx context.Context, accountID int64) (*U
var apiResp *ClaudeUsageResponse
// 1. 检查 API 缓存10 分钟)
if cached, ok := apiCacheMap.Load(accountID); ok {
if cached, ok := s.cache.apiCache.Load(accountID); ok {
if cache, ok := cached.(*apiUsageCache); ok && time.Since(cache.timestamp) < apiCacheTTL {
apiResp = cache.response
}
@@ -194,7 +213,7 @@ func (s *AccountUsageService) GetUsage(ctx context.Context, accountID int64) (*U
return nil, err
}
// 缓存 API 响应
apiCacheMap.Store(accountID, &apiUsageCache{
s.cache.apiCache.Store(accountID, &apiUsageCache{
response: apiResp,
timestamp: time.Now(),
})
@@ -252,14 +271,6 @@ func (s *AccountUsageService) getGeminiUsage(ctx context.Context, account *Accou
return usage, nil
}
// antigravityUsageCache 缓存 Antigravity 额度数据
type antigravityUsageCache struct {
usageInfo *UsageInfo
timestamp time.Time
}
var antigravityCacheMap = sync.Map{}
// getAntigravityUsage 获取 Antigravity 账户额度
func (s *AccountUsageService) getAntigravityUsage(ctx context.Context, account *Account) (*UsageInfo, error) {
if s.antigravityQuotaFetcher == nil || !s.antigravityQuotaFetcher.CanFetch(account) {
@@ -268,7 +279,7 @@ func (s *AccountUsageService) getAntigravityUsage(ctx context.Context, account *
}
// 1. 检查缓存10 分钟)
if cached, ok := antigravityCacheMap.Load(account.ID); ok {
if cached, ok := s.cache.antigravityCache.Load(account.ID); ok {
if cache, ok := cached.(*antigravityUsageCache); ok && time.Since(cache.timestamp) < apiCacheTTL {
// 重新计算 RemainingSeconds
usage := cache.usageInfo
@@ -289,7 +300,7 @@ func (s *AccountUsageService) getAntigravityUsage(ctx context.Context, account *
}
// 4. 缓存结果
antigravityCacheMap.Store(account.ID, &antigravityUsageCache{
s.cache.antigravityCache.Store(account.ID, &antigravityUsageCache{
usageInfo: result.UsageInfo,
timestamp: time.Now(),
})
@@ -308,7 +319,7 @@ func (s *AccountUsageService) addWindowStats(ctx context.Context, account *Accou
// 检查窗口统计缓存1 分钟)
var windowStats *WindowStats
if cached, ok := windowStatsCacheMap.Load(account.ID); ok {
if cached, ok := s.cache.windowStatsCache.Load(account.ID); ok {
if cache, ok := cached.(*windowStatsCache); ok && time.Since(cache.timestamp) < windowStatsCacheTTL {
windowStats = cache.stats
}
@@ -336,7 +347,7 @@ func (s *AccountUsageService) addWindowStats(ctx context.Context, account *Accou
}
// 缓存窗口统计1 分钟)
windowStatsCacheMap.Store(account.ID, &windowStatsCache{
s.cache.windowStatsCache.Store(account.ID, &windowStatsCache{
stats: windowStats,
timestamp: time.Now(),
})

View File

@@ -322,9 +322,6 @@ func (s *AntigravityGatewayService) Forward(ctx context.Context, c *gin.Context,
originalModel := claudeReq.Model
mappedModel := s.getMappedModel(account, claudeReq.Model)
if mappedModel != claudeReq.Model {
log.Printf("Antigravity model mapping: %s -> %s (account: %s)", claudeReq.Model, mappedModel, account.Name)
}
// 获取 access_token
if s.tokenProvider == nil {
@@ -350,15 +347,6 @@ func (s *AntigravityGatewayService) Forward(ctx context.Context, c *gin.Context,
return nil, fmt.Errorf("transform request: %w", err)
}
// 调试:记录转换后的请求体(仅记录前 2000 字符)
if bodyJSON, err := json.Marshal(geminiBody); err == nil {
truncated := string(bodyJSON)
if len(truncated) > 2000 {
truncated = truncated[:2000] + "..."
}
log.Printf("[Debug] Transformed Gemini request: %s", truncated)
}
// 构建上游 action
action := "generateContent"
if claudeReq.Stream {

View File

@@ -204,7 +204,7 @@ func (s *GatewayService) GenerateSessionHash(parsed *ParsedRequest) string {
// BindStickySession sets session -> account binding with standard TTL.
func (s *GatewayService) BindStickySession(ctx context.Context, sessionHash string, accountID int64) error {
if sessionHash == "" || accountID <= 0 {
if sessionHash == "" || accountID <= 0 || s.cache == nil {
return nil
}
return s.cache.SetSessionAccountID(ctx, sessionHash, accountID, stickySessionTTL)
@@ -429,7 +429,7 @@ func (s *GatewayService) SelectAccountWithLoadAwareness(ctx context.Context, gro
}
// ============ Layer 1: 粘性会话优先 ============
if sessionHash != "" {
if sessionHash != "" && s.cache != nil {
accountID, err := s.cache.GetSessionAccountID(ctx, sessionHash)
if err == nil && accountID > 0 && !isExcluded(accountID) {
account, err := s.accountRepo.GetByID(ctx, accountID)

View File

@@ -114,4 +114,5 @@ var ProviderSet = wire.NewSet(
ProvideDeferredService,
NewAntigravityQuotaFetcher,
NewUserAttributeService,
NewUsageCache,
)

51
deploy/.env.1panel Normal file
View File

@@ -0,0 +1,51 @@
# =============================================================================
# Sub2API 环境配置 - 1Panel 环境
# =============================================================================
# -----------------------------------------------------------------------------
# Server Configuration
# -----------------------------------------------------------------------------
BIND_HOST=0.0.0.0
SERVER_PORT=2080
SERVER_MODE=release
RUN_MODE=standard
TZ=Asia/Shanghai
# -----------------------------------------------------------------------------
# PostgreSQL Configuration (REQUIRED - 请修改密码!)
# -----------------------------------------------------------------------------
POSTGRES_USER=sub2api
POSTGRES_PASSWORD=请修改为你的强密码
POSTGRES_DB=sub2api
# -----------------------------------------------------------------------------
# Redis Configuration - 使用你现有的 1Panel Redis
# -----------------------------------------------------------------------------
# 方式1如果使用 Docker 网络连接,填写容器名
REDIS_HOST=1Panel-redis-S1KH
# 方式2如果上面不行使用宿主机 IP需要你填写服务器的内网 IP
# REDIS_HOST=172.17.0.1
REDIS_PORT=6379
REDIS_PASSWORD=redis_bJFKDk
REDIS_DB=1
# -----------------------------------------------------------------------------
# Admin Account
# -----------------------------------------------------------------------------
ADMIN_EMAIL=oadmin
ADMIN_PASSWORD=oadmin@123
# -----------------------------------------------------------------------------
# JWT Configuration
# -----------------------------------------------------------------------------
# 使用下面的命令生成随机密钥openssl rand -hex 32
JWT_SECRET=1fb3a3533039e0d1ea693f0508bbffca44b7157370f11dfe33ce352a5f07039a
JWT_EXPIRE_HOUR=24
# -----------------------------------------------------------------------------
# Gemini OAuth (可选 - 如果不使用 Gemini 可以忽略)
# -----------------------------------------------------------------------------
GEMINI_OAUTH_CLIENT_ID=
GEMINI_OAUTH_CLIENT_SECRET=
GEMINI_OAUTH_SCOPES=
GEMINI_QUOTA_POLICY=

136
deploy/README_DEPLOY.md Normal file
View File

@@ -0,0 +1,136 @@
# Sub2API 部署指南 - 端口 2080
## 快速部署(推荐)
### 1. 上传文件到服务器
`deploy` 目录上传到服务器的 `/opt/sub2api`
```bash
# 在服务器上创建目录
mkdir -p /opt/sub2api
```
### 2. 运行自动部署脚本
```bash
cd /opt/sub2api
chmod +x deploy.sh
bash deploy.sh
```
脚本会自动完成:
- ✅ 生成安全的密码和密钥
- ✅ 创建配置文件
- ✅ 拉取 Docker 镜像
- ✅ 启动服务
### 3. 访问系统
浏览器打开: `http://你的服务器IP:2080`
默认账号:
- 邮箱: `admin@example.com`
- 密码: `admin123`
---
## 手动部署
如果自动脚本失败,可以手动部署:
```bash
cd /opt/sub2api
# 1. 复制配置文件
cp docker-compose.2080.yml docker-compose.yml
cp .env.2080 .env
# 2. 生成密钥
sed -i "s/POSTGRES_PASSWORD=.*/POSTGRES_PASSWORD=$(openssl rand -base64 24)/" .env
sed -i "s/JWT_SECRET=.*/JWT_SECRET=$(openssl rand -hex 32)/" .env
# 3. 启动服务
docker-compose pull
docker-compose up -d
# 4. 查看日志
docker-compose logs -f
```
---
## 配置说明
### 端口映射
- 宿主机端口: `2080`
- 容器内端口: `8080`
### Redis 连接
- 使用现有的 1Panel Redis: `1Panel-redis-S1KH`
- 端口: `6379`
- 密码: `redis_bJFKDk`
- 数据库: `1` (避免与其他应用冲突)
### PostgreSQL
- 自动创建独立的 PostgreSQL 容器
- 数据持久化在 Docker volume 中
---
## 常用命令
```bash
# 查看服务状态
docker-compose ps
# 查看日志
docker-compose logs -f
# 重启服务
docker-compose restart
# 停止服务
docker-compose down
# 更新到最新版本
docker-compose pull
docker-compose up -d
```
---
## 故障排查
### Redis 连接失败
如果日志显示 Redis 连接失败,尝试修改 `docker-compose.yml`:
将:
```yaml
- REDIS_HOST=host.docker.internal
```
改为:
```yaml
- REDIS_HOST=172.17.0.1
```
然后重启:
```bash
docker-compose down
docker-compose up -d
```
### 查看详细日志
```bash
# 只看 sub2api 日志
docker-compose logs sub2api
# 只看 PostgreSQL 日志
docker-compose logs postgres
# 实时查看最新 50 行日志
docker-compose logs -f --tail=50
```

221
deploy/deploy-complete.sh Normal file
View File

@@ -0,0 +1,221 @@
#!/bin/bash
# Sub2API 完整部署脚本 - 双 Git Remote 配置
# 官方仓库: https://github.com/Wei-Shaw/sub2api.git
# 你的仓库: https://git.586vip.cn/oadmin/sub2api.git
set -e
echo "=========================================="
echo " Sub2API 二次开发环境部署"
echo " 端口: 2080"
echo " 双 Remote Git 配置"
echo "=========================================="
echo ""
# 检查是否已经存在项目目录
if [ -d "/opt/sub2api-dev" ]; then
echo "⚠️ 目录 /opt/sub2api-dev 已存在"
read -p "是否删除并重新部署?(y/N): " confirm
if [[ $confirm == [yY] ]]; then
rm -rf /opt/sub2api-dev
else
echo "❌ 取消部署"
exit 1
fi
fi
cd /opt
# ===================================
# 1. 克隆官方仓库
# ===================================
echo "📥 克隆官方仓库..."
git clone https://github.com/Wei-Shaw/sub2api.git sub2api-dev
cd sub2api-dev
echo "✅ 官方仓库克隆完成"
echo ""
# ===================================
# 2. 配置双 Remote
# ===================================
echo "🔧 配置 Git Remote..."
# 将 origin 改名为 upstream官方仓库
git remote rename origin upstream
# 添加你的仓库为 origin
git remote add origin https://git.586vip.cn/oadmin/sub2api.git
# 显示配置结果
echo "✅ Git Remote 配置完成:"
git remote -v
echo ""
# ===================================
# 3. 创建并切换到 main 分支
# ===================================
echo "🌿 创建 main 分支..."
git checkout -b main
echo "✅ 当前分支: $(git branch --show-current)"
echo ""
# ===================================
# 4. 首次推送到你的仓库
# ===================================
echo "📤 推送到你的仓库..."
git push -u origin main
echo "✅ 代码已推送到: https://git.586vip.cn/oadmin/sub2api.git"
echo ""
# ===================================
# 5. 配置部署文件
# ===================================
echo "📝 配置部署文件..."
cd deploy
# 创建生产环境 docker-compose 配置
cat > docker-compose.prod.yml << 'EOF'
services:
sub2api:
image: weishaw/sub2api:latest
container_name: sub2api
restart: unless-stopped
ports:
- "2080:8080"
volumes:
- ./data:/app/data
environment:
- AUTO_SETUP=true
- SERVER_HOST=0.0.0.0
- SERVER_PORT=8080
- SERVER_MODE=release
- DATABASE_HOST=postgres
- DATABASE_PORT=5432
- DATABASE_USER=sub2api
- DATABASE_PASSWORD=${POSTGRES_PASSWORD}
- DATABASE_DBNAME=sub2api
- DATABASE_SSLMODE=disable
- REDIS_HOST=host.docker.internal
- REDIS_PORT=6379
- REDIS_PASSWORD=redis_bJFKDk
- REDIS_DB=1
- ADMIN_EMAIL=${ADMIN_EMAIL}
- ADMIN_PASSWORD=${ADMIN_PASSWORD}
- JWT_SECRET=${JWT_SECRET}
- TZ=Asia/Shanghai
depends_on:
- postgres
extra_hosts:
- "host.docker.internal:host-gateway"
postgres:
image: postgres:18-alpine
container_name: sub2api-postgres
restart: unless-stopped
volumes:
- postgres_data:/var/lib/postgresql/data
environment:
- POSTGRES_USER=sub2api
- POSTGRES_PASSWORD=${POSTGRES_PASSWORD}
- POSTGRES_DB=sub2api
- TZ=Asia/Shanghai
healthcheck:
test: ["CMD-SHELL", "pg_isready -U sub2api"]
interval: 10s
timeout: 5s
retries: 5
volumes:
postgres_data:
EOF
echo "✅ docker-compose.prod.yml 已创建"
# 创建 .env 文件
if [ ! -f ".env" ]; then
echo "🔐 生成配置文件..."
POSTGRES_PASS=$(openssl rand -base64 24)
JWT_KEY=$(openssl rand -hex 32)
cat > .env << EOF
# PostgreSQL 配置
POSTGRES_USER=sub2api
POSTGRES_PASSWORD=${POSTGRES_PASS}
POSTGRES_DB=sub2api
# 管理员账号
ADMIN_EMAIL=admin@example.com
ADMIN_PASSWORD=admin123
# JWT 配置
JWT_SECRET=${JWT_KEY}
JWT_EXPIRE_HOUR=24
# 其他配置
TZ=Asia/Shanghai
SERVER_MODE=release
RUN_MODE=standard
EOF
echo "✅ .env 文件已创建"
echo ""
echo "🔑 生成的密钥(请保存):"
echo " PostgreSQL 密码: ${POSTGRES_PASS}"
echo " JWT Secret: ${JWT_KEY}"
echo ""
fi
# ===================================
# 6. 启动 Docker 服务
# ===================================
echo "🐳 拉取 Docker 镜像..."
docker-compose -f docker-compose.prod.yml pull
echo ""
echo "🚀 启动服务..."
docker-compose -f docker-compose.prod.yml up -d
echo ""
echo "⏳ 等待服务启动..."
sleep 15
# ===================================
# 7. 显示部署结果
# ===================================
echo ""
echo "📊 服务状态:"
docker-compose -f docker-compose.prod.yml ps
echo ""
echo "=========================================="
echo " ✅ 部署完成!"
echo "=========================================="
echo ""
echo "🌐 访问信息:"
echo " URL: http://$(hostname -I | awk '{print $1}'):2080"
echo " 管理员: admin@example.com / admin123"
echo ""
echo "📋 Git 配置:"
echo " 官方仓库 (upstream): https://github.com/Wei-Shaw/sub2api.git"
echo " 你的仓库 (origin): https://git.586vip.cn/oadmin/sub2api.git"
echo " 当前分支: main"
echo ""
echo "🔄 同步官方更新:"
echo " cd /opt/sub2api-dev"
echo " git fetch upstream"
echo " git merge upstream/main"
echo " git push origin main"
echo ""
echo "📝 常用命令:"
echo " 查看日志: cd /opt/sub2api-dev/deploy && docker-compose -f docker-compose.prod.yml logs -f"
echo " 重启服务: docker-compose -f docker-compose.prod.yml restart"
echo " 停止服务: docker-compose -f docker-compose.prod.yml down"
echo ""
echo "🎉 访问 http://$(hostname -I | awk '{print $1}'):2080 开始使用!"
echo ""

120
deploy/deploy-dev.sh Normal file
View File

@@ -0,0 +1,120 @@
#!/bin/bash
# Sub2API 二次开发部署脚本
# 使用方法: bash deploy-dev.sh
set -e
echo "====================================="
echo " Sub2API 二开版本部署"
echo " 端口: 2080"
echo "====================================="
echo ""
# 检查是否在 Git 仓库中
if [ ! -d ".git" ]; then
echo "❌ 错误: 当前目录不是 Git 仓库"
echo "请在项目根目录运行此脚本"
exit 1
fi
echo "📦 当前 Git 信息:"
echo " 分支: $(git branch --show-current)"
echo " 最新提交: $(git log -1 --oneline)"
echo ""
# 进入 deploy 目录
cd deploy
# 检查配置文件
if [ ! -f "docker-compose.prod.yml" ]; then
echo "📝 创建生产环境配置..."
cp docker-compose.yml docker-compose.prod.yml
# 修改端口为 2080
sed -i 's/\${SERVER_PORT:-8080}/2080/' docker-compose.prod.yml
# 添加 extra_hosts 配置(连接宿主机 Redis
if ! grep -q "extra_hosts" docker-compose.prod.yml; then
sed -i '/depends_on:/i\ extra_hosts:\n - "host.docker.internal:host-gateway"' docker-compose.prod.yml
fi
# 修改 Redis 配置
sed -i 's/REDIS_HOST=redis/REDIS_HOST=host.docker.internal/' docker-compose.prod.yml
sed -i 's/REDIS_PASSWORD=\${REDIS_PASSWORD:-}/REDIS_PASSWORD=redis_bJFKDk/' docker-compose.prod.yml
sed -i 's/REDIS_DB=\${REDIS_DB:-0}/REDIS_DB=1/' docker-compose.prod.yml
echo "✅ docker-compose.prod.yml 已创建"
fi
# 创建 .env 文件
if [ ! -f ".env" ]; then
echo "🔐 生成配置文件..."
cat > .env << EOF
# PostgreSQL 配置
POSTGRES_USER=sub2api
POSTGRES_PASSWORD=$(openssl rand -base64 24)
POSTGRES_DB=sub2api
# 管理员账号
ADMIN_EMAIL=admin@example.com
ADMIN_PASSWORD=admin123
# JWT 配置
JWT_SECRET=$(openssl rand -hex 32)
JWT_EXPIRE_HOUR=24
# 其他配置
TZ=Asia/Shanghai
SERVER_MODE=release
RUN_MODE=standard
EOF
echo "✅ .env 文件已创建"
echo ""
echo "📋 生成的配置(请保存):"
cat .env | grep -E "POSTGRES_PASSWORD|JWT_SECRET"
echo ""
else
echo "⚠️ .env 已存在,跳过创建"
fi
# 拉取镜像
echo ""
echo "📥 拉取 Docker 镜像..."
docker-compose -f docker-compose.prod.yml pull
# 启动服务
echo ""
echo "🚀 启动服务..."
docker-compose -f docker-compose.prod.yml up -d
# 等待启动
echo ""
echo "⏳ 等待服务启动..."
sleep 10
# 显示状态
echo ""
echo "📊 服务状态:"
docker-compose -f docker-compose.prod.yml ps
echo ""
echo "====================================="
echo " ✅ 部署完成!"
echo "====================================="
echo ""
echo "📝 访问信息:"
echo " URL: http://$(hostname -I | awk '{print $1}'):2080"
echo " 管理员: admin@example.com / admin123"
echo ""
echo "📋 常用命令:"
echo " 查看日志: docker-compose -f docker-compose.prod.yml logs -f"
echo " 重启服务: docker-compose -f docker-compose.prod.yml restart"
echo " 停止服务: docker-compose -f docker-compose.prod.yml down"
echo ""
echo "🔄 同步官方更新:"
echo " git fetch upstream"
echo " git merge upstream/main"
echo ""

104
deploy/deploy.sh Normal file
View File

@@ -0,0 +1,104 @@
#!/bin/bash
# Sub2API 自动部署脚本 - 端口 2080
# 使用方法: bash deploy.sh
set -e
echo "====================================="
echo " Sub2API 自动部署脚本"
echo " 端口: 2080"
echo "====================================="
echo ""
# 检查是否在正确的目录
if [ ! -f "docker-compose.2080.yml" ]; then
echo "❌ 错误: 请在包含 docker-compose.2080.yml 的目录中运行此脚本"
exit 1
fi
# 检查 Docker 是否安装
if ! command -v docker &> /dev/null; then
echo "❌ Docker 未安装,请先安装 Docker"
exit 1
fi
if ! command -v docker-compose &> /dev/null; then
echo "❌ Docker Compose 未安装,请先安装 Docker Compose"
exit 1
fi
echo "✅ Docker 环境检查通过"
echo ""
# 复制配置文件
if [ ! -f ".env" ]; then
echo "📝 创建配置文件..."
cp .env.2080 .env
# 生成随机密码和密钥
echo "🔐 生成安全密钥..."
POSTGRES_PASS=$(openssl rand -base64 24)
JWT_KEY=$(openssl rand -hex 32)
# 更新 .env 文件
sed -i "s/POSTGRES_PASSWORD=.*/POSTGRES_PASSWORD=${POSTGRES_PASS}/" .env
sed -i "s/JWT_SECRET=.*/JWT_SECRET=${JWT_KEY}/" .env
echo "✅ 配置文件已创建: .env"
echo ""
echo "📋 生成的密码(请保存):"
echo " PostgreSQL 密码: ${POSTGRES_PASS}"
echo ""
else
echo "⚠️ .env 文件已存在,跳过创建"
echo ""
fi
# 复制 docker-compose 文件
if [ ! -f "docker-compose.yml" ]; then
echo "📝 创建 docker-compose.yml..."
cp docker-compose.2080.yml docker-compose.yml
echo "✅ docker-compose.yml 已创建"
echo ""
fi
# 拉取镜像
echo "📥 拉取 Docker 镜像..."
docker-compose pull
# 启动服务
echo ""
echo "🚀 启动服务..."
docker-compose up -d
# 等待服务启动
echo ""
echo "⏳ 等待服务启动..."
sleep 10
# 检查服务状态
echo ""
echo "📊 服务状态:"
docker-compose ps
echo ""
echo "====================================="
echo " ✅ 部署完成!"
echo "====================================="
echo ""
echo "📝 访问信息:"
echo " URL: http://$(hostname -I | awk '{print $1}'):2080"
echo " 管理员邮箱: admin@example.com"
echo " 管理员密码: admin123"
echo ""
echo "⚠️ 安全建议:"
echo " 1. 登录后立即修改管理员密码"
echo " 2. 如需修改配置,编辑 .env 文件后执行: docker-compose restart"
echo ""
echo "📋 常用命令:"
echo " 查看日志: docker-compose logs -f"
echo " 重启服务: docker-compose restart"
echo " 停止服务: docker-compose down"
echo ""
echo "🎉 部署完成!请在浏览器中访问上述 URL"

View File

@@ -0,0 +1,111 @@
# =============================================================================
# Sub2API Docker Compose - 1Panel 环境配置
# =============================================================================
# 此配置文件适用于已有 1Panel 环境,复用现有 Redis
# =============================================================================
services:
# ===========================================================================
# Sub2API Application
# ===========================================================================
sub2api:
image: weishaw/sub2api:latest
container_name: sub2api
restart: unless-stopped
ulimits:
nofile:
soft: 100000
hard: 100000
ports:
- "${BIND_HOST:-0.0.0.0}:${SERVER_PORT:-8080}:8080"
volumes:
- sub2api_data:/app/data
environment:
# Auto Setup
- AUTO_SETUP=true
# Server Configuration
- SERVER_HOST=0.0.0.0
- SERVER_PORT=8080
- SERVER_MODE=${SERVER_MODE:-release}
- RUN_MODE=${RUN_MODE:-standard}
# Database Configuration
- DATABASE_HOST=postgres
- DATABASE_PORT=5432
- DATABASE_USER=${POSTGRES_USER:-sub2api}
- DATABASE_PASSWORD=${POSTGRES_PASSWORD:?POSTGRES_PASSWORD is required}
- DATABASE_DBNAME=${POSTGRES_DB:-sub2api}
- DATABASE_SSLMODE=disable
# Redis Configuration - 使用外部 Redis
- REDIS_HOST=${REDIS_HOST}
- REDIS_PORT=${REDIS_PORT:-6379}
- REDIS_PASSWORD=${REDIS_PASSWORD}
- REDIS_DB=${REDIS_DB:-0}
# Admin Account
- ADMIN_EMAIL=${ADMIN_EMAIL:-admin@sub2api.local}
- ADMIN_PASSWORD=${ADMIN_PASSWORD:-}
# JWT Configuration
- JWT_SECRET=${JWT_SECRET:-}
- JWT_EXPIRE_HOUR=${JWT_EXPIRE_HOUR:-24}
# Timezone
- TZ=${TZ:-Asia/Shanghai}
# Gemini OAuth (可选)
- GEMINI_OAUTH_CLIENT_ID=${GEMINI_OAUTH_CLIENT_ID:-}
- GEMINI_OAUTH_CLIENT_SECRET=${GEMINI_OAUTH_CLIENT_SECRET:-}
- GEMINI_OAUTH_SCOPES=${GEMINI_OAUTH_SCOPES:-}
- GEMINI_QUOTA_POLICY=${GEMINI_QUOTA_POLICY:-}
depends_on:
postgres:
condition: service_healthy
# 使用 host 网络模式以访问宿主机上的 1Panel Redis
# 或者可以使用 extra_hosts 添加到 1Panel 网络
network_mode: bridge
extra_hosts:
- "host.docker.internal:host-gateway"
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:8080/health"]
interval: 30s
timeout: 10s
retries: 3
start_period: 30s
# ===========================================================================
# PostgreSQL Database
# ===========================================================================
postgres:
image: postgres:18-alpine
container_name: sub2api-postgres
restart: unless-stopped
ulimits:
nofile:
soft: 100000
hard: 100000
volumes:
- postgres_data:/var/lib/postgresql/data
environment:
- POSTGRES_USER=${POSTGRES_USER:-sub2api}
- POSTGRES_PASSWORD=${POSTGRES_PASSWORD:?POSTGRES_PASSWORD is required}
- POSTGRES_DB=${POSTGRES_DB:-sub2api}
- TZ=${TZ:-Asia/Shanghai}
network_mode: bridge
healthcheck:
test: ["CMD-SHELL", "pg_isready -U ${POSTGRES_USER:-sub2api} -d ${POSTGRES_DB:-sub2api}"]
interval: 10s
timeout: 5s
retries: 5
start_period: 10s
# =============================================================================
# Volumes
# =============================================================================
volumes:
sub2api_data:
driver: local
postgres_data:
driver: local