Compare commits
2 Commits
v0.6.2-alp
...
dev
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
42569c83c0 | ||
|
|
b373882814 |
4
.github/workflows/linux-release.yml
vendored
4
.github/workflows/linux-release.yml
vendored
@@ -23,7 +23,7 @@ jobs:
|
|||||||
- uses: actions/setup-node@v3
|
- uses: actions/setup-node@v3
|
||||||
with:
|
with:
|
||||||
node-version: 16
|
node-version: 16
|
||||||
- name: Build Frontend
|
- name: Build Frontend (theme default)
|
||||||
env:
|
env:
|
||||||
CI: ""
|
CI: ""
|
||||||
run: |
|
run: |
|
||||||
@@ -38,7 +38,7 @@ jobs:
|
|||||||
- name: Build Backend (amd64)
|
- name: Build Backend (amd64)
|
||||||
run: |
|
run: |
|
||||||
go mod download
|
go mod download
|
||||||
go build -ldflags "-s -w -X 'github.com/songquanpeng/one-api/common.Version=$(git describe --tags)' -extldflags '-static'" -o one-api
|
go build -ldflags "-s -w -X 'one-api/common.Version=$(git describe --tags)' -extldflags '-static'" -o one-api
|
||||||
|
|
||||||
- name: Build Backend (arm64)
|
- name: Build Backend (arm64)
|
||||||
run: |
|
run: |
|
||||||
|
|||||||
4
.github/workflows/macos-release.yml
vendored
4
.github/workflows/macos-release.yml
vendored
@@ -23,7 +23,7 @@ jobs:
|
|||||||
- uses: actions/setup-node@v3
|
- uses: actions/setup-node@v3
|
||||||
with:
|
with:
|
||||||
node-version: 16
|
node-version: 16
|
||||||
- name: Build Frontend
|
- name: Build Frontend (theme default)
|
||||||
env:
|
env:
|
||||||
CI: ""
|
CI: ""
|
||||||
run: |
|
run: |
|
||||||
@@ -38,7 +38,7 @@ jobs:
|
|||||||
- name: Build Backend
|
- name: Build Backend
|
||||||
run: |
|
run: |
|
||||||
go mod download
|
go mod download
|
||||||
go build -ldflags "-X 'github.com/songquanpeng/one-api/common.Version=$(git describe --tags)'" -o one-api-macos
|
go build -ldflags "-X 'one-api/common.Version=$(git describe --tags)'" -o one-api-macos
|
||||||
- name: Release
|
- name: Release
|
||||||
uses: softprops/action-gh-release@v1
|
uses: softprops/action-gh-release@v1
|
||||||
if: startsWith(github.ref, 'refs/tags/')
|
if: startsWith(github.ref, 'refs/tags/')
|
||||||
|
|||||||
4
.github/workflows/windows-release.yml
vendored
4
.github/workflows/windows-release.yml
vendored
@@ -26,7 +26,7 @@ jobs:
|
|||||||
- uses: actions/setup-node@v3
|
- uses: actions/setup-node@v3
|
||||||
with:
|
with:
|
||||||
node-version: 16
|
node-version: 16
|
||||||
- name: Build Frontend
|
- name: Build Frontend (theme default)
|
||||||
env:
|
env:
|
||||||
CI: ""
|
CI: ""
|
||||||
run: |
|
run: |
|
||||||
@@ -41,7 +41,7 @@ jobs:
|
|||||||
- name: Build Backend
|
- name: Build Backend
|
||||||
run: |
|
run: |
|
||||||
go mod download
|
go mod download
|
||||||
go build -ldflags "-s -w -X 'github.com/songquanpeng/one-api/common.Version=$(git describe --tags)'" -o one-api.exe
|
go build -ldflags "-s -w -X 'one-api/common.Version=$(git describe --tags)'" -o one-api.exe
|
||||||
- name: Release
|
- name: Release
|
||||||
uses: softprops/action-gh-release@v1
|
uses: softprops/action-gh-release@v1
|
||||||
if: startsWith(github.ref, 'refs/tags/')
|
if: startsWith(github.ref, 'refs/tags/')
|
||||||
|
|||||||
3
.gitignore
vendored
3
.gitignore
vendored
@@ -6,5 +6,4 @@ upload
|
|||||||
build
|
build
|
||||||
*.db-journal
|
*.db-journal
|
||||||
logs
|
logs
|
||||||
data
|
data
|
||||||
/web/node_modules
|
|
||||||
@@ -23,7 +23,7 @@ ADD go.mod go.sum ./
|
|||||||
RUN go mod download
|
RUN go mod download
|
||||||
COPY . .
|
COPY . .
|
||||||
COPY --from=builder /web/build ./web/build
|
COPY --from=builder /web/build ./web/build
|
||||||
RUN go build -ldflags "-s -w -X 'github.com/songquanpeng/one-api/common.Version=$(cat VERSION)' -extldflags '-static'" -o one-api
|
RUN go build -ldflags "-s -w -X 'one-api/common.Version=$(cat VERSION)' -extldflags '-static'" -o one-api
|
||||||
|
|
||||||
FROM alpine
|
FROM alpine
|
||||||
|
|
||||||
|
|||||||
@@ -134,12 +134,12 @@ The initial account username is `root` and password is `123456`.
|
|||||||
git clone https://github.com/songquanpeng/one-api.git
|
git clone https://github.com/songquanpeng/one-api.git
|
||||||
|
|
||||||
# Build the frontend
|
# Build the frontend
|
||||||
cd one-api/web/default
|
cd one-api/web
|
||||||
npm install
|
npm install
|
||||||
npm run build
|
npm run build
|
||||||
|
|
||||||
# Build the backend
|
# Build the backend
|
||||||
cd ../..
|
cd ..
|
||||||
go mod download
|
go mod download
|
||||||
go build -ldflags "-s -w" -o one-api
|
go build -ldflags "-s -w" -o one-api
|
||||||
```
|
```
|
||||||
|
|||||||
@@ -135,12 +135,12 @@ sudo service nginx restart
|
|||||||
git clone https://github.com/songquanpeng/one-api.git
|
git clone https://github.com/songquanpeng/one-api.git
|
||||||
|
|
||||||
# フロントエンドのビルド
|
# フロントエンドのビルド
|
||||||
cd one-api/web/default
|
cd one-api/web
|
||||||
npm install
|
npm install
|
||||||
npm run build
|
npm run build
|
||||||
|
|
||||||
# バックエンドのビルド
|
# バックエンドのビルド
|
||||||
cd ../..
|
cd ..
|
||||||
go mod download
|
go mod download
|
||||||
go build -ldflags "-s -w" -o one-api
|
go build -ldflags "-s -w" -o one-api
|
||||||
```
|
```
|
||||||
|
|||||||
14
README.md
14
README.md
@@ -67,18 +67,12 @@ _✨ 通过标准的 OpenAI API 格式访问所有的大模型,开箱即用
|
|||||||
+ [x] [OpenAI ChatGPT 系列模型](https://platform.openai.com/docs/guides/gpt/chat-completions-api)(支持 [Azure OpenAI API](https://learn.microsoft.com/en-us/azure/ai-services/openai/reference))
|
+ [x] [OpenAI ChatGPT 系列模型](https://platform.openai.com/docs/guides/gpt/chat-completions-api)(支持 [Azure OpenAI API](https://learn.microsoft.com/en-us/azure/ai-services/openai/reference))
|
||||||
+ [x] [Anthropic Claude 系列模型](https://anthropic.com)
|
+ [x] [Anthropic Claude 系列模型](https://anthropic.com)
|
||||||
+ [x] [Google PaLM2/Gemini 系列模型](https://developers.generativeai.google)
|
+ [x] [Google PaLM2/Gemini 系列模型](https://developers.generativeai.google)
|
||||||
+ [x] [Mistral 系列模型](https://mistral.ai/)
|
|
||||||
+ [x] [百度文心一言系列模型](https://cloud.baidu.com/doc/WENXINWORKSHOP/index.html)
|
+ [x] [百度文心一言系列模型](https://cloud.baidu.com/doc/WENXINWORKSHOP/index.html)
|
||||||
+ [x] [阿里通义千问系列模型](https://help.aliyun.com/document_detail/2400395.html)
|
+ [x] [阿里通义千问系列模型](https://help.aliyun.com/document_detail/2400395.html)
|
||||||
+ [x] [讯飞星火认知大模型](https://www.xfyun.cn/doc/spark/Web.html)
|
+ [x] [讯飞星火认知大模型](https://www.xfyun.cn/doc/spark/Web.html)
|
||||||
+ [x] [智谱 ChatGLM 系列模型](https://bigmodel.cn)
|
+ [x] [智谱 ChatGLM 系列模型](https://bigmodel.cn)
|
||||||
+ [x] [360 智脑](https://ai.360.cn)
|
+ [x] [360 智脑](https://ai.360.cn)
|
||||||
+ [x] [腾讯混元大模型](https://cloud.tencent.com/document/product/1729)
|
+ [x] [腾讯混元大模型](https://cloud.tencent.com/document/product/1729)
|
||||||
+ [x] [Moonshot AI](https://platform.moonshot.cn/)
|
|
||||||
+ [x] [百川大模型](https://platform.baichuan-ai.com)
|
|
||||||
+ [ ] [字节云雀大模型](https://www.volcengine.com/product/ark) (WIP)
|
|
||||||
+ [x] [MINIMAX](https://api.minimax.chat/)
|
|
||||||
+ [x] [Groq](https://wow.groq.com/)
|
|
||||||
2. 支持配置镜像以及众多[第三方代理服务](https://iamazing.cn/page/openai-api-third-party-services)。
|
2. 支持配置镜像以及众多[第三方代理服务](https://iamazing.cn/page/openai-api-third-party-services)。
|
||||||
3. 支持通过**负载均衡**的方式访问多个渠道。
|
3. 支持通过**负载均衡**的方式访问多个渠道。
|
||||||
4. 支持 **stream 模式**,可以通过流式传输实现打字机效果。
|
4. 支持 **stream 模式**,可以通过流式传输实现打字机效果。
|
||||||
@@ -106,7 +100,6 @@ _✨ 通过标准的 OpenAI API 格式访问所有的大模型,开箱即用
|
|||||||
+ [GitHub 开放授权](https://github.com/settings/applications/new)。
|
+ [GitHub 开放授权](https://github.com/settings/applications/new)。
|
||||||
+ 微信公众号授权(需要额外部署 [WeChat Server](https://github.com/songquanpeng/wechat-server))。
|
+ 微信公众号授权(需要额外部署 [WeChat Server](https://github.com/songquanpeng/wechat-server))。
|
||||||
23. 支持主题切换,设置环境变量 `THEME` 即可,默认为 `default`,欢迎 PR 更多主题,具体参考[此处](./web/README.md)。
|
23. 支持主题切换,设置环境变量 `THEME` 即可,默认为 `default`,欢迎 PR 更多主题,具体参考[此处](./web/README.md)。
|
||||||
24. 配合 [Message Pusher](https://github.com/songquanpeng/message-pusher) 可将报警信息推送到多种 App 上。
|
|
||||||
|
|
||||||
## 部署
|
## 部署
|
||||||
### 基于 Docker 进行部署
|
### 基于 Docker 进行部署
|
||||||
@@ -181,12 +174,12 @@ docker-compose ps
|
|||||||
git clone https://github.com/songquanpeng/one-api.git
|
git clone https://github.com/songquanpeng/one-api.git
|
||||||
|
|
||||||
# 构建前端
|
# 构建前端
|
||||||
cd one-api/web/default
|
cd one-api/web
|
||||||
npm install
|
npm install
|
||||||
npm run build
|
npm run build
|
||||||
|
|
||||||
# 构建后端
|
# 构建后端
|
||||||
cd ../..
|
cd ..
|
||||||
go mod download
|
go mod download
|
||||||
go build -ldflags "-s -w" -o one-api
|
go build -ldflags "-s -w" -o one-api
|
||||||
````
|
````
|
||||||
@@ -376,9 +369,6 @@ graph LR
|
|||||||
16. `SQLITE_BUSY_TIMEOUT`:SQLite 锁等待超时设置,单位为毫秒,默认 `3000`。
|
16. `SQLITE_BUSY_TIMEOUT`:SQLite 锁等待超时设置,单位为毫秒,默认 `3000`。
|
||||||
17. `GEMINI_SAFETY_SETTING`:Gemini 的安全设置,默认 `BLOCK_NONE`。
|
17. `GEMINI_SAFETY_SETTING`:Gemini 的安全设置,默认 `BLOCK_NONE`。
|
||||||
18. `THEME`:系统的主题设置,默认为 `default`,具体可选值参考[此处](./web/README.md)。
|
18. `THEME`:系统的主题设置,默认为 `default`,具体可选值参考[此处](./web/README.md)。
|
||||||
19. `ENABLE_METRIC`:是否根据请求成功率禁用渠道,默认不开启,可选值为 `true` 和 `false`。
|
|
||||||
20. `METRIC_QUEUE_SIZE`:请求成功率统计队列大小,默认为 `10`。
|
|
||||||
21. `METRIC_SUCCESS_RATE_THRESHOLD`:请求成功率阈值,默认为 `0.8`。
|
|
||||||
|
|
||||||
### 命令行参数
|
### 命令行参数
|
||||||
1. `--port <port_number>`: 指定服务器监听的端口号,默认为 `3000`。
|
1. `--port <port_number>`: 指定服务器监听的端口号,默认为 `3000`。
|
||||||
|
|||||||
@@ -1,29 +0,0 @@
|
|||||||
package blacklist
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
"sync"
|
|
||||||
)
|
|
||||||
|
|
||||||
var blackList sync.Map
|
|
||||||
|
|
||||||
func init() {
|
|
||||||
blackList = sync.Map{}
|
|
||||||
}
|
|
||||||
|
|
||||||
func userId2Key(id int) string {
|
|
||||||
return fmt.Sprintf("userid_%d", id)
|
|
||||||
}
|
|
||||||
|
|
||||||
func BanUser(id int) {
|
|
||||||
blackList.Store(userId2Key(id), true)
|
|
||||||
}
|
|
||||||
|
|
||||||
func UnbanUser(id int) {
|
|
||||||
blackList.Delete(userId2Key(id))
|
|
||||||
}
|
|
||||||
|
|
||||||
func IsUserBanned(id int) bool {
|
|
||||||
_, ok := blackList.Load(userId2Key(id))
|
|
||||||
return ok
|
|
||||||
}
|
|
||||||
@@ -1,7 +1,7 @@
|
|||||||
package config
|
package config
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
"one-api/common/helper"
|
||||||
"os"
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
"sync"
|
"sync"
|
||||||
@@ -52,7 +52,6 @@ var EmailDomainWhitelist = []string{
|
|||||||
}
|
}
|
||||||
|
|
||||||
var DebugEnabled = os.Getenv("DEBUG") == "true"
|
var DebugEnabled = os.Getenv("DEBUG") == "true"
|
||||||
var DebugSQLEnabled = os.Getenv("DEBUG_SQL") == "true"
|
|
||||||
var MemoryCacheEnabled = os.Getenv("MEMORY_CACHE_ENABLED") == "true"
|
var MemoryCacheEnabled = os.Getenv("MEMORY_CACHE_ENABLED") == "true"
|
||||||
|
|
||||||
var LogConsumeEnabled = true
|
var LogConsumeEnabled = true
|
||||||
@@ -70,9 +69,6 @@ var WeChatServerAddress = ""
|
|||||||
var WeChatServerToken = ""
|
var WeChatServerToken = ""
|
||||||
var WeChatAccountQRCodeImageURL = ""
|
var WeChatAccountQRCodeImageURL = ""
|
||||||
|
|
||||||
var MessagePusherAddress = ""
|
|
||||||
var MessagePusherToken = ""
|
|
||||||
|
|
||||||
var TurnstileSiteKey = ""
|
var TurnstileSiteKey = ""
|
||||||
var TurnstileSecretKey = ""
|
var TurnstileSecretKey = ""
|
||||||
|
|
||||||
@@ -129,9 +125,3 @@ var (
|
|||||||
)
|
)
|
||||||
|
|
||||||
var RateLimitKeyExpirationDuration = 20 * time.Minute
|
var RateLimitKeyExpirationDuration = 20 * time.Minute
|
||||||
|
|
||||||
var EnableMetric = helper.GetOrDefaultEnvBool("ENABLE_METRIC", false)
|
|
||||||
var MetricQueueSize = helper.GetOrDefaultEnvInt("METRIC_QUEUE_SIZE", 10)
|
|
||||||
var MetricSuccessRateThreshold = helper.GetOrDefaultEnvFloat64("METRIC_SUCCESS_RATE_THRESHOLD", 0.8)
|
|
||||||
var MetricSuccessChanSize = helper.GetOrDefaultEnvInt("METRIC_SUCCESS_CHAN_SIZE", 1024)
|
|
||||||
var MetricFailChanSize = helper.GetOrDefaultEnvInt("METRIC_FAIL_CHAN_SIZE", 128)
|
|
||||||
|
|||||||
@@ -15,7 +15,6 @@ const (
|
|||||||
const (
|
const (
|
||||||
UserStatusEnabled = 1 // don't use 0, 0 is the default value!
|
UserStatusEnabled = 1 // don't use 0, 0 is the default value!
|
||||||
UserStatusDisabled = 2 // also don't use 0
|
UserStatusDisabled = 2 // also don't use 0
|
||||||
UserStatusDeleted = 3
|
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -39,38 +38,31 @@ const (
|
|||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
ChannelTypeUnknown = iota
|
ChannelTypeUnknown = 0
|
||||||
ChannelTypeOpenAI
|
ChannelTypeOpenAI = 1
|
||||||
ChannelTypeAPI2D
|
ChannelTypeAPI2D = 2
|
||||||
ChannelTypeAzure
|
ChannelTypeAzure = 3
|
||||||
ChannelTypeCloseAI
|
ChannelTypeCloseAI = 4
|
||||||
ChannelTypeOpenAISB
|
ChannelTypeOpenAISB = 5
|
||||||
ChannelTypeOpenAIMax
|
ChannelTypeOpenAIMax = 6
|
||||||
ChannelTypeOhMyGPT
|
ChannelTypeOhMyGPT = 7
|
||||||
ChannelTypeCustom
|
ChannelTypeCustom = 8
|
||||||
ChannelTypeAILS
|
ChannelTypeAILS = 9
|
||||||
ChannelTypeAIProxy
|
ChannelTypeAIProxy = 10
|
||||||
ChannelTypePaLM
|
ChannelTypePaLM = 11
|
||||||
ChannelTypeAPI2GPT
|
ChannelTypeAPI2GPT = 12
|
||||||
ChannelTypeAIGC2D
|
ChannelTypeAIGC2D = 13
|
||||||
ChannelTypeAnthropic
|
ChannelTypeAnthropic = 14
|
||||||
ChannelTypeBaidu
|
ChannelTypeBaidu = 15
|
||||||
ChannelTypeZhipu
|
ChannelTypeZhipu = 16
|
||||||
ChannelTypeAli
|
ChannelTypeAli = 17
|
||||||
ChannelTypeXunfei
|
ChannelTypeXunfei = 18
|
||||||
ChannelType360
|
ChannelType360 = 19
|
||||||
ChannelTypeOpenRouter
|
ChannelTypeOpenRouter = 20
|
||||||
ChannelTypeAIProxyLibrary
|
ChannelTypeAIProxyLibrary = 21
|
||||||
ChannelTypeFastGPT
|
ChannelTypeFastGPT = 22
|
||||||
ChannelTypeTencent
|
ChannelTypeTencent = 23
|
||||||
ChannelTypeGemini
|
ChannelTypeGemini = 24
|
||||||
ChannelTypeMoonshot
|
|
||||||
ChannelTypeBaichuan
|
|
||||||
ChannelTypeMinimax
|
|
||||||
ChannelTypeMistral
|
|
||||||
ChannelTypeGroq
|
|
||||||
|
|
||||||
ChannelTypeDummy
|
|
||||||
)
|
)
|
||||||
|
|
||||||
var ChannelBaseURLs = []string{
|
var ChannelBaseURLs = []string{
|
||||||
@@ -99,17 +91,4 @@ var ChannelBaseURLs = []string{
|
|||||||
"https://fastgpt.run/api/openapi", // 22
|
"https://fastgpt.run/api/openapi", // 22
|
||||||
"https://hunyuan.cloud.tencent.com", // 23
|
"https://hunyuan.cloud.tencent.com", // 23
|
||||||
"https://generativelanguage.googleapis.com", // 24
|
"https://generativelanguage.googleapis.com", // 24
|
||||||
"https://api.moonshot.cn", // 25
|
|
||||||
"https://api.baichuan-ai.com", // 26
|
|
||||||
"https://api.minimax.chat", // 27
|
|
||||||
"https://api.mistral.ai", // 28
|
|
||||||
"https://api.groq.com/openai", // 29
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const (
|
|
||||||
ConfigKeyPrefix = "cfg_"
|
|
||||||
|
|
||||||
ConfigKeyAPIVersion = ConfigKeyPrefix + "api_version"
|
|
||||||
ConfigKeyLibraryID = ConfigKeyPrefix + "library_id"
|
|
||||||
ConfigKeyPlugin = ConfigKeyPrefix + "plugin"
|
|
||||||
)
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
package common
|
package common
|
||||||
|
|
||||||
import "github.com/songquanpeng/one-api/common/helper"
|
import "one-api/common/helper"
|
||||||
|
|
||||||
var UsingSQLite = false
|
var UsingSQLite = false
|
||||||
var UsingPostgreSQL = false
|
var UsingPostgreSQL = false
|
||||||
|
|||||||
@@ -1,20 +1,17 @@
|
|||||||
package message
|
package common
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"crypto/rand"
|
"crypto/rand"
|
||||||
"crypto/tls"
|
"crypto/tls"
|
||||||
"encoding/base64"
|
"encoding/base64"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"net/smtp"
|
"net/smtp"
|
||||||
|
"one-api/common/config"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
func SendEmail(subject string, receiver string, content string) error {
|
func SendEmail(subject string, receiver string, content string) error {
|
||||||
if receiver == "" {
|
|
||||||
return fmt.Errorf("receiver is empty")
|
|
||||||
}
|
|
||||||
if config.SMTPFrom == "" { // for compatibility
|
if config.SMTPFrom == "" { // for compatibility
|
||||||
config.SMTPFrom = config.SMTPAccount
|
config.SMTPFrom = config.SMTPAccount
|
||||||
}
|
}
|
||||||
@@ -15,7 +15,10 @@ type embedFileSystem struct {
|
|||||||
|
|
||||||
func (e embedFileSystem) Exists(prefix string, path string) bool {
|
func (e embedFileSystem) Exists(prefix string, path string) bool {
|
||||||
_, err := e.Open(path)
|
_, err := e.Open(path)
|
||||||
return err == nil
|
if err != nil {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
func EmbedFolder(fsEmbed embed.FS, targetPath string) static.ServeFileSystem {
|
func EmbedFolder(fsEmbed embed.FS, targetPath string) static.ServeFileSystem {
|
||||||
|
|||||||
@@ -8,24 +8,12 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
const KeyRequestBody = "key_request_body"
|
func UnmarshalBodyReusable(c *gin.Context, v any) error {
|
||||||
|
|
||||||
func GetRequestBody(c *gin.Context) ([]byte, error) {
|
|
||||||
requestBody, _ := c.Get(KeyRequestBody)
|
|
||||||
if requestBody != nil {
|
|
||||||
return requestBody.([]byte), nil
|
|
||||||
}
|
|
||||||
requestBody, err := io.ReadAll(c.Request.Body)
|
requestBody, err := io.ReadAll(c.Request.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return err
|
||||||
}
|
}
|
||||||
_ = c.Request.Body.Close()
|
err = c.Request.Body.Close()
|
||||||
c.Set(KeyRequestBody, requestBody)
|
|
||||||
return requestBody.([]byte), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func UnmarshalBodyReusable(c *gin.Context, v any) error {
|
|
||||||
requestBody, err := GetRequestBody(c)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ package common
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
var GroupRatio = map[string]float64{
|
var GroupRatio = map[string]float64{
|
||||||
|
|||||||
@@ -3,11 +3,11 @@ package helper
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/google/uuid"
|
"github.com/google/uuid"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"html/template"
|
"html/template"
|
||||||
"log"
|
"log"
|
||||||
"math/rand"
|
"math/rand"
|
||||||
"net"
|
"net"
|
||||||
|
"one-api/common/logger"
|
||||||
"os"
|
"os"
|
||||||
"os/exec"
|
"os/exec"
|
||||||
"runtime"
|
"runtime"
|
||||||
@@ -107,13 +107,13 @@ func Seconds2Time(num int) (time string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func Interface2String(inter interface{}) string {
|
func Interface2String(inter interface{}) string {
|
||||||
switch inter := inter.(type) {
|
switch inter.(type) {
|
||||||
case string:
|
case string:
|
||||||
return inter
|
return inter.(string)
|
||||||
case int:
|
case int:
|
||||||
return fmt.Sprintf("%d", inter)
|
return fmt.Sprintf("%d", inter.(int))
|
||||||
case float64:
|
case float64:
|
||||||
return fmt.Sprintf("%f", inter)
|
return fmt.Sprintf("%f", inter.(float64))
|
||||||
}
|
}
|
||||||
return "Not Implemented"
|
return "Not Implemented"
|
||||||
}
|
}
|
||||||
@@ -137,7 +137,6 @@ func GetUUID() string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
const keyChars = "0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ"
|
const keyChars = "0123456789abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ"
|
||||||
const keyNumbers = "0123456789"
|
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
rand.Seed(time.Now().UnixNano())
|
rand.Seed(time.Now().UnixNano())
|
||||||
@@ -169,15 +168,6 @@ func GetRandomString(length int) string {
|
|||||||
return string(key)
|
return string(key)
|
||||||
}
|
}
|
||||||
|
|
||||||
func GetRandomNumberString(length int) string {
|
|
||||||
rand.Seed(time.Now().UnixNano())
|
|
||||||
key := make([]byte, length)
|
|
||||||
for i := 0; i < length; i++ {
|
|
||||||
key[i] = keyNumbers[rand.Intn(len(keyNumbers))]
|
|
||||||
}
|
|
||||||
return string(key)
|
|
||||||
}
|
|
||||||
|
|
||||||
func GetTimestamp() int64 {
|
func GetTimestamp() int64 {
|
||||||
return time.Now().Unix()
|
return time.Now().Unix()
|
||||||
}
|
}
|
||||||
@@ -195,13 +185,6 @@ func Max(a int, b int) int {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func GetOrDefaultEnvBool(env string, defaultValue bool) bool {
|
|
||||||
if env == "" || os.Getenv(env) == "" {
|
|
||||||
return defaultValue
|
|
||||||
}
|
|
||||||
return os.Getenv(env) == "true"
|
|
||||||
}
|
|
||||||
|
|
||||||
func GetOrDefaultEnvInt(env string, defaultValue int) int {
|
func GetOrDefaultEnvInt(env string, defaultValue int) int {
|
||||||
if env == "" || os.Getenv(env) == "" {
|
if env == "" || os.Getenv(env) == "" {
|
||||||
return defaultValue
|
return defaultValue
|
||||||
@@ -214,18 +197,6 @@ func GetOrDefaultEnvInt(env string, defaultValue int) int {
|
|||||||
return num
|
return num
|
||||||
}
|
}
|
||||||
|
|
||||||
func GetOrDefaultEnvFloat64(env string, defaultValue float64) float64 {
|
|
||||||
if env == "" || os.Getenv(env) == "" {
|
|
||||||
return defaultValue
|
|
||||||
}
|
|
||||||
num, err := strconv.ParseFloat(os.Getenv(env), 64)
|
|
||||||
if err != nil {
|
|
||||||
logger.SysError(fmt.Sprintf("failed to parse %s: %s, using default value: %f", env, err.Error(), defaultValue))
|
|
||||||
return defaultValue
|
|
||||||
}
|
|
||||||
return num
|
|
||||||
}
|
|
||||||
|
|
||||||
func GetOrDefaultEnvString(env string, defaultValue string) string {
|
func GetOrDefaultEnvString(env string, defaultValue string) string {
|
||||||
if env == "" || os.Getenv(env) == "" {
|
if env == "" || os.Getenv(env) == "" {
|
||||||
return defaultValue
|
return defaultValue
|
||||||
|
|||||||
@@ -12,7 +12,7 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
img "github.com/songquanpeng/one-api/common/image"
|
img "one-api/common/image"
|
||||||
|
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
_ "golang.org/x/image/webp"
|
_ "golang.org/x/image/webp"
|
||||||
|
|||||||
@@ -3,9 +3,9 @@ package common
|
|||||||
import (
|
import (
|
||||||
"flag"
|
"flag"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"log"
|
"log"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/logger"
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -13,7 +13,6 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
loggerDEBUG = "DEBUG"
|
|
||||||
loggerINFO = "INFO"
|
loggerINFO = "INFO"
|
||||||
loggerWarn = "WARN"
|
loggerWarn = "WARN"
|
||||||
loggerError = "ERR"
|
loggerError = "ERR"
|
||||||
@@ -56,10 +55,6 @@ func SysError(s string) {
|
|||||||
_, _ = fmt.Fprintf(gin.DefaultErrorWriter, "[SYS] %v | %s \n", t.Format("2006/01/02 - 15:04:05"), s)
|
_, _ = fmt.Fprintf(gin.DefaultErrorWriter, "[SYS] %v | %s \n", t.Format("2006/01/02 - 15:04:05"), s)
|
||||||
}
|
}
|
||||||
|
|
||||||
func Debug(ctx context.Context, msg string) {
|
|
||||||
logHelper(ctx, loggerDEBUG, msg)
|
|
||||||
}
|
|
||||||
|
|
||||||
func Info(ctx context.Context, msg string) {
|
func Info(ctx context.Context, msg string) {
|
||||||
logHelper(ctx, loggerINFO, msg)
|
logHelper(ctx, loggerINFO, msg)
|
||||||
}
|
}
|
||||||
@@ -72,20 +67,16 @@ func Error(ctx context.Context, msg string) {
|
|||||||
logHelper(ctx, loggerError, msg)
|
logHelper(ctx, loggerError, msg)
|
||||||
}
|
}
|
||||||
|
|
||||||
func Debugf(ctx context.Context, format string, a ...any) {
|
|
||||||
Debug(ctx, fmt.Sprintf(format, a...))
|
|
||||||
}
|
|
||||||
|
|
||||||
func Infof(ctx context.Context, format string, a ...any) {
|
func Infof(ctx context.Context, format string, a ...any) {
|
||||||
Info(ctx, fmt.Sprintf(format, a...))
|
Info(ctx, fmt.Sprintf(format, a))
|
||||||
}
|
}
|
||||||
|
|
||||||
func Warnf(ctx context.Context, format string, a ...any) {
|
func Warnf(ctx context.Context, format string, a ...any) {
|
||||||
Warn(ctx, fmt.Sprintf(format, a...))
|
Warn(ctx, fmt.Sprintf(format, a))
|
||||||
}
|
}
|
||||||
|
|
||||||
func Errorf(ctx context.Context, format string, a ...any) {
|
func Errorf(ctx context.Context, format string, a ...any) {
|
||||||
Error(ctx, fmt.Sprintf(format, a...))
|
Error(ctx, fmt.Sprintf(format, a))
|
||||||
}
|
}
|
||||||
|
|
||||||
func logHelper(ctx context.Context, level string, msg string) {
|
func logHelper(ctx context.Context, level string, msg string) {
|
||||||
|
|||||||
@@ -1,22 +0,0 @@
|
|||||||
package message
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
)
|
|
||||||
|
|
||||||
const (
|
|
||||||
ByAll = "all"
|
|
||||||
ByEmail = "email"
|
|
||||||
ByMessagePusher = "message_pusher"
|
|
||||||
)
|
|
||||||
|
|
||||||
func Notify(by string, title string, description string, content string) error {
|
|
||||||
if by == ByEmail {
|
|
||||||
return SendEmail(title, config.RootUserEmail, content)
|
|
||||||
}
|
|
||||||
if by == ByMessagePusher {
|
|
||||||
return SendMessage(title, description, content)
|
|
||||||
}
|
|
||||||
return fmt.Errorf("unknown notify method: %s", by)
|
|
||||||
}
|
|
||||||
@@ -1,53 +0,0 @@
|
|||||||
package message
|
|
||||||
|
|
||||||
import (
|
|
||||||
"bytes"
|
|
||||||
"encoding/json"
|
|
||||||
"errors"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"net/http"
|
|
||||||
)
|
|
||||||
|
|
||||||
type request struct {
|
|
||||||
Title string `json:"title"`
|
|
||||||
Description string `json:"description"`
|
|
||||||
Content string `json:"content"`
|
|
||||||
URL string `json:"url"`
|
|
||||||
Channel string `json:"channel"`
|
|
||||||
Token string `json:"token"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type response struct {
|
|
||||||
Success bool `json:"success"`
|
|
||||||
Message string `json:"message"`
|
|
||||||
}
|
|
||||||
|
|
||||||
func SendMessage(title string, description string, content string) error {
|
|
||||||
if config.MessagePusherAddress == "" {
|
|
||||||
return errors.New("message pusher address is not set")
|
|
||||||
}
|
|
||||||
req := request{
|
|
||||||
Title: title,
|
|
||||||
Description: description,
|
|
||||||
Content: content,
|
|
||||||
Token: config.MessagePusherToken,
|
|
||||||
}
|
|
||||||
data, err := json.Marshal(req)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
resp, err := http.Post(config.MessagePusherAddress,
|
|
||||||
"application/json", bytes.NewBuffer(data))
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
var res response
|
|
||||||
err = json.NewDecoder(resp.Body).Decode(&res)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
if !res.Success {
|
|
||||||
return errors.New(res.Message)
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
@@ -2,89 +2,92 @@ package common
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
var DalleSizeRatios = map[string]map[string]float64{
|
||||||
USD2RMB = 7
|
"dall-e-2": {
|
||||||
USD = 500 // $0.002 = 1 -> $1 = 500
|
"256x256": 1,
|
||||||
RMB = USD / USD2RMB
|
"512x512": 1.125,
|
||||||
)
|
"1024x1024": 1.25,
|
||||||
|
},
|
||||||
|
"dall-e-3": {
|
||||||
|
"1024x1024": 1,
|
||||||
|
"1024x1792": 2,
|
||||||
|
"1792x1024": 2,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
var DalleGenerationImageAmounts = map[string][2]int{
|
||||||
|
"dall-e-2": {1, 10},
|
||||||
|
"dall-e-3": {1, 1}, // OpenAI allows n=1 currently.
|
||||||
|
}
|
||||||
|
|
||||||
|
var DalleImagePromptLengthLimitations = map[string]int{
|
||||||
|
"dall-e-2": 1000,
|
||||||
|
"dall-e-3": 4000,
|
||||||
|
}
|
||||||
|
|
||||||
// ModelRatio
|
// ModelRatio
|
||||||
// https://platform.openai.com/docs/models/model-endpoint-compatibility
|
// https://platform.openai.com/docs/models/model-endpoint-compatibility
|
||||||
// https://cloud.baidu.com/doc/WENXINWORKSHOP/s/Blfmc9dlf
|
// https://cloud.baidu.com/doc/WENXINWORKSHOP/s/Blfmc9dlf
|
||||||
// https://openai.com/pricing
|
// https://openai.com/pricing
|
||||||
|
// TODO: when a new api is enabled, check the pricing here
|
||||||
// 1 === $0.002 / 1K tokens
|
// 1 === $0.002 / 1K tokens
|
||||||
// 1 === ¥0.014 / 1k tokens
|
// 1 === ¥0.014 / 1k tokens
|
||||||
var ModelRatio = map[string]float64{
|
var ModelRatio = map[string]float64{
|
||||||
// https://openai.com/pricing
|
"gpt-4": 15,
|
||||||
"gpt-4": 15,
|
"gpt-4-0314": 15,
|
||||||
"gpt-4-0314": 15,
|
"gpt-4-0613": 15,
|
||||||
"gpt-4-0613": 15,
|
"gpt-4-32k": 30,
|
||||||
"gpt-4-32k": 30,
|
"gpt-4-32k-0314": 30,
|
||||||
"gpt-4-32k-0314": 30,
|
"gpt-4-32k-0613": 30,
|
||||||
"gpt-4-32k-0613": 30,
|
"gpt-4-1106-preview": 5, // $0.01 / 1K tokens
|
||||||
"gpt-4-1106-preview": 5, // $0.01 / 1K tokens
|
"gpt-4-vision-preview": 5, // $0.01 / 1K tokens
|
||||||
"gpt-4-0125-preview": 5, // $0.01 / 1K tokens
|
"gpt-3.5-turbo": 0.75, // $0.0015 / 1K tokens
|
||||||
"gpt-4-turbo-preview": 5, // $0.01 / 1K tokens
|
"gpt-3.5-turbo-0301": 0.75,
|
||||||
"gpt-4-vision-preview": 5, // $0.01 / 1K tokens
|
"gpt-3.5-turbo-0613": 0.75,
|
||||||
"gpt-3.5-turbo": 0.75, // $0.0015 / 1K tokens
|
"gpt-3.5-turbo-16k": 1.5, // $0.003 / 1K tokens
|
||||||
"gpt-3.5-turbo-0301": 0.75,
|
"gpt-3.5-turbo-16k-0613": 1.5,
|
||||||
"gpt-3.5-turbo-0613": 0.75,
|
"gpt-3.5-turbo-instruct": 0.75, // $0.0015 / 1K tokens
|
||||||
"gpt-3.5-turbo-16k": 1.5, // $0.003 / 1K tokens
|
"gpt-3.5-turbo-1106": 0.5, // $0.001 / 1K tokens
|
||||||
"gpt-3.5-turbo-16k-0613": 1.5,
|
"davinci-002": 1, // $0.002 / 1K tokens
|
||||||
"gpt-3.5-turbo-instruct": 0.75, // $0.0015 / 1K tokens
|
"babbage-002": 0.2, // $0.0004 / 1K tokens
|
||||||
"gpt-3.5-turbo-1106": 0.5, // $0.001 / 1K tokens
|
"text-ada-001": 0.2,
|
||||||
"gpt-3.5-turbo-0125": 0.25, // $0.0005 / 1K tokens
|
"text-babbage-001": 0.25,
|
||||||
"davinci-002": 1, // $0.002 / 1K tokens
|
"text-curie-001": 1,
|
||||||
"babbage-002": 0.2, // $0.0004 / 1K tokens
|
"text-davinci-002": 10,
|
||||||
"text-ada-001": 0.2,
|
"text-davinci-003": 10,
|
||||||
"text-babbage-001": 0.25,
|
"text-davinci-edit-001": 10,
|
||||||
"text-curie-001": 1,
|
"code-davinci-edit-001": 10,
|
||||||
"text-davinci-002": 10,
|
"whisper-1": 15, // $0.006 / minute -> $0.006 / 150 words -> $0.006 / 200 tokens -> $0.03 / 1k tokens
|
||||||
"text-davinci-003": 10,
|
"tts-1": 7.5, // $0.015 / 1K characters
|
||||||
"text-davinci-edit-001": 10,
|
"tts-1-1106": 7.5,
|
||||||
"code-davinci-edit-001": 10,
|
"tts-1-hd": 15, // $0.030 / 1K characters
|
||||||
"whisper-1": 15, // $0.006 / minute -> $0.006 / 150 words -> $0.006 / 200 tokens -> $0.03 / 1k tokens
|
"tts-1-hd-1106": 15,
|
||||||
"tts-1": 7.5, // $0.015 / 1K characters
|
"davinci": 10,
|
||||||
"tts-1-1106": 7.5,
|
"curie": 10,
|
||||||
"tts-1-hd": 15, // $0.030 / 1K characters
|
"babbage": 10,
|
||||||
"tts-1-hd-1106": 15,
|
"ada": 10,
|
||||||
"davinci": 10,
|
"text-embedding-ada-002": 0.05,
|
||||||
"curie": 10,
|
"text-search-ada-doc-001": 10,
|
||||||
"babbage": 10,
|
"text-moderation-stable": 0.1,
|
||||||
"ada": 10,
|
"text-moderation-latest": 0.1,
|
||||||
"text-embedding-ada-002": 0.05,
|
"dall-e-2": 8, // $0.016 - $0.020 / image
|
||||||
"text-embedding-3-small": 0.01,
|
"dall-e-3": 20, // $0.040 - $0.120 / image
|
||||||
"text-embedding-3-large": 0.065,
|
"claude-instant-1": 0.815, // $1.63 / 1M tokens
|
||||||
"text-search-ada-doc-001": 10,
|
"claude-2": 5.51, // $11.02 / 1M tokens
|
||||||
"text-moderation-stable": 0.1,
|
"claude-2.0": 5.51, // $11.02 / 1M tokens
|
||||||
"text-moderation-latest": 0.1,
|
"claude-2.1": 5.51, // $11.02 / 1M tokens
|
||||||
"dall-e-2": 8, // $0.016 - $0.020 / image
|
"ERNIE-Bot": 0.8572, // ¥0.012 / 1k tokens
|
||||||
"dall-e-3": 20, // $0.040 - $0.120 / image
|
"ERNIE-Bot-turbo": 0.5715, // ¥0.008 / 1k tokens
|
||||||
// https://www.anthropic.com/api#pricing
|
"ERNIE-Bot-4": 8.572, // ¥0.12 / 1k tokens
|
||||||
"claude-instant-1.2": 0.8 / 1000 * USD,
|
"Embedding-V1": 0.1429, // ¥0.002 / 1k tokens
|
||||||
"claude-2.0": 8.0 / 1000 * USD,
|
"PaLM-2": 1,
|
||||||
"claude-2.1": 8.0 / 1000 * USD,
|
"gemini-pro": 1, // $0.00025 / 1k characters -> $0.001 / 1k tokens
|
||||||
"claude-3-haiku-20240229": 0.25 / 1000 * USD,
|
"gemini-pro-vision": 1, // $0.00025 / 1k characters -> $0.001 / 1k tokens
|
||||||
"claude-3-sonnet-20240229": 3.0 / 1000 * USD,
|
|
||||||
"claude-3-opus-20240229": 15.0 / 1000 * USD,
|
|
||||||
// https://cloud.baidu.com/doc/WENXINWORKSHOP/s/hlrk4akp7
|
|
||||||
"ERNIE-Bot": 0.8572, // ¥0.012 / 1k tokens
|
|
||||||
"ERNIE-Bot-turbo": 0.5715, // ¥0.008 / 1k tokens
|
|
||||||
"ERNIE-Bot-4": 0.12 * RMB, // ¥0.12 / 1k tokens
|
|
||||||
"ERNIE-Bot-8k": 0.024 * RMB,
|
|
||||||
"Embedding-V1": 0.1429, // ¥0.002 / 1k tokens
|
|
||||||
"PaLM-2": 1,
|
|
||||||
"gemini-pro": 1, // $0.00025 / 1k characters -> $0.001 / 1k tokens
|
|
||||||
"gemini-pro-vision": 1, // $0.00025 / 1k characters -> $0.001 / 1k tokens
|
|
||||||
// https://open.bigmodel.cn/pricing
|
|
||||||
"glm-4": 0.1 * RMB,
|
|
||||||
"glm-4v": 0.1 * RMB,
|
|
||||||
"glm-3-turbo": 0.005 * RMB,
|
|
||||||
"chatglm_turbo": 0.3572, // ¥0.005 / 1k tokens
|
"chatglm_turbo": 0.3572, // ¥0.005 / 1k tokens
|
||||||
"chatglm_pro": 0.7143, // ¥0.01 / 1k tokens
|
"chatglm_pro": 0.7143, // ¥0.01 / 1k tokens
|
||||||
"chatglm_std": 0.3572, // ¥0.005 / 1k tokens
|
"chatglm_std": 0.3572, // ¥0.005 / 1k tokens
|
||||||
@@ -95,57 +98,11 @@ var ModelRatio = map[string]float64{
|
|||||||
"qwen-max-longcontext": 1.4286, // ¥0.02 / 1k tokens
|
"qwen-max-longcontext": 1.4286, // ¥0.02 / 1k tokens
|
||||||
"text-embedding-v1": 0.05, // ¥0.0007 / 1k tokens
|
"text-embedding-v1": 0.05, // ¥0.0007 / 1k tokens
|
||||||
"SparkDesk": 1.2858, // ¥0.018 / 1k tokens
|
"SparkDesk": 1.2858, // ¥0.018 / 1k tokens
|
||||||
"SparkDesk-v1.1": 1.2858, // ¥0.018 / 1k tokens
|
|
||||||
"SparkDesk-v2.1": 1.2858, // ¥0.018 / 1k tokens
|
|
||||||
"SparkDesk-v3.1": 1.2858, // ¥0.018 / 1k tokens
|
|
||||||
"SparkDesk-v3.5": 1.2858, // ¥0.018 / 1k tokens
|
|
||||||
"360GPT_S2_V9": 0.8572, // ¥0.012 / 1k tokens
|
"360GPT_S2_V9": 0.8572, // ¥0.012 / 1k tokens
|
||||||
"embedding-bert-512-v1": 0.0715, // ¥0.001 / 1k tokens
|
"embedding-bert-512-v1": 0.0715, // ¥0.001 / 1k tokens
|
||||||
"embedding_s1_v1": 0.0715, // ¥0.001 / 1k tokens
|
"embedding_s1_v1": 0.0715, // ¥0.001 / 1k tokens
|
||||||
"semantic_similarity_s1_v1": 0.0715, // ¥0.001 / 1k tokens
|
"semantic_similarity_s1_v1": 0.0715, // ¥0.001 / 1k tokens
|
||||||
"hunyuan": 7.143, // ¥0.1 / 1k tokens // https://cloud.tencent.com/document/product/1729/97731#e0e6be58-60c8-469f-bdeb-6c264ce3b4d0
|
"hunyuan": 7.143, // ¥0.1 / 1k tokens // https://cloud.tencent.com/document/product/1729/97731#e0e6be58-60c8-469f-bdeb-6c264ce3b4d0
|
||||||
"ChatStd": 0.01 * RMB,
|
|
||||||
"ChatPro": 0.1 * RMB,
|
|
||||||
// https://platform.moonshot.cn/pricing
|
|
||||||
"moonshot-v1-8k": 0.012 * RMB,
|
|
||||||
"moonshot-v1-32k": 0.024 * RMB,
|
|
||||||
"moonshot-v1-128k": 0.06 * RMB,
|
|
||||||
// https://platform.baichuan-ai.com/price
|
|
||||||
"Baichuan2-Turbo": 0.008 * RMB,
|
|
||||||
"Baichuan2-Turbo-192k": 0.016 * RMB,
|
|
||||||
"Baichuan2-53B": 0.02 * RMB,
|
|
||||||
// https://api.minimax.chat/document/price
|
|
||||||
"abab6-chat": 0.1 * RMB,
|
|
||||||
"abab5.5-chat": 0.015 * RMB,
|
|
||||||
"abab5.5s-chat": 0.005 * RMB,
|
|
||||||
// https://docs.mistral.ai/platform/pricing/
|
|
||||||
"open-mistral-7b": 0.25 / 1000 * USD,
|
|
||||||
"open-mixtral-8x7b": 0.7 / 1000 * USD,
|
|
||||||
"mistral-small-latest": 2.0 / 1000 * USD,
|
|
||||||
"mistral-medium-latest": 2.7 / 1000 * USD,
|
|
||||||
"mistral-large-latest": 8.0 / 1000 * USD,
|
|
||||||
"mistral-embed": 0.1 / 1000 * USD,
|
|
||||||
// https://wow.groq.com/
|
|
||||||
"llama2-70b-4096": 0.7 / 1000 * USD,
|
|
||||||
"llama2-7b-2048": 0.1 / 1000 * USD,
|
|
||||||
"mixtral-8x7b-32768": 0.27 / 1000 * USD,
|
|
||||||
"gemma-7b-it": 0.1 / 1000 * USD,
|
|
||||||
}
|
|
||||||
|
|
||||||
var CompletionRatio = map[string]float64{}
|
|
||||||
|
|
||||||
var DefaultModelRatio map[string]float64
|
|
||||||
var DefaultCompletionRatio map[string]float64
|
|
||||||
|
|
||||||
func init() {
|
|
||||||
DefaultModelRatio = make(map[string]float64)
|
|
||||||
for k, v := range ModelRatio {
|
|
||||||
DefaultModelRatio[k] = v
|
|
||||||
}
|
|
||||||
DefaultCompletionRatio = make(map[string]float64)
|
|
||||||
for k, v := range CompletionRatio {
|
|
||||||
DefaultCompletionRatio[k] = v
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func ModelRatio2JSONString() string {
|
func ModelRatio2JSONString() string {
|
||||||
@@ -166,9 +123,6 @@ func GetModelRatio(name string) float64 {
|
|||||||
name = strings.TrimSuffix(name, "-internet")
|
name = strings.TrimSuffix(name, "-internet")
|
||||||
}
|
}
|
||||||
ratio, ok := ModelRatio[name]
|
ratio, ok := ModelRatio[name]
|
||||||
if !ok {
|
|
||||||
ratio, ok = DefaultModelRatio[name]
|
|
||||||
}
|
|
||||||
if !ok {
|
if !ok {
|
||||||
logger.SysError("model ratio not found: " + name)
|
logger.SysError("model ratio not found: " + name)
|
||||||
return 30
|
return 30
|
||||||
@@ -176,32 +130,8 @@ func GetModelRatio(name string) float64 {
|
|||||||
return ratio
|
return ratio
|
||||||
}
|
}
|
||||||
|
|
||||||
func CompletionRatio2JSONString() string {
|
|
||||||
jsonBytes, err := json.Marshal(CompletionRatio)
|
|
||||||
if err != nil {
|
|
||||||
logger.SysError("error marshalling completion ratio: " + err.Error())
|
|
||||||
}
|
|
||||||
return string(jsonBytes)
|
|
||||||
}
|
|
||||||
|
|
||||||
func UpdateCompletionRatioByJSONString(jsonStr string) error {
|
|
||||||
CompletionRatio = make(map[string]float64)
|
|
||||||
return json.Unmarshal([]byte(jsonStr), &CompletionRatio)
|
|
||||||
}
|
|
||||||
|
|
||||||
func GetCompletionRatio(name string) float64 {
|
func GetCompletionRatio(name string) float64 {
|
||||||
if ratio, ok := CompletionRatio[name]; ok {
|
|
||||||
return ratio
|
|
||||||
}
|
|
||||||
if ratio, ok := DefaultCompletionRatio[name]; ok {
|
|
||||||
return ratio
|
|
||||||
}
|
|
||||||
if strings.HasPrefix(name, "gpt-3.5") {
|
if strings.HasPrefix(name, "gpt-3.5") {
|
||||||
if strings.HasSuffix(name, "0125") {
|
|
||||||
// https://openai.com/blog/new-embedding-models-and-api-updates
|
|
||||||
// Updated GPT-3.5 Turbo model and lower pricing
|
|
||||||
return 3
|
|
||||||
}
|
|
||||||
if strings.HasSuffix(name, "1106") {
|
if strings.HasSuffix(name, "1106") {
|
||||||
return 2
|
return 2
|
||||||
}
|
}
|
||||||
@@ -214,7 +144,7 @@ func GetCompletionRatio(name string) float64 {
|
|||||||
return 2
|
return 2
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return 4.0 / 3.0
|
return 1.333333
|
||||||
}
|
}
|
||||||
if strings.HasPrefix(name, "gpt-4") {
|
if strings.HasPrefix(name, "gpt-4") {
|
||||||
if strings.HasSuffix(name, "preview") {
|
if strings.HasSuffix(name, "preview") {
|
||||||
@@ -222,18 +152,11 @@ func GetCompletionRatio(name string) float64 {
|
|||||||
}
|
}
|
||||||
return 2
|
return 2
|
||||||
}
|
}
|
||||||
if strings.HasPrefix(name, "claude-3") {
|
if strings.HasPrefix(name, "claude-instant-1") {
|
||||||
return 5
|
return 3.38
|
||||||
}
|
}
|
||||||
if strings.HasPrefix(name, "claude-") {
|
if strings.HasPrefix(name, "claude-2") {
|
||||||
return 3
|
return 2.965517
|
||||||
}
|
|
||||||
if strings.HasPrefix(name, "mistral-") {
|
|
||||||
return 3
|
|
||||||
}
|
|
||||||
switch name {
|
|
||||||
case "llama2-70b-4096":
|
|
||||||
return 0.8 / 0.7
|
|
||||||
}
|
}
|
||||||
return 1
|
return 1
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,8 +0,0 @@
|
|||||||
package common
|
|
||||||
|
|
||||||
import "math/rand"
|
|
||||||
|
|
||||||
// RandRange returns a random number between min and max (max is not included)
|
|
||||||
func RandRange(min, max int) int {
|
|
||||||
return min + rand.Intn(max-min)
|
|
||||||
}
|
|
||||||
@@ -3,7 +3,7 @@ package common
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/go-redis/redis/v8"
|
"github.com/go-redis/redis/v8"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
"os"
|
"os"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -2,7 +2,7 @@ package common
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
"one-api/common/config"
|
||||||
)
|
)
|
||||||
|
|
||||||
func LogQuota(quota int) string {
|
func LogQuota(quota int) string {
|
||||||
|
|||||||
@@ -2,9 +2,9 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
"one-api/common/config"
|
||||||
"github.com/songquanpeng/one-api/model"
|
"one-api/model"
|
||||||
relaymodel "github.com/songquanpeng/one-api/relay/model"
|
"one-api/relay/channel/openai"
|
||||||
)
|
)
|
||||||
|
|
||||||
func GetSubscription(c *gin.Context) {
|
func GetSubscription(c *gin.Context) {
|
||||||
@@ -22,15 +22,13 @@ func GetSubscription(c *gin.Context) {
|
|||||||
} else {
|
} else {
|
||||||
userId := c.GetInt("id")
|
userId := c.GetInt("id")
|
||||||
remainQuota, err = model.GetUserQuota(userId)
|
remainQuota, err = model.GetUserQuota(userId)
|
||||||
if err != nil {
|
usedQuota, err = model.GetUserUsedQuota(userId)
|
||||||
usedQuota, err = model.GetUserUsedQuota(userId)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
if expiredTime <= 0 {
|
if expiredTime <= 0 {
|
||||||
expiredTime = 0
|
expiredTime = 0
|
||||||
}
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
Error := relaymodel.Error{
|
Error := openai.Error{
|
||||||
Message: err.Error(),
|
Message: err.Error(),
|
||||||
Type: "upstream_error",
|
Type: "upstream_error",
|
||||||
}
|
}
|
||||||
@@ -72,7 +70,7 @@ func GetUsage(c *gin.Context) {
|
|||||||
quota, err = model.GetUserUsedQuota(userId)
|
quota, err = model.GetUserUsedQuota(userId)
|
||||||
}
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
Error := relaymodel.Error{
|
Error := openai.Error{
|
||||||
Message: err.Error(),
|
Message: err.Error(),
|
||||||
Type: "one_api_error",
|
Type: "one_api_error",
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -4,14 +4,13 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"github.com/songquanpeng/one-api/monitor"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/model"
|
||||||
|
"one-api/relay/util"
|
||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
@@ -296,7 +295,7 @@ func UpdateChannelBalance(c *gin.Context) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func updateAllChannelsBalance() error {
|
func updateAllChannelsBalance() error {
|
||||||
channels, err := model.GetAllChannels(0, 0, "all")
|
channels, err := model.GetAllChannels(0, 0, true)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -314,7 +313,7 @@ func updateAllChannelsBalance() error {
|
|||||||
} else {
|
} else {
|
||||||
// err is nil & balance <= 0 means quota is used up
|
// err is nil & balance <= 0 means quota is used up
|
||||||
if balance <= 0 {
|
if balance <= 0 {
|
||||||
monitor.DisableChannel(channel.Id, channel.Name, "余额不足")
|
disableChannel(channel.Id, channel.Name, "余额不足")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
time.Sleep(config.RequestInterval)
|
time.Sleep(config.RequestInterval)
|
||||||
|
|||||||
@@ -5,36 +5,102 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/common/message"
|
|
||||||
"github.com/songquanpeng/one-api/middleware"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"github.com/songquanpeng/one-api/monitor"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/helper"
|
|
||||||
relaymodel "github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/http/httptest"
|
"one-api/common"
|
||||||
"net/url"
|
"one-api/common/config"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/model"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
|
"one-api/relay/util"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
)
|
)
|
||||||
|
|
||||||
func buildTestRequest() *relaymodel.GeneralOpenAIRequest {
|
func testChannel(channel *model.Channel, request openai.ChatRequest) (err error, openaiErr *openai.Error) {
|
||||||
testRequest := &relaymodel.GeneralOpenAIRequest{
|
switch channel.Type {
|
||||||
MaxTokens: 1,
|
case common.ChannelTypePaLM:
|
||||||
Stream: false,
|
fallthrough
|
||||||
Model: "gpt-3.5-turbo",
|
case common.ChannelTypeGemini:
|
||||||
|
fallthrough
|
||||||
|
case common.ChannelTypeAnthropic:
|
||||||
|
fallthrough
|
||||||
|
case common.ChannelTypeBaidu:
|
||||||
|
fallthrough
|
||||||
|
case common.ChannelTypeZhipu:
|
||||||
|
fallthrough
|
||||||
|
case common.ChannelTypeAli:
|
||||||
|
fallthrough
|
||||||
|
case common.ChannelType360:
|
||||||
|
fallthrough
|
||||||
|
case common.ChannelTypeXunfei:
|
||||||
|
return errors.New("该渠道类型当前版本不支持测试,请手动测试"), nil
|
||||||
|
case common.ChannelTypeAzure:
|
||||||
|
request.Model = "gpt-35-turbo"
|
||||||
|
defer func() {
|
||||||
|
if err != nil {
|
||||||
|
err = errors.New("请确保已在 Azure 上创建了 gpt-35-turbo 模型,并且 apiVersion 已正确填写!")
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
default:
|
||||||
|
request.Model = "gpt-3.5-turbo"
|
||||||
}
|
}
|
||||||
testMessage := relaymodel.Message{
|
requestURL := common.ChannelBaseURLs[channel.Type]
|
||||||
|
if channel.Type == common.ChannelTypeAzure {
|
||||||
|
requestURL = util.GetFullRequestURL(channel.GetBaseURL(), fmt.Sprintf("/openai/deployments/%s/chat/completions?api-version=2023-03-15-preview", request.Model), channel.Type)
|
||||||
|
} else {
|
||||||
|
if baseURL := channel.GetBaseURL(); len(baseURL) > 0 {
|
||||||
|
requestURL = baseURL
|
||||||
|
}
|
||||||
|
|
||||||
|
requestURL = util.GetFullRequestURL(requestURL, "/v1/chat/completions", channel.Type)
|
||||||
|
}
|
||||||
|
jsonData, err := json.Marshal(request)
|
||||||
|
if err != nil {
|
||||||
|
return err, nil
|
||||||
|
}
|
||||||
|
req, err := http.NewRequest("POST", requestURL, bytes.NewBuffer(jsonData))
|
||||||
|
if err != nil {
|
||||||
|
return err, nil
|
||||||
|
}
|
||||||
|
if channel.Type == common.ChannelTypeAzure {
|
||||||
|
req.Header.Set("api-key", channel.Key)
|
||||||
|
} else {
|
||||||
|
req.Header.Set("Authorization", "Bearer "+channel.Key)
|
||||||
|
}
|
||||||
|
req.Header.Set("Content-Type", "application/json")
|
||||||
|
resp, err := util.HTTPClient.Do(req)
|
||||||
|
if err != nil {
|
||||||
|
return err, nil
|
||||||
|
}
|
||||||
|
defer resp.Body.Close()
|
||||||
|
var response openai.SlimTextResponse
|
||||||
|
body, err := io.ReadAll(resp.Body)
|
||||||
|
if err != nil {
|
||||||
|
return err, nil
|
||||||
|
}
|
||||||
|
err = json.Unmarshal(body, &response)
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("Error: %s\nResp body: %s", err, body), nil
|
||||||
|
}
|
||||||
|
if response.Usage.CompletionTokens == 0 {
|
||||||
|
if response.Error.Message == "" {
|
||||||
|
response.Error.Message = "补全 tokens 非预期返回 0"
|
||||||
|
}
|
||||||
|
return errors.New(fmt.Sprintf("type %s, code %v, message %s", response.Error.Type, response.Error.Code, response.Error.Message)), &response.Error
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func buildTestRequest() *openai.ChatRequest {
|
||||||
|
testRequest := &openai.ChatRequest{
|
||||||
|
Model: "", // this will be set later
|
||||||
|
MaxTokens: 1,
|
||||||
|
}
|
||||||
|
testMessage := openai.Message{
|
||||||
Role: "user",
|
Role: "user",
|
||||||
Content: "hi",
|
Content: "hi",
|
||||||
}
|
}
|
||||||
@@ -42,72 +108,6 @@ func buildTestRequest() *relaymodel.GeneralOpenAIRequest {
|
|||||||
return testRequest
|
return testRequest
|
||||||
}
|
}
|
||||||
|
|
||||||
func testChannel(channel *model.Channel) (err error, openaiErr *relaymodel.Error) {
|
|
||||||
w := httptest.NewRecorder()
|
|
||||||
c, _ := gin.CreateTestContext(w)
|
|
||||||
c.Request = &http.Request{
|
|
||||||
Method: "POST",
|
|
||||||
URL: &url.URL{Path: "/v1/chat/completions"},
|
|
||||||
Body: nil,
|
|
||||||
Header: make(http.Header),
|
|
||||||
}
|
|
||||||
c.Request.Header.Set("Authorization", "Bearer "+channel.Key)
|
|
||||||
c.Request.Header.Set("Content-Type", "application/json")
|
|
||||||
c.Set("channel", channel.Type)
|
|
||||||
c.Set("base_url", channel.GetBaseURL())
|
|
||||||
middleware.SetupContextForSelectedChannel(c, channel, "")
|
|
||||||
meta := util.GetRelayMeta(c)
|
|
||||||
apiType := constant.ChannelType2APIType(channel.Type)
|
|
||||||
adaptor := helper.GetAdaptor(apiType)
|
|
||||||
if adaptor == nil {
|
|
||||||
return fmt.Errorf("invalid api type: %d, adaptor is nil", apiType), nil
|
|
||||||
}
|
|
||||||
adaptor.Init(meta)
|
|
||||||
modelName := adaptor.GetModelList()[0]
|
|
||||||
if !strings.Contains(channel.Models, modelName) {
|
|
||||||
modelNames := strings.Split(channel.Models, ",")
|
|
||||||
if len(modelNames) > 0 {
|
|
||||||
modelName = modelNames[0]
|
|
||||||
}
|
|
||||||
}
|
|
||||||
request := buildTestRequest()
|
|
||||||
request.Model = modelName
|
|
||||||
meta.OriginModelName, meta.ActualModelName = modelName, modelName
|
|
||||||
convertedRequest, err := adaptor.ConvertRequest(c, constant.RelayModeChatCompletions, request)
|
|
||||||
if err != nil {
|
|
||||||
return err, nil
|
|
||||||
}
|
|
||||||
jsonData, err := json.Marshal(convertedRequest)
|
|
||||||
if err != nil {
|
|
||||||
return err, nil
|
|
||||||
}
|
|
||||||
requestBody := bytes.NewBuffer(jsonData)
|
|
||||||
c.Request.Body = io.NopCloser(requestBody)
|
|
||||||
resp, err := adaptor.DoRequest(c, meta, requestBody)
|
|
||||||
if err != nil {
|
|
||||||
return err, nil
|
|
||||||
}
|
|
||||||
if resp.StatusCode != http.StatusOK {
|
|
||||||
err := util.RelayErrorHandler(resp)
|
|
||||||
return fmt.Errorf("status code %d: %s", resp.StatusCode, err.Error.Message), &err.Error
|
|
||||||
}
|
|
||||||
usage, respErr := adaptor.DoResponse(c, resp, meta)
|
|
||||||
if respErr != nil {
|
|
||||||
return fmt.Errorf("%s", respErr.Error.Message), &respErr.Error
|
|
||||||
}
|
|
||||||
if usage == nil {
|
|
||||||
return errors.New("usage is nil"), nil
|
|
||||||
}
|
|
||||||
result := w.Result()
|
|
||||||
// print result.Body
|
|
||||||
respBody, err := io.ReadAll(result.Body)
|
|
||||||
if err != nil {
|
|
||||||
return err, nil
|
|
||||||
}
|
|
||||||
logger.SysLog(fmt.Sprintf("testing channel #%d, response: \n%s", channel.Id, string(respBody)))
|
|
||||||
return nil, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestChannel(c *gin.Context) {
|
func TestChannel(c *gin.Context) {
|
||||||
id, err := strconv.Atoi(c.Param("id"))
|
id, err := strconv.Atoi(c.Param("id"))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -125,8 +125,9 @@ func TestChannel(c *gin.Context) {
|
|||||||
})
|
})
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
testRequest := buildTestRequest()
|
||||||
tik := time.Now()
|
tik := time.Now()
|
||||||
err, _ = testChannel(channel)
|
err, _ = testChannel(channel, *testRequest)
|
||||||
tok := time.Now()
|
tok := time.Now()
|
||||||
milliseconds := tok.Sub(tik).Milliseconds()
|
milliseconds := tok.Sub(tik).Milliseconds()
|
||||||
go channel.UpdateResponseTime(milliseconds)
|
go channel.UpdateResponseTime(milliseconds)
|
||||||
@@ -150,7 +151,33 @@ func TestChannel(c *gin.Context) {
|
|||||||
var testAllChannelsLock sync.Mutex
|
var testAllChannelsLock sync.Mutex
|
||||||
var testAllChannelsRunning bool = false
|
var testAllChannelsRunning bool = false
|
||||||
|
|
||||||
func testChannels(notify bool, scope string) error {
|
func notifyRootUser(subject string, content string) {
|
||||||
|
if config.RootUserEmail == "" {
|
||||||
|
config.RootUserEmail = model.GetRootUserEmail()
|
||||||
|
}
|
||||||
|
err := common.SendEmail(subject, config.RootUserEmail, content)
|
||||||
|
if err != nil {
|
||||||
|
logger.SysError(fmt.Sprintf("failed to send email: %s", err.Error()))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// disable & notify
|
||||||
|
func disableChannel(channelId int, channelName string, reason string) {
|
||||||
|
model.UpdateChannelStatusById(channelId, common.ChannelStatusAutoDisabled)
|
||||||
|
subject := fmt.Sprintf("通道「%s」(#%d)已被禁用", channelName, channelId)
|
||||||
|
content := fmt.Sprintf("通道「%s」(#%d)已被禁用,原因:%s", channelName, channelId, reason)
|
||||||
|
notifyRootUser(subject, content)
|
||||||
|
}
|
||||||
|
|
||||||
|
// enable & notify
|
||||||
|
func enableChannel(channelId int, channelName string) {
|
||||||
|
model.UpdateChannelStatusById(channelId, common.ChannelStatusEnabled)
|
||||||
|
subject := fmt.Sprintf("通道「%s」(#%d)已被启用", channelName, channelId)
|
||||||
|
content := fmt.Sprintf("通道「%s」(#%d)已被启用", channelName, channelId)
|
||||||
|
notifyRootUser(subject, content)
|
||||||
|
}
|
||||||
|
|
||||||
|
func testAllChannels(notify bool) error {
|
||||||
if config.RootUserEmail == "" {
|
if config.RootUserEmail == "" {
|
||||||
config.RootUserEmail = model.GetRootUserEmail()
|
config.RootUserEmail = model.GetRootUserEmail()
|
||||||
}
|
}
|
||||||
@@ -161,10 +188,11 @@ func testChannels(notify bool, scope string) error {
|
|||||||
}
|
}
|
||||||
testAllChannelsRunning = true
|
testAllChannelsRunning = true
|
||||||
testAllChannelsLock.Unlock()
|
testAllChannelsLock.Unlock()
|
||||||
channels, err := model.GetAllChannels(0, 0, scope)
|
channels, err := model.GetAllChannels(0, 0, true)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
testRequest := buildTestRequest()
|
||||||
var disableThreshold = int64(config.ChannelDisableThreshold * 1000)
|
var disableThreshold = int64(config.ChannelDisableThreshold * 1000)
|
||||||
if disableThreshold == 0 {
|
if disableThreshold == 0 {
|
||||||
disableThreshold = 10000000 // a impossible value
|
disableThreshold = 10000000 // a impossible value
|
||||||
@@ -173,18 +201,18 @@ func testChannels(notify bool, scope string) error {
|
|||||||
for _, channel := range channels {
|
for _, channel := range channels {
|
||||||
isChannelEnabled := channel.Status == common.ChannelStatusEnabled
|
isChannelEnabled := channel.Status == common.ChannelStatusEnabled
|
||||||
tik := time.Now()
|
tik := time.Now()
|
||||||
err, openaiErr := testChannel(channel)
|
err, openaiErr := testChannel(channel, *testRequest)
|
||||||
tok := time.Now()
|
tok := time.Now()
|
||||||
milliseconds := tok.Sub(tik).Milliseconds()
|
milliseconds := tok.Sub(tik).Milliseconds()
|
||||||
if isChannelEnabled && milliseconds > disableThreshold {
|
if isChannelEnabled && milliseconds > disableThreshold {
|
||||||
err = errors.New(fmt.Sprintf("响应时间 %.2fs 超过阈值 %.2fs", float64(milliseconds)/1000.0, float64(disableThreshold)/1000.0))
|
err = errors.New(fmt.Sprintf("响应时间 %.2fs 超过阈值 %.2fs", float64(milliseconds)/1000.0, float64(disableThreshold)/1000.0))
|
||||||
monitor.DisableChannel(channel.Id, channel.Name, err.Error())
|
disableChannel(channel.Id, channel.Name, err.Error())
|
||||||
}
|
}
|
||||||
if isChannelEnabled && util.ShouldDisableChannel(openaiErr, -1) {
|
if isChannelEnabled && util.ShouldDisableChannel(openaiErr, -1) {
|
||||||
monitor.DisableChannel(channel.Id, channel.Name, err.Error())
|
disableChannel(channel.Id, channel.Name, err.Error())
|
||||||
}
|
}
|
||||||
if !isChannelEnabled && util.ShouldEnableChannel(err, openaiErr) {
|
if !isChannelEnabled && util.ShouldEnableChannel(err, openaiErr) {
|
||||||
monitor.EnableChannel(channel.Id, channel.Name)
|
enableChannel(channel.Id, channel.Name)
|
||||||
}
|
}
|
||||||
channel.UpdateResponseTime(milliseconds)
|
channel.UpdateResponseTime(milliseconds)
|
||||||
time.Sleep(config.RequestInterval)
|
time.Sleep(config.RequestInterval)
|
||||||
@@ -193,7 +221,7 @@ func testChannels(notify bool, scope string) error {
|
|||||||
testAllChannelsRunning = false
|
testAllChannelsRunning = false
|
||||||
testAllChannelsLock.Unlock()
|
testAllChannelsLock.Unlock()
|
||||||
if notify {
|
if notify {
|
||||||
err := message.Notify(message.ByAll, "通道测试完成", "", "通道测试完成,如果没有收到禁用通知,说明所有通道都正常")
|
err := common.SendEmail("通道测试完成", config.RootUserEmail, "通道测试完成,如果没有收到禁用通知,说明所有通道都正常")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logger.SysError(fmt.Sprintf("failed to send email: %s", err.Error()))
|
logger.SysError(fmt.Sprintf("failed to send email: %s", err.Error()))
|
||||||
}
|
}
|
||||||
@@ -202,12 +230,8 @@ func testChannels(notify bool, scope string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestChannels(c *gin.Context) {
|
func TestAllChannels(c *gin.Context) {
|
||||||
scope := c.Query("scope")
|
err := testAllChannels(true)
|
||||||
if scope == "" {
|
|
||||||
scope = "all"
|
|
||||||
}
|
|
||||||
err := testChannels(true, scope)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.JSON(http.StatusOK, gin.H{
|
c.JSON(http.StatusOK, gin.H{
|
||||||
"success": false,
|
"success": false,
|
||||||
@@ -226,7 +250,7 @@ func AutomaticallyTestChannels(frequency int) {
|
|||||||
for {
|
for {
|
||||||
time.Sleep(time.Duration(frequency) * time.Minute)
|
time.Sleep(time.Duration(frequency) * time.Minute)
|
||||||
logger.SysLog("testing all channels")
|
logger.SysLog("testing all channels")
|
||||||
_ = testChannels(false, "all")
|
_ = testAllChannels(false)
|
||||||
logger.SysLog("channel test finished")
|
logger.SysLog("channel test finished")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,10 +2,10 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/model"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
@@ -15,7 +15,7 @@ func GetAllChannels(c *gin.Context) {
|
|||||||
if p < 0 {
|
if p < 0 {
|
||||||
p = 0
|
p = 0
|
||||||
}
|
}
|
||||||
channels, err := model.GetAllChannels(p*config.ItemsPerPage, config.ItemsPerPage, "limited")
|
channels, err := model.GetAllChannels(p*config.ItemsPerPage, config.ItemsPerPage, false)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.JSON(http.StatusOK, gin.H{
|
c.JSON(http.StatusOK, gin.H{
|
||||||
"success": false,
|
"success": false,
|
||||||
|
|||||||
@@ -7,12 +7,12 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-contrib/sessions"
|
"github.com/gin-contrib/sessions"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/model"
|
||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -2,13 +2,13 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
)
|
)
|
||||||
|
|
||||||
func GetGroups(c *gin.Context) {
|
func GetGroups(c *gin.Context) {
|
||||||
groupNames := make([]string, 0)
|
groupNames := make([]string, 0)
|
||||||
for groupName := range common.GroupRatio {
|
for groupName, _ := range common.GroupRatio {
|
||||||
groupNames = append(groupNames, groupName)
|
groupNames = append(groupNames, groupName)
|
||||||
}
|
}
|
||||||
c.JSON(http.StatusOK, gin.H{
|
c.JSON(http.StatusOK, gin.H{
|
||||||
|
|||||||
@@ -2,9 +2,9 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/model"
|
||||||
"strconv"
|
"strconv"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -3,11 +3,10 @@ package controller
|
|||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/message"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/model"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
@@ -111,7 +110,7 @@ func SendEmailVerification(c *gin.Context) {
|
|||||||
content := fmt.Sprintf("<p>您好,你正在进行%s邮箱验证。</p>"+
|
content := fmt.Sprintf("<p>您好,你正在进行%s邮箱验证。</p>"+
|
||||||
"<p>您的验证码为: <strong>%s</strong></p>"+
|
"<p>您的验证码为: <strong>%s</strong></p>"+
|
||||||
"<p>验证码 %d 分钟内有效,如果不是本人操作,请忽略。</p>", config.SystemName, code, common.VerificationValidMinutes)
|
"<p>验证码 %d 分钟内有效,如果不是本人操作,请忽略。</p>", config.SystemName, code, common.VerificationValidMinutes)
|
||||||
err := message.SendEmail(subject, email, content)
|
err := common.SendEmail(subject, email, content)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.JSON(http.StatusOK, gin.H{
|
c.JSON(http.StatusOK, gin.H{
|
||||||
"success": false,
|
"success": false,
|
||||||
@@ -150,7 +149,7 @@ func SendPasswordResetEmail(c *gin.Context) {
|
|||||||
"<p>点击 <a href='%s'>此处</a> 进行密码重置。</p>"+
|
"<p>点击 <a href='%s'>此处</a> 进行密码重置。</p>"+
|
||||||
"<p>如果链接无法点击,请尝试点击下面的链接或将其复制到浏览器中打开:<br> %s </p>"+
|
"<p>如果链接无法点击,请尝试点击下面的链接或将其复制到浏览器中打开:<br> %s </p>"+
|
||||||
"<p>重置链接 %d 分钟内有效,如果不是本人操作,请忽略。</p>", config.SystemName, link, link, common.VerificationValidMinutes)
|
"<p>重置链接 %d 分钟内有效,如果不是本人操作,请忽略。</p>", config.SystemName, link, link, common.VerificationValidMinutes)
|
||||||
err := message.SendEmail(subject, email, content)
|
err := common.SendEmail(subject, email, content)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.JSON(http.StatusOK, gin.H{
|
c.JSON(http.StatusOK, gin.H{
|
||||||
"success": false,
|
"success": false,
|
||||||
|
|||||||
@@ -3,13 +3,7 @@ package controller
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
"one-api/relay/channel/openai"
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/helper"
|
|
||||||
relaymodel "github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"net/http"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// https://platform.openai.com/docs/api-reference/models/list
|
// https://platform.openai.com/docs/api-reference/models/list
|
||||||
@@ -41,7 +35,6 @@ type OpenAIModels struct {
|
|||||||
|
|
||||||
var openAIModels []OpenAIModels
|
var openAIModels []OpenAIModels
|
||||||
var openAIModelsMap map[string]OpenAIModels
|
var openAIModelsMap map[string]OpenAIModels
|
||||||
var channelId2Models map[int][]string
|
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
var permission []OpenAIModelPermission
|
var permission []OpenAIModelPermission
|
||||||
@@ -60,63 +53,552 @@ func init() {
|
|||||||
IsBlocking: false,
|
IsBlocking: false,
|
||||||
})
|
})
|
||||||
// https://platform.openai.com/docs/models/model-endpoint-compatibility
|
// https://platform.openai.com/docs/models/model-endpoint-compatibility
|
||||||
for i := 0; i < constant.APITypeDummy; i++ {
|
openAIModels = []OpenAIModels{
|
||||||
if i == constant.APITypeAIProxyLibrary {
|
{
|
||||||
continue
|
Id: "dall-e-2",
|
||||||
}
|
Object: "model",
|
||||||
adaptor := helper.GetAdaptor(i)
|
Created: 1677649963,
|
||||||
channelName := adaptor.GetChannelName()
|
OwnedBy: "openai",
|
||||||
modelNames := adaptor.GetModelList()
|
Permission: permission,
|
||||||
for _, modelName := range modelNames {
|
Root: "dall-e-2",
|
||||||
openAIModels = append(openAIModels, OpenAIModels{
|
Parent: nil,
|
||||||
Id: modelName,
|
},
|
||||||
Object: "model",
|
{
|
||||||
Created: 1626777600,
|
Id: "dall-e-3",
|
||||||
OwnedBy: channelName,
|
Object: "model",
|
||||||
Permission: permission,
|
Created: 1677649963,
|
||||||
Root: modelName,
|
OwnedBy: "openai",
|
||||||
Parent: nil,
|
Permission: permission,
|
||||||
})
|
Root: "dall-e-3",
|
||||||
}
|
Parent: nil,
|
||||||
}
|
},
|
||||||
for _, channelType := range openai.CompatibleChannels {
|
{
|
||||||
if channelType == common.ChannelTypeAzure {
|
Id: "whisper-1",
|
||||||
continue
|
Object: "model",
|
||||||
}
|
Created: 1677649963,
|
||||||
channelName, channelModelList := openai.GetCompatibleChannelMeta(channelType)
|
OwnedBy: "openai",
|
||||||
for _, modelName := range channelModelList {
|
Permission: permission,
|
||||||
openAIModels = append(openAIModels, OpenAIModels{
|
Root: "whisper-1",
|
||||||
Id: modelName,
|
Parent: nil,
|
||||||
Object: "model",
|
},
|
||||||
Created: 1626777600,
|
{
|
||||||
OwnedBy: channelName,
|
Id: "tts-1",
|
||||||
Permission: permission,
|
Object: "model",
|
||||||
Root: modelName,
|
Created: 1677649963,
|
||||||
Parent: nil,
|
OwnedBy: "openai",
|
||||||
})
|
Permission: permission,
|
||||||
}
|
Root: "tts-1",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "tts-1-1106",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "tts-1-1106",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "tts-1-hd",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "tts-1-hd",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "tts-1-hd-1106",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "tts-1-hd-1106",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-3.5-turbo",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-3.5-turbo",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-3.5-turbo-0301",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-3.5-turbo-0301",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-3.5-turbo-0613",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-3.5-turbo-0613",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-3.5-turbo-16k",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-3.5-turbo-16k",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-3.5-turbo-16k-0613",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-3.5-turbo-16k-0613",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-3.5-turbo-1106",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1699593571,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-3.5-turbo-1106",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-3.5-turbo-instruct",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-3.5-turbo-instruct",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-4",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-4",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-4-0314",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-4-0314",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-4-0613",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-4-0613",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-4-32k",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-4-32k",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-4-32k-0314",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-4-32k-0314",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-4-32k-0613",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-4-32k-0613",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-4-1106-preview",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1699593571,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-4-1106-preview",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gpt-4-vision-preview",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1699593571,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gpt-4-vision-preview",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-embedding-ada-002",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-embedding-ada-002",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-davinci-003",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-davinci-003",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-davinci-002",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-davinci-002",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-curie-001",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-curie-001",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-babbage-001",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-babbage-001",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-ada-001",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-ada-001",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-moderation-latest",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-moderation-latest",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-moderation-stable",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-moderation-stable",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-davinci-edit-001",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-davinci-edit-001",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "code-davinci-edit-001",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "code-davinci-edit-001",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "davinci-002",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "davinci-002",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "babbage-002",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "openai",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "babbage-002",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "claude-instant-1",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "anthropic",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "claude-instant-1",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "claude-2",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "anthropic",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "claude-2",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "claude-2.1",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "anthropic",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "claude-2.1",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "claude-2.0",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "anthropic",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "claude-2.0",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "ERNIE-Bot",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "baidu",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "ERNIE-Bot",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "ERNIE-Bot-turbo",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "baidu",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "ERNIE-Bot-turbo",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "ERNIE-Bot-4",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "baidu",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "ERNIE-Bot-4",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "Embedding-V1",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "baidu",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "Embedding-V1",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "PaLM-2",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "google palm",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "PaLM-2",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gemini-pro",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "google gemini",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gemini-pro",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "gemini-pro-vision",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "google gemini",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "gemini-pro-vision",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "chatglm_turbo",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "zhipu",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "chatglm_turbo",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "chatglm_pro",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "zhipu",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "chatglm_pro",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "chatglm_std",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "zhipu",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "chatglm_std",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "chatglm_lite",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "zhipu",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "chatglm_lite",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "qwen-turbo",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "ali",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "qwen-turbo",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "qwen-plus",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "ali",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "qwen-plus",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "qwen-max",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "ali",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "qwen-max",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "qwen-max-longcontext",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "ali",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "qwen-max-longcontext",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "text-embedding-v1",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "ali",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "text-embedding-v1",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "SparkDesk",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "xunfei",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "SparkDesk",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "360GPT_S2_V9",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "360",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "360GPT_S2_V9",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "embedding-bert-512-v1",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "360",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "embedding-bert-512-v1",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "embedding_s1_v1",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "360",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "embedding_s1_v1",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "semantic_similarity_s1_v1",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "360",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "semantic_similarity_s1_v1",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Id: "hunyuan",
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: "tencent",
|
||||||
|
Permission: permission,
|
||||||
|
Root: "hunyuan",
|
||||||
|
Parent: nil,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
openAIModelsMap = make(map[string]OpenAIModels)
|
openAIModelsMap = make(map[string]OpenAIModels)
|
||||||
for _, model := range openAIModels {
|
for _, model := range openAIModels {
|
||||||
openAIModelsMap[model.Id] = model
|
openAIModelsMap[model.Id] = model
|
||||||
}
|
}
|
||||||
channelId2Models = make(map[int][]string)
|
|
||||||
for i := 1; i < common.ChannelTypeDummy; i++ {
|
|
||||||
adaptor := helper.GetAdaptor(constant.ChannelType2APIType(i))
|
|
||||||
meta := &util.RelayMeta{
|
|
||||||
ChannelType: i,
|
|
||||||
}
|
|
||||||
adaptor.Init(meta)
|
|
||||||
channelId2Models[i] = adaptor.GetModelList()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func DashboardListModels(c *gin.Context) {
|
|
||||||
c.JSON(http.StatusOK, gin.H{
|
|
||||||
"success": true,
|
|
||||||
"message": "",
|
|
||||||
"data": channelId2Models,
|
|
||||||
})
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func ListModels(c *gin.Context) {
|
func ListModels(c *gin.Context) {
|
||||||
@@ -131,7 +613,7 @@ func RetrieveModel(c *gin.Context) {
|
|||||||
if model, ok := openAIModelsMap[modelId]; ok {
|
if model, ok := openAIModelsMap[modelId]; ok {
|
||||||
c.JSON(200, model)
|
c.JSON(200, model)
|
||||||
} else {
|
} else {
|
||||||
Error := relaymodel.Error{
|
Error := openai.Error{
|
||||||
Message: fmt.Sprintf("The model '%s' does not exist", modelId),
|
Message: fmt.Sprintf("The model '%s' does not exist", modelId),
|
||||||
Type: "invalid_request_error",
|
Type: "invalid_request_error",
|
||||||
Param: "model",
|
Param: "model",
|
||||||
|
|||||||
@@ -2,10 +2,10 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/model"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
|
|||||||
@@ -2,10 +2,10 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/model"
|
||||||
"strconv"
|
"strconv"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -1,29 +1,24 @@
|
|||||||
package controller
|
package controller
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
|
||||||
"context"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/middleware"
|
|
||||||
dbmodel "github.com/songquanpeng/one-api/model"
|
|
||||||
"github.com/songquanpeng/one-api/monitor"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/controller"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
|
"one-api/relay/constant"
|
||||||
|
"one-api/relay/controller"
|
||||||
|
"one-api/relay/util"
|
||||||
|
"strconv"
|
||||||
)
|
)
|
||||||
|
|
||||||
// https://platform.openai.com/docs/api-reference/chat
|
// https://platform.openai.com/docs/api-reference/chat
|
||||||
|
|
||||||
func relay(c *gin.Context, relayMode int) *model.ErrorWithStatusCode {
|
func Relay(c *gin.Context) {
|
||||||
var err *model.ErrorWithStatusCode
|
relayMode := constant.Path2RelayMode(c.Request.URL.Path)
|
||||||
|
var err *openai.ErrorWithStatusCode
|
||||||
switch relayMode {
|
switch relayMode {
|
||||||
case constant.RelayModeImagesGenerations:
|
case constant.RelayModeImagesGenerations:
|
||||||
err = controller.RelayImageHelper(c, relayMode)
|
err = controller.RelayImageHelper(c, relayMode)
|
||||||
@@ -34,99 +29,39 @@ func relay(c *gin.Context, relayMode int) *model.ErrorWithStatusCode {
|
|||||||
case constant.RelayModeAudioTranscription:
|
case constant.RelayModeAudioTranscription:
|
||||||
err = controller.RelayAudioHelper(c, relayMode)
|
err = controller.RelayAudioHelper(c, relayMode)
|
||||||
default:
|
default:
|
||||||
err = controller.RelayTextHelper(c)
|
err = controller.RelayTextHelper(c, relayMode)
|
||||||
}
|
}
|
||||||
return err
|
if err != nil {
|
||||||
}
|
requestId := c.GetString(logger.RequestIdKey)
|
||||||
|
retryTimesStr := c.Query("retry")
|
||||||
func Relay(c *gin.Context) {
|
retryTimes, _ := strconv.Atoi(retryTimesStr)
|
||||||
ctx := c.Request.Context()
|
if retryTimesStr == "" {
|
||||||
relayMode := constant.Path2RelayMode(c.Request.URL.Path)
|
retryTimes = config.RetryTimes
|
||||||
if config.DebugEnabled {
|
|
||||||
requestBody, _ := common.GetRequestBody(c)
|
|
||||||
logger.Debugf(ctx, "request body: %s", string(requestBody))
|
|
||||||
}
|
|
||||||
channelId := c.GetInt("channel_id")
|
|
||||||
bizErr := relay(c, relayMode)
|
|
||||||
if bizErr == nil {
|
|
||||||
monitor.Emit(channelId, true)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
lastFailedChannelId := channelId
|
|
||||||
channelName := c.GetString("channel_name")
|
|
||||||
group := c.GetString("group")
|
|
||||||
originalModel := c.GetString("original_model")
|
|
||||||
go processChannelRelayError(ctx, channelId, channelName, bizErr)
|
|
||||||
requestId := c.GetString(logger.RequestIdKey)
|
|
||||||
retryTimes := config.RetryTimes
|
|
||||||
if !shouldRetry(c, bizErr.StatusCode) {
|
|
||||||
logger.Errorf(ctx, "relay error happen, status code is %d, won't retry in this case", bizErr.StatusCode)
|
|
||||||
retryTimes = 0
|
|
||||||
}
|
|
||||||
for i := retryTimes; i > 0; i-- {
|
|
||||||
channel, err := dbmodel.CacheGetRandomSatisfiedChannel(group, originalModel, i != retryTimes)
|
|
||||||
if err != nil {
|
|
||||||
logger.Errorf(ctx, "CacheGetRandomSatisfiedChannel failed: %w", err)
|
|
||||||
break
|
|
||||||
}
|
}
|
||||||
logger.Infof(ctx, "using channel #%d to retry (remain times %d)", channel.Id, i)
|
if retryTimes > 0 {
|
||||||
if channel.Id == lastFailedChannelId {
|
c.Redirect(http.StatusTemporaryRedirect, fmt.Sprintf("%s?retry=%d", c.Request.URL.Path, retryTimes-1))
|
||||||
continue
|
} else {
|
||||||
}
|
if err.StatusCode == http.StatusTooManyRequests {
|
||||||
middleware.SetupContextForSelectedChannel(c, channel, originalModel)
|
err.Error.Message = "当前分组上游负载已饱和,请稍后再试"
|
||||||
requestBody, err := common.GetRequestBody(c)
|
}
|
||||||
c.Request.Body = io.NopCloser(bytes.NewBuffer(requestBody))
|
err.Error.Message = helper.MessageWithRequestId(err.Error.Message, requestId)
|
||||||
bizErr = relay(c, relayMode)
|
c.JSON(err.StatusCode, gin.H{
|
||||||
if bizErr == nil {
|
"error": err.Error,
|
||||||
return
|
})
|
||||||
}
|
}
|
||||||
channelId := c.GetInt("channel_id")
|
channelId := c.GetInt("channel_id")
|
||||||
lastFailedChannelId = channelId
|
logger.Error(c.Request.Context(), fmt.Sprintf("relay error (channel #%d): %s", channelId, err.Message))
|
||||||
channelName := c.GetString("channel_name")
|
// https://platform.openai.com/docs/guides/error-codes/api-errors
|
||||||
go processChannelRelayError(ctx, channelId, channelName, bizErr)
|
if util.ShouldDisableChannel(&err.Error, err.StatusCode) {
|
||||||
}
|
channelId := c.GetInt("channel_id")
|
||||||
if bizErr != nil {
|
channelName := c.GetString("channel_name")
|
||||||
if bizErr.StatusCode == http.StatusTooManyRequests {
|
disableChannel(channelId, channelName, err.Message)
|
||||||
bizErr.Error.Message = "当前分组上游负载已饱和,请稍后再试"
|
|
||||||
}
|
}
|
||||||
bizErr.Error.Message = helper.MessageWithRequestId(bizErr.Error.Message, requestId)
|
|
||||||
c.JSON(bizErr.StatusCode, gin.H{
|
|
||||||
"error": bizErr.Error,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func shouldRetry(c *gin.Context, statusCode int) bool {
|
|
||||||
if _, ok := c.Get("specific_channel_id"); ok {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
if statusCode == http.StatusTooManyRequests {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
if statusCode/100 == 5 {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
if statusCode == http.StatusBadRequest {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
if statusCode/100 == 2 {
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
|
|
||||||
func processChannelRelayError(ctx context.Context, channelId int, channelName string, err *model.ErrorWithStatusCode) {
|
|
||||||
logger.Errorf(ctx, "relay error (channel #%d): %s", channelId, err.Message)
|
|
||||||
// https://platform.openai.com/docs/guides/error-codes/api-errors
|
|
||||||
if util.ShouldDisableChannel(&err.Error, err.StatusCode) {
|
|
||||||
monitor.DisableChannel(channelId, channelName, err.Message)
|
|
||||||
} else {
|
|
||||||
monitor.Emit(channelId, false)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func RelayNotImplemented(c *gin.Context) {
|
func RelayNotImplemented(c *gin.Context) {
|
||||||
err := model.Error{
|
err := openai.Error{
|
||||||
Message: "API not implemented",
|
Message: "API not implemented",
|
||||||
Type: "one_api_error",
|
Type: "one_api_error",
|
||||||
Param: "",
|
Param: "",
|
||||||
@@ -138,7 +73,7 @@ func RelayNotImplemented(c *gin.Context) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func RelayNotFound(c *gin.Context) {
|
func RelayNotFound(c *gin.Context) {
|
||||||
err := model.Error{
|
err := openai.Error{
|
||||||
Message: fmt.Sprintf("Invalid URL (%s %s)", c.Request.Method, c.Request.URL.Path),
|
Message: fmt.Sprintf("Invalid URL (%s %s)", c.Request.Method, c.Request.URL.Path),
|
||||||
Type: "invalid_request_error",
|
Type: "invalid_request_error",
|
||||||
Param: "",
|
Param: "",
|
||||||
|
|||||||
@@ -2,11 +2,11 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/model"
|
||||||
"strconv"
|
"strconv"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -3,11 +3,11 @@ package controller
|
|||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/model"
|
||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
|||||||
@@ -5,10 +5,10 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/model"
|
||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|||||||
2
go.mod
2
go.mod
@@ -1,4 +1,4 @@
|
|||||||
module github.com/songquanpeng/one-api
|
module one-api
|
||||||
|
|
||||||
// +heroku goVersion go1.18
|
// +heroku goVersion go1.18
|
||||||
go 1.18
|
go 1.18
|
||||||
|
|||||||
@@ -456,7 +456,6 @@
|
|||||||
"已绑定的邮箱账户": "Email Account Bound",
|
"已绑定的邮箱账户": "Email Account Bound",
|
||||||
"用户信息更新成功!": "User information updated successfully!",
|
"用户信息更新成功!": "User information updated successfully!",
|
||||||
"模型倍率 %.2f,分组倍率 %.2f": "model rate %.2f, group rate %.2f",
|
"模型倍率 %.2f,分组倍率 %.2f": "model rate %.2f, group rate %.2f",
|
||||||
"模型倍率 %.2f,分组倍率 %.2f,补全倍率 %.2f": "model rate %.2f, group rate %.2f, completion rate %.2f",
|
|
||||||
"使用明细(总消耗额度:{renderQuota(stat.quota)})": "Usage Details (Total Consumption Quota: {renderQuota(stat.quota)})",
|
"使用明细(总消耗额度:{renderQuota(stat.quota)})": "Usage Details (Total Consumption Quota: {renderQuota(stat.quota)})",
|
||||||
"用户名称": "User Name",
|
"用户名称": "User Name",
|
||||||
"令牌名称": "Token Name",
|
"令牌名称": "Token Name",
|
||||||
|
|||||||
21
main.go
21
main.go
@@ -6,15 +6,14 @@ import (
|
|||||||
"github.com/gin-contrib/sessions"
|
"github.com/gin-contrib/sessions"
|
||||||
"github.com/gin-contrib/sessions/cookie"
|
"github.com/gin-contrib/sessions/cookie"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
"one-api/common"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
"one-api/common/config"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
"github.com/songquanpeng/one-api/common/message"
|
"one-api/controller"
|
||||||
"github.com/songquanpeng/one-api/controller"
|
"one-api/middleware"
|
||||||
"github.com/songquanpeng/one-api/middleware"
|
"one-api/model"
|
||||||
"github.com/songquanpeng/one-api/model"
|
"one-api/relay/channel/openai"
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
"one-api/router"
|
||||||
"github.com/songquanpeng/one-api/router"
|
|
||||||
"os"
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
)
|
)
|
||||||
@@ -84,11 +83,7 @@ func main() {
|
|||||||
logger.SysLog("batch update enabled with interval " + strconv.Itoa(config.BatchUpdateInterval) + "s")
|
logger.SysLog("batch update enabled with interval " + strconv.Itoa(config.BatchUpdateInterval) + "s")
|
||||||
model.InitBatchUpdater()
|
model.InitBatchUpdater()
|
||||||
}
|
}
|
||||||
if config.EnableMetric {
|
|
||||||
logger.SysLog("metric enabled, will disable channel if too much request failed")
|
|
||||||
}
|
|
||||||
openai.InitTokenEncoders()
|
openai.InitTokenEncoders()
|
||||||
_ = message.SendMessage("One API", "", fmt.Sprintf("One API %s started", common.Version))
|
|
||||||
|
|
||||||
// Initialize HTTP server
|
// Initialize HTTP server
|
||||||
server := gin.New()
|
server := gin.New()
|
||||||
|
|||||||
@@ -3,10 +3,9 @@ package middleware
|
|||||||
import (
|
import (
|
||||||
"github.com/gin-contrib/sessions"
|
"github.com/gin-contrib/sessions"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/blacklist"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/model"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -43,14 +42,11 @@ func authHelper(c *gin.Context, minRole int) {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if status.(int) == common.UserStatusDisabled || blacklist.IsUserBanned(id.(int)) {
|
if status.(int) == common.UserStatusDisabled {
|
||||||
c.JSON(http.StatusOK, gin.H{
|
c.JSON(http.StatusOK, gin.H{
|
||||||
"success": false,
|
"success": false,
|
||||||
"message": "用户已被封禁",
|
"message": "用户已被封禁",
|
||||||
})
|
})
|
||||||
session := sessions.Default(c)
|
|
||||||
session.Clear()
|
|
||||||
_ = session.Save()
|
|
||||||
c.Abort()
|
c.Abort()
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
@@ -103,7 +99,7 @@ func TokenAuth() func(c *gin.Context) {
|
|||||||
abortWithMessage(c, http.StatusInternalServerError, err.Error())
|
abortWithMessage(c, http.StatusInternalServerError, err.Error())
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
if !userEnabled || blacklist.IsUserBanned(token.UserId) {
|
if !userEnabled {
|
||||||
abortWithMessage(c, http.StatusForbidden, "用户已被封禁")
|
abortWithMessage(c, http.StatusForbidden, "用户已被封禁")
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
@@ -112,7 +108,7 @@ func TokenAuth() func(c *gin.Context) {
|
|||||||
c.Set("token_name", token.Name)
|
c.Set("token_name", token.Name)
|
||||||
if len(parts) > 1 {
|
if len(parts) > 1 {
|
||||||
if model.IsAdmin(token.UserId) {
|
if model.IsAdmin(token.UserId) {
|
||||||
c.Set("specific_channel_id", parts[1])
|
c.Set("channelId", parts[1])
|
||||||
} else {
|
} else {
|
||||||
abortWithMessage(c, http.StatusForbidden, "普通用户不支持指定渠道")
|
abortWithMessage(c, http.StatusForbidden, "普通用户不支持指定渠道")
|
||||||
return
|
return
|
||||||
|
|||||||
@@ -2,10 +2,10 @@ package middleware
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/model"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
@@ -21,9 +21,8 @@ func Distribute() func(c *gin.Context) {
|
|||||||
userId := c.GetInt("id")
|
userId := c.GetInt("id")
|
||||||
userGroup, _ := model.CacheGetUserGroup(userId)
|
userGroup, _ := model.CacheGetUserGroup(userId)
|
||||||
c.Set("group", userGroup)
|
c.Set("group", userGroup)
|
||||||
var requestModel string
|
|
||||||
var channel *model.Channel
|
var channel *model.Channel
|
||||||
channelId, ok := c.Get("specific_channel_id")
|
channelId, ok := c.Get("channelId")
|
||||||
if ok {
|
if ok {
|
||||||
id, err := strconv.Atoi(channelId.(string))
|
id, err := strconv.Atoi(channelId.(string))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -67,8 +66,7 @@ func Distribute() func(c *gin.Context) {
|
|||||||
modelRequest.Model = "whisper-1"
|
modelRequest.Model = "whisper-1"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
requestModel = modelRequest.Model
|
channel, err = model.CacheGetRandomSatisfiedChannel(userGroup, modelRequest.Model)
|
||||||
channel, err = model.CacheGetRandomSatisfiedChannel(userGroup, modelRequest.Model, false)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
message := fmt.Sprintf("当前分组 %s 下对于模型 %s 无可用渠道", userGroup, modelRequest.Model)
|
message := fmt.Sprintf("当前分组 %s 下对于模型 %s 无可用渠道", userGroup, modelRequest.Model)
|
||||||
if channel != nil {
|
if channel != nil {
|
||||||
@@ -79,34 +77,24 @@ func Distribute() func(c *gin.Context) {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
SetupContextForSelectedChannel(c, channel, requestModel)
|
c.Set("channel", channel.Type)
|
||||||
|
c.Set("channel_id", channel.Id)
|
||||||
|
c.Set("channel_name", channel.Name)
|
||||||
|
c.Set("model_mapping", channel.GetModelMapping())
|
||||||
|
c.Request.Header.Set("Authorization", fmt.Sprintf("Bearer %s", channel.Key))
|
||||||
|
c.Set("base_url", channel.GetBaseURL())
|
||||||
|
switch channel.Type {
|
||||||
|
case common.ChannelTypeAzure:
|
||||||
|
c.Set("api_version", channel.Other)
|
||||||
|
case common.ChannelTypeXunfei:
|
||||||
|
c.Set("api_version", channel.Other)
|
||||||
|
case common.ChannelTypeGemini:
|
||||||
|
c.Set("api_version", channel.Other)
|
||||||
|
case common.ChannelTypeAIProxyLibrary:
|
||||||
|
c.Set("library_id", channel.Other)
|
||||||
|
case common.ChannelTypeAli:
|
||||||
|
c.Set("plugin", channel.Other)
|
||||||
|
}
|
||||||
c.Next()
|
c.Next()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func SetupContextForSelectedChannel(c *gin.Context, channel *model.Channel, modelName string) {
|
|
||||||
c.Set("channel", channel.Type)
|
|
||||||
c.Set("channel_id", channel.Id)
|
|
||||||
c.Set("channel_name", channel.Name)
|
|
||||||
c.Set("model_mapping", channel.GetModelMapping())
|
|
||||||
c.Set("original_model", modelName) // for retry
|
|
||||||
c.Request.Header.Set("Authorization", fmt.Sprintf("Bearer %s", channel.Key))
|
|
||||||
c.Set("base_url", channel.GetBaseURL())
|
|
||||||
// this is for backward compatibility
|
|
||||||
switch channel.Type {
|
|
||||||
case common.ChannelTypeAzure:
|
|
||||||
c.Set(common.ConfigKeyAPIVersion, channel.Other)
|
|
||||||
case common.ChannelTypeXunfei:
|
|
||||||
c.Set(common.ConfigKeyAPIVersion, channel.Other)
|
|
||||||
case common.ChannelTypeGemini:
|
|
||||||
c.Set(common.ConfigKeyAPIVersion, channel.Other)
|
|
||||||
case common.ChannelTypeAIProxyLibrary:
|
|
||||||
c.Set(common.ConfigKeyLibraryID, channel.Other)
|
|
||||||
case common.ChannelTypeAli:
|
|
||||||
c.Set(common.ConfigKeyPlugin, channel.Other)
|
|
||||||
}
|
|
||||||
cfg, _ := channel.LoadConfig()
|
|
||||||
for k, v := range cfg {
|
|
||||||
c.Set(common.ConfigKeyPrefix+k, v)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ package middleware
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
func SetUpLogger(server *gin.Engine) {
|
func SetUpLogger(server *gin.Engine) {
|
||||||
|
|||||||
@@ -4,9 +4,9 @@ import (
|
|||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -3,8 +3,8 @@ package middleware
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common/logger"
|
||||||
"runtime/debug"
|
"runtime/debug"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -3,13 +3,13 @@ package middleware
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
"one-api/common/helper"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
func RequestId() func(c *gin.Context) {
|
func RequestId() func(c *gin.Context) {
|
||||||
return func(c *gin.Context) {
|
return func(c *gin.Context) {
|
||||||
id := helper.GetTimeString() + helper.GetRandomNumberString(8)
|
id := helper.GetTimeString() + helper.GetRandomString(8)
|
||||||
c.Set(logger.RequestIdKey, id)
|
c.Set(logger.RequestIdKey, id)
|
||||||
ctx := context.WithValue(c.Request.Context(), logger.RequestIdKey, id)
|
ctx := context.WithValue(c.Request.Context(), logger.RequestIdKey, id)
|
||||||
c.Request = c.Request.WithContext(ctx)
|
c.Request = c.Request.WithContext(ctx)
|
||||||
|
|||||||
@@ -4,10 +4,10 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/gin-contrib/sessions"
|
"github.com/gin-contrib/sessions"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
"net/url"
|
"net/url"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
type turnstileCheckResponse struct {
|
type turnstileCheckResponse struct {
|
||||||
|
|||||||
@@ -2,8 +2,8 @@ package middleware
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
"one-api/common/helper"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
func abortWithMessage(c *gin.Context, statusCode int, message string) {
|
func abortWithMessage(c *gin.Context, statusCode int, message string) {
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
package model
|
package model
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/songquanpeng/one-api/common"
|
"one-api/common"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -4,10 +4,10 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"math/rand"
|
"math/rand"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/logger"
|
||||||
"sort"
|
"sort"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
@@ -94,7 +94,7 @@ func CacheUpdateUserQuota(id int) error {
|
|||||||
if !common.RedisEnabled {
|
if !common.RedisEnabled {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
quota, err := CacheGetUserQuota(id)
|
quota, err := GetUserQuota(id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -191,7 +191,7 @@ func SyncChannelCache(frequency int) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func CacheGetRandomSatisfiedChannel(group string, model string, ignoreFirstPriority bool) (*Channel, error) {
|
func CacheGetRandomSatisfiedChannel(group string, model string) (*Channel, error) {
|
||||||
if !config.MemoryCacheEnabled {
|
if !config.MemoryCacheEnabled {
|
||||||
return GetRandomSatisfiedChannel(group, model)
|
return GetRandomSatisfiedChannel(group, model)
|
||||||
}
|
}
|
||||||
@@ -213,10 +213,5 @@ func CacheGetRandomSatisfiedChannel(group string, model string, ignoreFirstPrior
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
idx := rand.Intn(endIdx)
|
idx := rand.Intn(endIdx)
|
||||||
if ignoreFirstPriority {
|
|
||||||
if endIdx < len(channels) { // which means there are more than one priority
|
|
||||||
idx = common.RandRange(endIdx, len(channels))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return channels[idx], nil
|
return channels[idx], nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -3,11 +3,11 @@ package model
|
|||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"gorm.io/gorm"
|
"gorm.io/gorm"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Channel struct {
|
type Channel struct {
|
||||||
@@ -21,7 +21,7 @@ type Channel struct {
|
|||||||
TestTime int64 `json:"test_time" gorm:"bigint"`
|
TestTime int64 `json:"test_time" gorm:"bigint"`
|
||||||
ResponseTime int `json:"response_time"` // in milliseconds
|
ResponseTime int `json:"response_time"` // in milliseconds
|
||||||
BaseURL *string `json:"base_url" gorm:"column:base_url;default:''"`
|
BaseURL *string `json:"base_url" gorm:"column:base_url;default:''"`
|
||||||
Other string `json:"other"` // DEPRECATED: please save config to field Config
|
Other string `json:"other"`
|
||||||
Balance float64 `json:"balance"` // in USD
|
Balance float64 `json:"balance"` // in USD
|
||||||
BalanceUpdatedTime int64 `json:"balance_updated_time" gorm:"bigint"`
|
BalanceUpdatedTime int64 `json:"balance_updated_time" gorm:"bigint"`
|
||||||
Models string `json:"models"`
|
Models string `json:"models"`
|
||||||
@@ -29,18 +29,14 @@ type Channel struct {
|
|||||||
UsedQuota int64 `json:"used_quota" gorm:"bigint;default:0"`
|
UsedQuota int64 `json:"used_quota" gorm:"bigint;default:0"`
|
||||||
ModelMapping *string `json:"model_mapping" gorm:"type:varchar(1024);default:''"`
|
ModelMapping *string `json:"model_mapping" gorm:"type:varchar(1024);default:''"`
|
||||||
Priority *int64 `json:"priority" gorm:"bigint;default:0"`
|
Priority *int64 `json:"priority" gorm:"bigint;default:0"`
|
||||||
Config string `json:"config"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func GetAllChannels(startIdx int, num int, scope string) ([]*Channel, error) {
|
func GetAllChannels(startIdx int, num int, selectAll bool) ([]*Channel, error) {
|
||||||
var channels []*Channel
|
var channels []*Channel
|
||||||
var err error
|
var err error
|
||||||
switch scope {
|
if selectAll {
|
||||||
case "all":
|
|
||||||
err = DB.Order("id desc").Find(&channels).Error
|
err = DB.Order("id desc").Find(&channels).Error
|
||||||
case "disabled":
|
} else {
|
||||||
err = DB.Order("id desc").Where("status = ? or status = ?", common.ChannelStatusAutoDisabled, common.ChannelStatusManuallyDisabled).Find(&channels).Error
|
|
||||||
default:
|
|
||||||
err = DB.Order("id desc").Limit(num).Offset(startIdx).Omit("key").Find(&channels).Error
|
err = DB.Order("id desc").Limit(num).Offset(startIdx).Omit("key").Find(&channels).Error
|
||||||
}
|
}
|
||||||
return channels, err
|
return channels, err
|
||||||
@@ -159,18 +155,6 @@ func (channel *Channel) Delete() error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (channel *Channel) LoadConfig() (map[string]string, error) {
|
|
||||||
if channel.Config == "" {
|
|
||||||
return nil, nil
|
|
||||||
}
|
|
||||||
cfg := make(map[string]string)
|
|
||||||
err := json.Unmarshal([]byte(channel.Config), &cfg)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
return cfg, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func UpdateChannelStatusById(id int, status int) {
|
func UpdateChannelStatusById(id int, status int) {
|
||||||
err := UpdateAbilityStatus(id, status == common.ChannelStatusEnabled)
|
err := UpdateAbilityStatus(id, status == common.ChannelStatusEnabled)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
|||||||
@@ -3,10 +3,10 @@ package model
|
|||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
"one-api/common"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
"one-api/common/config"
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
"one-api/common/helper"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
|
|
||||||
"gorm.io/gorm"
|
"gorm.io/gorm"
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -2,14 +2,14 @@ package model
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"gorm.io/driver/mysql"
|
"gorm.io/driver/mysql"
|
||||||
"gorm.io/driver/postgres"
|
"gorm.io/driver/postgres"
|
||||||
"gorm.io/driver/sqlite"
|
"gorm.io/driver/sqlite"
|
||||||
"gorm.io/gorm"
|
"gorm.io/gorm"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
"os"
|
"os"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
@@ -72,7 +72,7 @@ func chooseDB() (*gorm.DB, error) {
|
|||||||
func InitDB() (err error) {
|
func InitDB() (err error) {
|
||||||
db, err := chooseDB()
|
db, err := chooseDB()
|
||||||
if err == nil {
|
if err == nil {
|
||||||
if config.DebugSQLEnabled {
|
if config.DebugEnabled {
|
||||||
db = db.Debug()
|
db = db.Debug()
|
||||||
}
|
}
|
||||||
DB = db
|
DB = db
|
||||||
|
|||||||
@@ -1,9 +1,9 @@
|
|||||||
package model
|
package model
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/songquanpeng/one-api/common"
|
"one-api/common"
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
"one-api/common/config"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
@@ -57,8 +57,6 @@ func InitOptionMap() {
|
|||||||
config.OptionMap["WeChatServerAddress"] = ""
|
config.OptionMap["WeChatServerAddress"] = ""
|
||||||
config.OptionMap["WeChatServerToken"] = ""
|
config.OptionMap["WeChatServerToken"] = ""
|
||||||
config.OptionMap["WeChatAccountQRCodeImageURL"] = ""
|
config.OptionMap["WeChatAccountQRCodeImageURL"] = ""
|
||||||
config.OptionMap["MessagePusherAddress"] = ""
|
|
||||||
config.OptionMap["MessagePusherToken"] = ""
|
|
||||||
config.OptionMap["TurnstileSiteKey"] = ""
|
config.OptionMap["TurnstileSiteKey"] = ""
|
||||||
config.OptionMap["TurnstileSecretKey"] = ""
|
config.OptionMap["TurnstileSecretKey"] = ""
|
||||||
config.OptionMap["QuotaForNewUser"] = strconv.Itoa(config.QuotaForNewUser)
|
config.OptionMap["QuotaForNewUser"] = strconv.Itoa(config.QuotaForNewUser)
|
||||||
@@ -68,7 +66,6 @@ func InitOptionMap() {
|
|||||||
config.OptionMap["PreConsumedQuota"] = strconv.Itoa(config.PreConsumedQuota)
|
config.OptionMap["PreConsumedQuota"] = strconv.Itoa(config.PreConsumedQuota)
|
||||||
config.OptionMap["ModelRatio"] = common.ModelRatio2JSONString()
|
config.OptionMap["ModelRatio"] = common.ModelRatio2JSONString()
|
||||||
config.OptionMap["GroupRatio"] = common.GroupRatio2JSONString()
|
config.OptionMap["GroupRatio"] = common.GroupRatio2JSONString()
|
||||||
config.OptionMap["CompletionRatio"] = common.CompletionRatio2JSONString()
|
|
||||||
config.OptionMap["TopUpLink"] = config.TopUpLink
|
config.OptionMap["TopUpLink"] = config.TopUpLink
|
||||||
config.OptionMap["ChatLink"] = config.ChatLink
|
config.OptionMap["ChatLink"] = config.ChatLink
|
||||||
config.OptionMap["QuotaPerUnit"] = strconv.FormatFloat(config.QuotaPerUnit, 'f', -1, 64)
|
config.OptionMap["QuotaPerUnit"] = strconv.FormatFloat(config.QuotaPerUnit, 'f', -1, 64)
|
||||||
@@ -181,10 +178,6 @@ func updateOptionMap(key string, value string) (err error) {
|
|||||||
config.WeChatServerToken = value
|
config.WeChatServerToken = value
|
||||||
case "WeChatAccountQRCodeImageURL":
|
case "WeChatAccountQRCodeImageURL":
|
||||||
config.WeChatAccountQRCodeImageURL = value
|
config.WeChatAccountQRCodeImageURL = value
|
||||||
case "MessagePusherAddress":
|
|
||||||
config.MessagePusherAddress = value
|
|
||||||
case "MessagePusherToken":
|
|
||||||
config.MessagePusherToken = value
|
|
||||||
case "TurnstileSiteKey":
|
case "TurnstileSiteKey":
|
||||||
config.TurnstileSiteKey = value
|
config.TurnstileSiteKey = value
|
||||||
case "TurnstileSecretKey":
|
case "TurnstileSecretKey":
|
||||||
@@ -205,8 +198,6 @@ func updateOptionMap(key string, value string) (err error) {
|
|||||||
err = common.UpdateModelRatioByJSONString(value)
|
err = common.UpdateModelRatioByJSONString(value)
|
||||||
case "GroupRatio":
|
case "GroupRatio":
|
||||||
err = common.UpdateGroupRatioByJSONString(value)
|
err = common.UpdateGroupRatioByJSONString(value)
|
||||||
case "CompletionRatio":
|
|
||||||
err = common.UpdateCompletionRatioByJSONString(value)
|
|
||||||
case "TopUpLink":
|
case "TopUpLink":
|
||||||
config.TopUpLink = value
|
config.TopUpLink = value
|
||||||
case "ChatLink":
|
case "ChatLink":
|
||||||
|
|||||||
@@ -3,9 +3,9 @@ package model
|
|||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"gorm.io/gorm"
|
"gorm.io/gorm"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/helper"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Redemption struct {
|
type Redemption struct {
|
||||||
|
|||||||
@@ -3,12 +3,11 @@ package model
|
|||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/common/message"
|
|
||||||
"gorm.io/gorm"
|
"gorm.io/gorm"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Token struct {
|
type Token struct {
|
||||||
@@ -214,7 +213,7 @@ func PreConsumeTokenQuota(tokenId int, quota int) (err error) {
|
|||||||
}
|
}
|
||||||
if email != "" {
|
if email != "" {
|
||||||
topUpLink := fmt.Sprintf("%s/topup", config.ServerAddress)
|
topUpLink := fmt.Sprintf("%s/topup", config.ServerAddress)
|
||||||
err = message.SendEmail(prompt, email,
|
err = common.SendEmail(prompt, email,
|
||||||
fmt.Sprintf("%s,当前剩余额度为 %d,为了不影响您的使用,请及时充值。<br/>充值链接:<a href='%s'>%s</a>", prompt, userQuota, topUpLink, topUpLink))
|
fmt.Sprintf("%s,当前剩余额度为 %d,为了不影响您的使用,请及时充值。<br/>充值链接:<a href='%s'>%s</a>", prompt, userQuota, topUpLink, topUpLink))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logger.SysError("failed to send email" + err.Error())
|
logger.SysError("failed to send email" + err.Error())
|
||||||
|
|||||||
@@ -3,12 +3,11 @@ package model
|
|||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/blacklist"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"gorm.io/gorm"
|
"gorm.io/gorm"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -41,7 +40,7 @@ func GetMaxUserId() int {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func GetAllUsers(startIdx int, num int) (users []*User, err error) {
|
func GetAllUsers(startIdx int, num int) (users []*User, err error) {
|
||||||
err = DB.Order("id desc").Limit(num).Offset(startIdx).Omit("password").Where("status != ?", common.UserStatusDeleted).Find(&users).Error
|
err = DB.Order("id desc").Limit(num).Offset(startIdx).Omit("password").Find(&users).Error
|
||||||
return users, err
|
return users, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -124,11 +123,6 @@ func (user *User) Update(updatePassword bool) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if user.Status == common.UserStatusDisabled {
|
|
||||||
blacklist.BanUser(user.Id)
|
|
||||||
} else if user.Status == common.UserStatusEnabled {
|
|
||||||
blacklist.UnbanUser(user.Id)
|
|
||||||
}
|
|
||||||
err = DB.Model(user).Updates(user).Error
|
err = DB.Model(user).Updates(user).Error
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -137,10 +131,7 @@ func (user *User) Delete() error {
|
|||||||
if user.Id == 0 {
|
if user.Id == 0 {
|
||||||
return errors.New("id 为空!")
|
return errors.New("id 为空!")
|
||||||
}
|
}
|
||||||
blacklist.BanUser(user.Id)
|
err := DB.Delete(user).Error
|
||||||
user.Username = fmt.Sprintf("deleted_%s", helper.GetUUID())
|
|
||||||
user.Status = common.UserStatusDeleted
|
|
||||||
err := DB.Model(user).Updates(user).Error
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
package model
|
package model
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
"one-api/common/config"
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
"one-api/common/logger"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|||||||
@@ -1,55 +0,0 @@
|
|||||||
package monitor
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/common/message"
|
|
||||||
"github.com/songquanpeng/one-api/model"
|
|
||||||
)
|
|
||||||
|
|
||||||
func notifyRootUser(subject string, content string) {
|
|
||||||
if config.MessagePusherAddress != "" {
|
|
||||||
err := message.SendMessage(subject, content, content)
|
|
||||||
if err != nil {
|
|
||||||
logger.SysError(fmt.Sprintf("failed to send message: %s", err.Error()))
|
|
||||||
} else {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if config.RootUserEmail == "" {
|
|
||||||
config.RootUserEmail = model.GetRootUserEmail()
|
|
||||||
}
|
|
||||||
err := message.SendEmail(subject, config.RootUserEmail, content)
|
|
||||||
if err != nil {
|
|
||||||
logger.SysError(fmt.Sprintf("failed to send email: %s", err.Error()))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// DisableChannel disable & notify
|
|
||||||
func DisableChannel(channelId int, channelName string, reason string) {
|
|
||||||
model.UpdateChannelStatusById(channelId, common.ChannelStatusAutoDisabled)
|
|
||||||
logger.SysLog(fmt.Sprintf("channel #%d has been disabled: %s", channelId, reason))
|
|
||||||
subject := fmt.Sprintf("通道「%s」(#%d)已被禁用", channelName, channelId)
|
|
||||||
content := fmt.Sprintf("通道「%s」(#%d)已被禁用,原因:%s", channelName, channelId, reason)
|
|
||||||
notifyRootUser(subject, content)
|
|
||||||
}
|
|
||||||
|
|
||||||
func MetricDisableChannel(channelId int, successRate float64) {
|
|
||||||
model.UpdateChannelStatusById(channelId, common.ChannelStatusAutoDisabled)
|
|
||||||
logger.SysLog(fmt.Sprintf("channel #%d has been disabled due to low success rate: %.2f", channelId, successRate*100))
|
|
||||||
subject := fmt.Sprintf("通道 #%d 已被禁用", channelId)
|
|
||||||
content := fmt.Sprintf("该渠道在最近 %d 次调用中成功率为 %.2f%%,低于阈值 %.2f%%,因此被系统自动禁用。",
|
|
||||||
config.MetricQueueSize, successRate*100, config.MetricSuccessRateThreshold*100)
|
|
||||||
notifyRootUser(subject, content)
|
|
||||||
}
|
|
||||||
|
|
||||||
// EnableChannel enable & notify
|
|
||||||
func EnableChannel(channelId int, channelName string) {
|
|
||||||
model.UpdateChannelStatusById(channelId, common.ChannelStatusEnabled)
|
|
||||||
logger.SysLog(fmt.Sprintf("channel #%d has been enabled", channelId))
|
|
||||||
subject := fmt.Sprintf("通道「%s」(#%d)已被启用", channelName, channelId)
|
|
||||||
content := fmt.Sprintf("通道「%s」(#%d)已被启用", channelName, channelId)
|
|
||||||
notifyRootUser(subject, content)
|
|
||||||
}
|
|
||||||
@@ -1,79 +0,0 @@
|
|||||||
package monitor
|
|
||||||
|
|
||||||
import (
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
)
|
|
||||||
|
|
||||||
var store = make(map[int][]bool)
|
|
||||||
var metricSuccessChan = make(chan int, config.MetricSuccessChanSize)
|
|
||||||
var metricFailChan = make(chan int, config.MetricFailChanSize)
|
|
||||||
|
|
||||||
func consumeSuccess(channelId int) {
|
|
||||||
if len(store[channelId]) > config.MetricQueueSize {
|
|
||||||
store[channelId] = store[channelId][1:]
|
|
||||||
}
|
|
||||||
store[channelId] = append(store[channelId], true)
|
|
||||||
}
|
|
||||||
|
|
||||||
func consumeFail(channelId int) (bool, float64) {
|
|
||||||
if len(store[channelId]) > config.MetricQueueSize {
|
|
||||||
store[channelId] = store[channelId][1:]
|
|
||||||
}
|
|
||||||
store[channelId] = append(store[channelId], false)
|
|
||||||
successCount := 0
|
|
||||||
for _, success := range store[channelId] {
|
|
||||||
if success {
|
|
||||||
successCount++
|
|
||||||
}
|
|
||||||
}
|
|
||||||
successRate := float64(successCount) / float64(len(store[channelId]))
|
|
||||||
if len(store[channelId]) < config.MetricQueueSize {
|
|
||||||
return false, successRate
|
|
||||||
}
|
|
||||||
if successRate < config.MetricSuccessRateThreshold {
|
|
||||||
store[channelId] = make([]bool, 0)
|
|
||||||
return true, successRate
|
|
||||||
}
|
|
||||||
return false, successRate
|
|
||||||
}
|
|
||||||
|
|
||||||
func metricSuccessConsumer() {
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case channelId := <-metricSuccessChan:
|
|
||||||
consumeSuccess(channelId)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func metricFailConsumer() {
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case channelId := <-metricFailChan:
|
|
||||||
disable, successRate := consumeFail(channelId)
|
|
||||||
if disable {
|
|
||||||
go MetricDisableChannel(channelId, successRate)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func init() {
|
|
||||||
if config.EnableMetric {
|
|
||||||
go metricSuccessConsumer()
|
|
||||||
go metricFailConsumer()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func Emit(channelId int, success bool) {
|
|
||||||
if !config.EnableMetric {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
go func() {
|
|
||||||
if success {
|
|
||||||
metricSuccessChan <- channelId
|
|
||||||
} else {
|
|
||||||
metricFailChan <- channelId
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
}
|
|
||||||
@@ -1,8 +0,0 @@
|
|||||||
package ai360
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"360GPT_S2_V9",
|
|
||||||
"embedding-bert-512-v1",
|
|
||||||
"embedding_s1_v1",
|
|
||||||
"semantic_similarity_s1_v1",
|
|
||||||
}
|
|
||||||
@@ -1,60 +1,22 @@
|
|||||||
package aiproxy
|
package aiproxy
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Adaptor struct {
|
type Adaptor struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) Init(meta *util.RelayMeta) {
|
func (a *Adaptor) Auth(c *gin.Context) error {
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetRequestURL(meta *util.RelayMeta) (string, error) {
|
|
||||||
return fmt.Sprintf("%s/api/library/ask", meta.BaseURL), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) SetupRequestHeader(c *gin.Context, req *http.Request, meta *util.RelayMeta) error {
|
|
||||||
channel.SetupCommonRequestHeader(c, req, meta)
|
|
||||||
req.Header.Set("Authorization", "Bearer "+meta.APIKey)
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) ConvertRequest(c *gin.Context, relayMode int, request *model.GeneralOpenAIRequest) (any, error) {
|
func (a *Adaptor) ConvertRequest(request *openai.GeneralOpenAIRequest) (any, error) {
|
||||||
if request == nil {
|
return nil, nil
|
||||||
return nil, errors.New("request is nil")
|
|
||||||
}
|
|
||||||
aiProxyLibraryRequest := ConvertRequest(*request)
|
|
||||||
aiProxyLibraryRequest.LibraryId = c.GetString(common.ConfigKeyLibraryID)
|
|
||||||
return aiProxyLibraryRequest, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) DoRequest(c *gin.Context, meta *util.RelayMeta, requestBody io.Reader) (*http.Response, error) {
|
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage, error) {
|
||||||
return channel.DoRequestHelper(a, c, meta, requestBody)
|
return nil, nil, nil
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, meta *util.RelayMeta) (usage *model.Usage, err *model.ErrorWithStatusCode) {
|
|
||||||
if meta.IsStream {
|
|
||||||
err, usage = StreamHandler(c, resp)
|
|
||||||
} else {
|
|
||||||
err, usage = Handler(c, resp)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetModelList() []string {
|
|
||||||
return ModelList
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetChannelName() string {
|
|
||||||
return "aiproxy"
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,9 +0,0 @@
|
|||||||
package aiproxy
|
|
||||||
|
|
||||||
import "github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
|
|
||||||
var ModelList = []string{""}
|
|
||||||
|
|
||||||
func init() {
|
|
||||||
ModelList = openai.ModelList
|
|
||||||
}
|
|
||||||
@@ -5,21 +5,20 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
|
"one-api/relay/constant"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
// https://docs.aiproxy.io/dev/library#使用已经定制好的知识库进行对话问答
|
// https://docs.aiproxy.io/dev/library#使用已经定制好的知识库进行对话问答
|
||||||
|
|
||||||
func ConvertRequest(request model.GeneralOpenAIRequest) *LibraryRequest {
|
func ConvertRequest(request openai.GeneralOpenAIRequest) *LibraryRequest {
|
||||||
query := ""
|
query := ""
|
||||||
if len(request.Messages) != 0 {
|
if len(request.Messages) != 0 {
|
||||||
query = request.Messages[len(request.Messages)-1].StringContent()
|
query = request.Messages[len(request.Messages)-1].StringContent()
|
||||||
@@ -46,14 +45,14 @@ func responseAIProxyLibrary2OpenAI(response *LibraryResponse) *openai.TextRespon
|
|||||||
content := response.Answer + aiProxyDocuments2Markdown(response.Documents)
|
content := response.Answer + aiProxyDocuments2Markdown(response.Documents)
|
||||||
choice := openai.TextResponseChoice{
|
choice := openai.TextResponseChoice{
|
||||||
Index: 0,
|
Index: 0,
|
||||||
Message: model.Message{
|
Message: openai.Message{
|
||||||
Role: "assistant",
|
Role: "assistant",
|
||||||
Content: content,
|
Content: content,
|
||||||
},
|
},
|
||||||
FinishReason: "stop",
|
FinishReason: "stop",
|
||||||
}
|
}
|
||||||
fullTextResponse := openai.TextResponse{
|
fullTextResponse := openai.TextResponse{
|
||||||
Id: fmt.Sprintf("chatcmpl-%s", helper.GetUUID()),
|
Id: helper.GetUUID(),
|
||||||
Object: "chat.completion",
|
Object: "chat.completion",
|
||||||
Created: helper.GetTimestamp(),
|
Created: helper.GetTimestamp(),
|
||||||
Choices: []openai.TextResponseChoice{choice},
|
Choices: []openai.TextResponseChoice{choice},
|
||||||
@@ -66,7 +65,7 @@ func documentsAIProxyLibrary(documents []LibraryDocument) *openai.ChatCompletion
|
|||||||
choice.Delta.Content = aiProxyDocuments2Markdown(documents)
|
choice.Delta.Content = aiProxyDocuments2Markdown(documents)
|
||||||
choice.FinishReason = &constant.StopFinishReason
|
choice.FinishReason = &constant.StopFinishReason
|
||||||
return &openai.ChatCompletionsStreamResponse{
|
return &openai.ChatCompletionsStreamResponse{
|
||||||
Id: fmt.Sprintf("chatcmpl-%s", helper.GetUUID()),
|
Id: helper.GetUUID(),
|
||||||
Object: "chat.completion.chunk",
|
Object: "chat.completion.chunk",
|
||||||
Created: helper.GetTimestamp(),
|
Created: helper.GetTimestamp(),
|
||||||
Model: "",
|
Model: "",
|
||||||
@@ -78,7 +77,7 @@ func streamResponseAIProxyLibrary2OpenAI(response *LibraryStreamResponse) *opena
|
|||||||
var choice openai.ChatCompletionsStreamResponseChoice
|
var choice openai.ChatCompletionsStreamResponseChoice
|
||||||
choice.Delta.Content = response.Content
|
choice.Delta.Content = response.Content
|
||||||
return &openai.ChatCompletionsStreamResponse{
|
return &openai.ChatCompletionsStreamResponse{
|
||||||
Id: fmt.Sprintf("chatcmpl-%s", helper.GetUUID()),
|
Id: helper.GetUUID(),
|
||||||
Object: "chat.completion.chunk",
|
Object: "chat.completion.chunk",
|
||||||
Created: helper.GetTimestamp(),
|
Created: helper.GetTimestamp(),
|
||||||
Model: response.Model,
|
Model: response.Model,
|
||||||
@@ -86,8 +85,8 @@ func streamResponseAIProxyLibrary2OpenAI(response *LibraryStreamResponse) *opena
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func StreamHandler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
var usage model.Usage
|
var usage openai.Usage
|
||||||
scanner := bufio.NewScanner(resp.Body)
|
scanner := bufio.NewScanner(resp.Body)
|
||||||
scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
|
scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
|
||||||
if atEOF && len(data) == 0 {
|
if atEOF && len(data) == 0 {
|
||||||
@@ -158,7 +157,7 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
|
|||||||
return nil, &usage
|
return nil, &usage
|
||||||
}
|
}
|
||||||
|
|
||||||
func Handler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func Handler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
var AIProxyLibraryResponse LibraryResponse
|
var AIProxyLibraryResponse LibraryResponse
|
||||||
responseBody, err := io.ReadAll(resp.Body)
|
responseBody, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -173,8 +172,8 @@ func Handler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *
|
|||||||
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
if AIProxyLibraryResponse.ErrCode != 0 {
|
if AIProxyLibraryResponse.ErrCode != 0 {
|
||||||
return &model.ErrorWithStatusCode{
|
return &openai.ErrorWithStatusCode{
|
||||||
Error: model.Error{
|
Error: openai.Error{
|
||||||
Message: AIProxyLibraryResponse.Message,
|
Message: AIProxyLibraryResponse.Message,
|
||||||
Type: strconv.Itoa(AIProxyLibraryResponse.ErrCode),
|
Type: strconv.Itoa(AIProxyLibraryResponse.ErrCode),
|
||||||
Code: AIProxyLibraryResponse.ErrCode,
|
Code: AIProxyLibraryResponse.ErrCode,
|
||||||
@@ -190,8 +189,5 @@ func Handler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *
|
|||||||
c.Writer.Header().Set("Content-Type", "application/json")
|
c.Writer.Header().Set("Content-Type", "application/json")
|
||||||
c.Writer.WriteHeader(resp.StatusCode)
|
c.Writer.WriteHeader(resp.StatusCode)
|
||||||
_, err = c.Writer.Write(jsonResponse)
|
_, err = c.Writer.Write(jsonResponse)
|
||||||
if err != nil {
|
|
||||||
return openai.ErrorWrapper(err, "write_response_body_failed", http.StatusInternalServerError), nil
|
|
||||||
}
|
|
||||||
return nil, &fullTextResponse.Usage
|
return nil, &fullTextResponse.Usage
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,83 +1,22 @@
|
|||||||
package ali
|
package ali
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
)
|
)
|
||||||
|
|
||||||
// https://help.aliyun.com/zh/dashscope/developer-reference/api-details
|
|
||||||
|
|
||||||
type Adaptor struct {
|
type Adaptor struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) Init(meta *util.RelayMeta) {
|
func (a *Adaptor) Auth(c *gin.Context) error {
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetRequestURL(meta *util.RelayMeta) (string, error) {
|
|
||||||
fullRequestURL := fmt.Sprintf("%s/api/v1/services/aigc/text-generation/generation", meta.BaseURL)
|
|
||||||
if meta.Mode == constant.RelayModeEmbeddings {
|
|
||||||
fullRequestURL = fmt.Sprintf("%s/api/v1/services/embeddings/text-embedding/text-embedding", meta.BaseURL)
|
|
||||||
}
|
|
||||||
return fullRequestURL, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) SetupRequestHeader(c *gin.Context, req *http.Request, meta *util.RelayMeta) error {
|
|
||||||
channel.SetupCommonRequestHeader(c, req, meta)
|
|
||||||
req.Header.Set("Authorization", "Bearer "+meta.APIKey)
|
|
||||||
if meta.IsStream {
|
|
||||||
req.Header.Set("X-DashScope-SSE", "enable")
|
|
||||||
}
|
|
||||||
if c.GetString(common.ConfigKeyPlugin) != "" {
|
|
||||||
req.Header.Set("X-DashScope-Plugin", c.GetString(common.ConfigKeyPlugin))
|
|
||||||
}
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) ConvertRequest(c *gin.Context, relayMode int, request *model.GeneralOpenAIRequest) (any, error) {
|
func (a *Adaptor) ConvertRequest(request *openai.GeneralOpenAIRequest) (any, error) {
|
||||||
if request == nil {
|
return nil, nil
|
||||||
return nil, errors.New("request is nil")
|
|
||||||
}
|
|
||||||
switch relayMode {
|
|
||||||
case constant.RelayModeEmbeddings:
|
|
||||||
baiduEmbeddingRequest := ConvertEmbeddingRequest(*request)
|
|
||||||
return baiduEmbeddingRequest, nil
|
|
||||||
default:
|
|
||||||
baiduRequest := ConvertRequest(*request)
|
|
||||||
return baiduRequest, nil
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) DoRequest(c *gin.Context, meta *util.RelayMeta, requestBody io.Reader) (*http.Response, error) {
|
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage, error) {
|
||||||
return channel.DoRequestHelper(a, c, meta, requestBody)
|
return nil, nil, nil
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, meta *util.RelayMeta) (usage *model.Usage, err *model.ErrorWithStatusCode) {
|
|
||||||
if meta.IsStream {
|
|
||||||
err, usage = StreamHandler(c, resp)
|
|
||||||
} else {
|
|
||||||
switch meta.Mode {
|
|
||||||
case constant.RelayModeEmbeddings:
|
|
||||||
err, usage = EmbeddingHandler(c, resp)
|
|
||||||
default:
|
|
||||||
err, usage = Handler(c, resp)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetModelList() []string {
|
|
||||||
return ModelList
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetChannelName() string {
|
|
||||||
return "ali"
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +0,0 @@
|
|||||||
package ali
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"qwen-turbo", "qwen-plus", "qwen-max", "qwen-max-longcontext",
|
|
||||||
"text-embedding-v1",
|
|
||||||
}
|
|
||||||
@@ -4,13 +4,12 @@ import (
|
|||||||
"bufio"
|
"bufio"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -18,7 +17,7 @@ import (
|
|||||||
|
|
||||||
const EnableSearchModelSuffix = "-internet"
|
const EnableSearchModelSuffix = "-internet"
|
||||||
|
|
||||||
func ConvertRequest(request model.GeneralOpenAIRequest) *ChatRequest {
|
func ConvertRequest(request openai.GeneralOpenAIRequest) *ChatRequest {
|
||||||
messages := make([]Message, 0, len(request.Messages))
|
messages := make([]Message, 0, len(request.Messages))
|
||||||
for i := 0; i < len(request.Messages); i++ {
|
for i := 0; i < len(request.Messages); i++ {
|
||||||
message := request.Messages[i]
|
message := request.Messages[i]
|
||||||
@@ -33,9 +32,6 @@ func ConvertRequest(request model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
enableSearch = true
|
enableSearch = true
|
||||||
aliModel = strings.TrimSuffix(aliModel, EnableSearchModelSuffix)
|
aliModel = strings.TrimSuffix(aliModel, EnableSearchModelSuffix)
|
||||||
}
|
}
|
||||||
if request.TopP >= 1 {
|
|
||||||
request.TopP = 0.9999
|
|
||||||
}
|
|
||||||
return &ChatRequest{
|
return &ChatRequest{
|
||||||
Model: aliModel,
|
Model: aliModel,
|
||||||
Input: Input{
|
Input: Input{
|
||||||
@@ -44,15 +40,11 @@ func ConvertRequest(request model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
Parameters: Parameters{
|
Parameters: Parameters{
|
||||||
EnableSearch: enableSearch,
|
EnableSearch: enableSearch,
|
||||||
IncrementalOutput: request.Stream,
|
IncrementalOutput: request.Stream,
|
||||||
Seed: uint64(request.Seed),
|
|
||||||
MaxTokens: request.MaxTokens,
|
|
||||||
Temperature: request.Temperature,
|
|
||||||
TopP: request.TopP,
|
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func ConvertEmbeddingRequest(request model.GeneralOpenAIRequest) *EmbeddingRequest {
|
func ConvertEmbeddingRequest(request openai.GeneralOpenAIRequest) *EmbeddingRequest {
|
||||||
return &EmbeddingRequest{
|
return &EmbeddingRequest{
|
||||||
Model: "text-embedding-v1",
|
Model: "text-embedding-v1",
|
||||||
Input: struct {
|
Input: struct {
|
||||||
@@ -63,7 +55,7 @@ func ConvertEmbeddingRequest(request model.GeneralOpenAIRequest) *EmbeddingReque
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func EmbeddingHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func EmbeddingHandler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
var aliResponse EmbeddingResponse
|
var aliResponse EmbeddingResponse
|
||||||
err := json.NewDecoder(resp.Body).Decode(&aliResponse)
|
err := json.NewDecoder(resp.Body).Decode(&aliResponse)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -76,8 +68,8 @@ func EmbeddingHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStat
|
|||||||
}
|
}
|
||||||
|
|
||||||
if aliResponse.Code != "" {
|
if aliResponse.Code != "" {
|
||||||
return &model.ErrorWithStatusCode{
|
return &openai.ErrorWithStatusCode{
|
||||||
Error: model.Error{
|
Error: openai.Error{
|
||||||
Message: aliResponse.Message,
|
Message: aliResponse.Message,
|
||||||
Type: aliResponse.Code,
|
Type: aliResponse.Code,
|
||||||
Param: aliResponse.RequestId,
|
Param: aliResponse.RequestId,
|
||||||
@@ -103,7 +95,7 @@ func embeddingResponseAli2OpenAI(response *EmbeddingResponse) *openai.EmbeddingR
|
|||||||
Object: "list",
|
Object: "list",
|
||||||
Data: make([]openai.EmbeddingResponseItem, 0, len(response.Output.Embeddings)),
|
Data: make([]openai.EmbeddingResponseItem, 0, len(response.Output.Embeddings)),
|
||||||
Model: "text-embedding-v1",
|
Model: "text-embedding-v1",
|
||||||
Usage: model.Usage{TotalTokens: response.Usage.TotalTokens},
|
Usage: openai.Usage{TotalTokens: response.Usage.TotalTokens},
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, item := range response.Output.Embeddings {
|
for _, item := range response.Output.Embeddings {
|
||||||
@@ -119,7 +111,7 @@ func embeddingResponseAli2OpenAI(response *EmbeddingResponse) *openai.EmbeddingR
|
|||||||
func responseAli2OpenAI(response *ChatResponse) *openai.TextResponse {
|
func responseAli2OpenAI(response *ChatResponse) *openai.TextResponse {
|
||||||
choice := openai.TextResponseChoice{
|
choice := openai.TextResponseChoice{
|
||||||
Index: 0,
|
Index: 0,
|
||||||
Message: model.Message{
|
Message: openai.Message{
|
||||||
Role: "assistant",
|
Role: "assistant",
|
||||||
Content: response.Output.Text,
|
Content: response.Output.Text,
|
||||||
},
|
},
|
||||||
@@ -130,7 +122,7 @@ func responseAli2OpenAI(response *ChatResponse) *openai.TextResponse {
|
|||||||
Object: "chat.completion",
|
Object: "chat.completion",
|
||||||
Created: helper.GetTimestamp(),
|
Created: helper.GetTimestamp(),
|
||||||
Choices: []openai.TextResponseChoice{choice},
|
Choices: []openai.TextResponseChoice{choice},
|
||||||
Usage: model.Usage{
|
Usage: openai.Usage{
|
||||||
PromptTokens: response.Usage.InputTokens,
|
PromptTokens: response.Usage.InputTokens,
|
||||||
CompletionTokens: response.Usage.OutputTokens,
|
CompletionTokens: response.Usage.OutputTokens,
|
||||||
TotalTokens: response.Usage.InputTokens + response.Usage.OutputTokens,
|
TotalTokens: response.Usage.InputTokens + response.Usage.OutputTokens,
|
||||||
@@ -156,8 +148,8 @@ func streamResponseAli2OpenAI(aliResponse *ChatResponse) *openai.ChatCompletions
|
|||||||
return &response
|
return &response
|
||||||
}
|
}
|
||||||
|
|
||||||
func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func StreamHandler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
var usage model.Usage
|
var usage openai.Usage
|
||||||
scanner := bufio.NewScanner(resp.Body)
|
scanner := bufio.NewScanner(resp.Body)
|
||||||
scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
|
scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
|
||||||
if atEOF && len(data) == 0 {
|
if atEOF && len(data) == 0 {
|
||||||
@@ -225,7 +217,7 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
|
|||||||
return nil, &usage
|
return nil, &usage
|
||||||
}
|
}
|
||||||
|
|
||||||
func Handler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func Handler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
var aliResponse ChatResponse
|
var aliResponse ChatResponse
|
||||||
responseBody, err := io.ReadAll(resp.Body)
|
responseBody, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -240,8 +232,8 @@ func Handler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *
|
|||||||
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
if aliResponse.Code != "" {
|
if aliResponse.Code != "" {
|
||||||
return &model.ErrorWithStatusCode{
|
return &openai.ErrorWithStatusCode{
|
||||||
Error: model.Error{
|
Error: openai.Error{
|
||||||
Message: aliResponse.Message,
|
Message: aliResponse.Message,
|
||||||
Type: aliResponse.Code,
|
Type: aliResponse.Code,
|
||||||
Param: aliResponse.RequestId,
|
Param: aliResponse.RequestId,
|
||||||
|
|||||||
@@ -16,8 +16,6 @@ type Parameters struct {
|
|||||||
Seed uint64 `json:"seed,omitempty"`
|
Seed uint64 `json:"seed,omitempty"`
|
||||||
EnableSearch bool `json:"enable_search,omitempty"`
|
EnableSearch bool `json:"enable_search,omitempty"`
|
||||||
IncrementalOutput bool `json:"incremental_output,omitempty"`
|
IncrementalOutput bool `json:"incremental_output,omitempty"`
|
||||||
MaxTokens int `json:"max_tokens,omitempty"`
|
|
||||||
Temperature float64 `json:"temperature,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type ChatRequest struct {
|
type ChatRequest struct {
|
||||||
|
|||||||
@@ -1,63 +1,22 @@
|
|||||||
package anthropic
|
package anthropic
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/relay/channel"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Adaptor struct {
|
type Adaptor struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) Init(meta *util.RelayMeta) {
|
func (a *Adaptor) Auth(c *gin.Context) error {
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetRequestURL(meta *util.RelayMeta) (string, error) {
|
|
||||||
return fmt.Sprintf("%s/v1/messages", meta.BaseURL), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) SetupRequestHeader(c *gin.Context, req *http.Request, meta *util.RelayMeta) error {
|
|
||||||
channel.SetupCommonRequestHeader(c, req, meta)
|
|
||||||
req.Header.Set("x-api-key", meta.APIKey)
|
|
||||||
anthropicVersion := c.Request.Header.Get("anthropic-version")
|
|
||||||
if anthropicVersion == "" {
|
|
||||||
anthropicVersion = "2023-06-01"
|
|
||||||
}
|
|
||||||
req.Header.Set("anthropic-version", anthropicVersion)
|
|
||||||
req.Header.Set("anthropic-beta", "messages-2023-12-15")
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) ConvertRequest(c *gin.Context, relayMode int, request *model.GeneralOpenAIRequest) (any, error) {
|
func (a *Adaptor) ConvertRequest(request *openai.GeneralOpenAIRequest) (any, error) {
|
||||||
if request == nil {
|
return nil, nil
|
||||||
return nil, errors.New("request is nil")
|
|
||||||
}
|
|
||||||
return ConvertRequest(*request), nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) DoRequest(c *gin.Context, meta *util.RelayMeta, requestBody io.Reader) (*http.Response, error) {
|
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage, error) {
|
||||||
return channel.DoRequestHelper(a, c, meta, requestBody)
|
return nil, nil, nil
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, meta *util.RelayMeta) (usage *model.Usage, err *model.ErrorWithStatusCode) {
|
|
||||||
if meta.IsStream {
|
|
||||||
err, usage = StreamHandler(c, resp)
|
|
||||||
} else {
|
|
||||||
err, usage = Handler(c, resp, meta.PromptTokens, meta.ActualModelName)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetModelList() []string {
|
|
||||||
return ModelList
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetChannelName() string {
|
|
||||||
return "authropic"
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,8 +0,0 @@
|
|||||||
package anthropic
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"claude-instant-1.2", "claude-2.0", "claude-2.1",
|
|
||||||
"claude-3-haiku-20240229",
|
|
||||||
"claude-3-sonnet-20240229",
|
|
||||||
"claude-3-opus-20240229",
|
|
||||||
}
|
|
||||||
@@ -5,146 +5,82 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/image"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
"strings"
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
func stopReasonClaude2OpenAI(reason *string) string {
|
func stopReasonClaude2OpenAI(reason string) string {
|
||||||
if reason == nil {
|
switch reason {
|
||||||
return ""
|
|
||||||
}
|
|
||||||
switch *reason {
|
|
||||||
case "end_turn":
|
|
||||||
return "stop"
|
|
||||||
case "stop_sequence":
|
case "stop_sequence":
|
||||||
return "stop"
|
return "stop"
|
||||||
case "max_tokens":
|
case "max_tokens":
|
||||||
return "length"
|
return "length"
|
||||||
default:
|
default:
|
||||||
return *reason
|
return reason
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func ConvertRequest(textRequest model.GeneralOpenAIRequest) *Request {
|
func ConvertRequest(textRequest openai.GeneralOpenAIRequest) *Request {
|
||||||
claudeRequest := Request{
|
claudeRequest := Request{
|
||||||
Model: textRequest.Model,
|
Model: textRequest.Model,
|
||||||
MaxTokens: textRequest.MaxTokens,
|
Prompt: "",
|
||||||
Temperature: textRequest.Temperature,
|
MaxTokensToSample: textRequest.MaxTokens,
|
||||||
TopP: textRequest.TopP,
|
StopSequences: nil,
|
||||||
Stream: textRequest.Stream,
|
Temperature: textRequest.Temperature,
|
||||||
|
TopP: textRequest.TopP,
|
||||||
|
Stream: textRequest.Stream,
|
||||||
}
|
}
|
||||||
if claudeRequest.MaxTokens == 0 {
|
if claudeRequest.MaxTokensToSample == 0 {
|
||||||
claudeRequest.MaxTokens = 4096
|
claudeRequest.MaxTokensToSample = 1000000
|
||||||
}
|
|
||||||
// legacy model name mapping
|
|
||||||
if claudeRequest.Model == "claude-instant-1" {
|
|
||||||
claudeRequest.Model = "claude-instant-1.1"
|
|
||||||
} else if claudeRequest.Model == "claude-2" {
|
|
||||||
claudeRequest.Model = "claude-2.1"
|
|
||||||
}
|
}
|
||||||
|
prompt := ""
|
||||||
for _, message := range textRequest.Messages {
|
for _, message := range textRequest.Messages {
|
||||||
if message.Role == "system" && claudeRequest.System == "" {
|
if message.Role == "user" {
|
||||||
claudeRequest.System = message.StringContent()
|
prompt += fmt.Sprintf("\n\nHuman: %s", message.Content)
|
||||||
continue
|
} else if message.Role == "assistant" {
|
||||||
}
|
prompt += fmt.Sprintf("\n\nAssistant: %s", message.Content)
|
||||||
claudeMessage := Message{
|
} else if message.Role == "system" {
|
||||||
Role: message.Role,
|
if prompt == "" {
|
||||||
}
|
prompt = message.StringContent()
|
||||||
var content Content
|
|
||||||
if message.IsStringContent() {
|
|
||||||
content.Type = "text"
|
|
||||||
content.Text = message.StringContent()
|
|
||||||
claudeMessage.Content = append(claudeMessage.Content, content)
|
|
||||||
claudeRequest.Messages = append(claudeRequest.Messages, claudeMessage)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
var contents []Content
|
|
||||||
openaiContent := message.ParseContent()
|
|
||||||
for _, part := range openaiContent {
|
|
||||||
var content Content
|
|
||||||
if part.Type == model.ContentTypeText {
|
|
||||||
content.Type = "text"
|
|
||||||
content.Text = part.Text
|
|
||||||
} else if part.Type == model.ContentTypeImageURL {
|
|
||||||
content.Type = "image"
|
|
||||||
content.Source = &ImageSource{
|
|
||||||
Type: "base64",
|
|
||||||
}
|
|
||||||
mimeType, data, _ := image.GetImageFromUrl(part.ImageURL.Url)
|
|
||||||
content.Source.MediaType = mimeType
|
|
||||||
content.Source.Data = data
|
|
||||||
}
|
}
|
||||||
contents = append(contents, content)
|
|
||||||
}
|
}
|
||||||
claudeMessage.Content = contents
|
|
||||||
claudeRequest.Messages = append(claudeRequest.Messages, claudeMessage)
|
|
||||||
}
|
}
|
||||||
|
prompt += "\n\nAssistant:"
|
||||||
|
claudeRequest.Prompt = prompt
|
||||||
return &claudeRequest
|
return &claudeRequest
|
||||||
}
|
}
|
||||||
|
|
||||||
// https://docs.anthropic.com/claude/reference/messages-streaming
|
func streamResponseClaude2OpenAI(claudeResponse *Response) *openai.ChatCompletionsStreamResponse {
|
||||||
func streamResponseClaude2OpenAI(claudeResponse *StreamResponse) (*openai.ChatCompletionsStreamResponse, *Response) {
|
|
||||||
var response *Response
|
|
||||||
var responseText string
|
|
||||||
var stopReason string
|
|
||||||
switch claudeResponse.Type {
|
|
||||||
case "message_start":
|
|
||||||
return nil, claudeResponse.Message
|
|
||||||
case "content_block_start":
|
|
||||||
if claudeResponse.ContentBlock != nil {
|
|
||||||
responseText = claudeResponse.ContentBlock.Text
|
|
||||||
}
|
|
||||||
case "content_block_delta":
|
|
||||||
if claudeResponse.Delta != nil {
|
|
||||||
responseText = claudeResponse.Delta.Text
|
|
||||||
}
|
|
||||||
case "message_delta":
|
|
||||||
if claudeResponse.Usage != nil {
|
|
||||||
response = &Response{
|
|
||||||
Usage: *claudeResponse.Usage,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if claudeResponse.Delta != nil && claudeResponse.Delta.StopReason != nil {
|
|
||||||
stopReason = *claudeResponse.Delta.StopReason
|
|
||||||
}
|
|
||||||
}
|
|
||||||
var choice openai.ChatCompletionsStreamResponseChoice
|
var choice openai.ChatCompletionsStreamResponseChoice
|
||||||
choice.Delta.Content = responseText
|
choice.Delta.Content = claudeResponse.Completion
|
||||||
choice.Delta.Role = "assistant"
|
finishReason := stopReasonClaude2OpenAI(claudeResponse.StopReason)
|
||||||
finishReason := stopReasonClaude2OpenAI(&stopReason)
|
|
||||||
if finishReason != "null" {
|
if finishReason != "null" {
|
||||||
choice.FinishReason = &finishReason
|
choice.FinishReason = &finishReason
|
||||||
}
|
}
|
||||||
var openaiResponse openai.ChatCompletionsStreamResponse
|
var response openai.ChatCompletionsStreamResponse
|
||||||
openaiResponse.Object = "chat.completion.chunk"
|
response.Object = "chat.completion.chunk"
|
||||||
openaiResponse.Choices = []openai.ChatCompletionsStreamResponseChoice{choice}
|
response.Model = claudeResponse.Model
|
||||||
return &openaiResponse, response
|
response.Choices = []openai.ChatCompletionsStreamResponseChoice{choice}
|
||||||
|
return &response
|
||||||
}
|
}
|
||||||
|
|
||||||
func responseClaude2OpenAI(claudeResponse *Response) *openai.TextResponse {
|
func responseClaude2OpenAI(claudeResponse *Response) *openai.TextResponse {
|
||||||
var responseText string
|
|
||||||
if len(claudeResponse.Content) > 0 {
|
|
||||||
responseText = claudeResponse.Content[0].Text
|
|
||||||
}
|
|
||||||
choice := openai.TextResponseChoice{
|
choice := openai.TextResponseChoice{
|
||||||
Index: 0,
|
Index: 0,
|
||||||
Message: model.Message{
|
Message: openai.Message{
|
||||||
Role: "assistant",
|
Role: "assistant",
|
||||||
Content: responseText,
|
Content: strings.TrimPrefix(claudeResponse.Completion, " "),
|
||||||
Name: nil,
|
Name: nil,
|
||||||
},
|
},
|
||||||
FinishReason: stopReasonClaude2OpenAI(claudeResponse.StopReason),
|
FinishReason: stopReasonClaude2OpenAI(claudeResponse.StopReason),
|
||||||
}
|
}
|
||||||
fullTextResponse := openai.TextResponse{
|
fullTextResponse := openai.TextResponse{
|
||||||
Id: fmt.Sprintf("chatcmpl-%s", claudeResponse.Id),
|
Id: fmt.Sprintf("chatcmpl-%s", helper.GetUUID()),
|
||||||
Model: claudeResponse.Model,
|
|
||||||
Object: "chat.completion",
|
Object: "chat.completion",
|
||||||
Created: helper.GetTimestamp(),
|
Created: helper.GetTimestamp(),
|
||||||
Choices: []openai.TextResponseChoice{choice},
|
Choices: []openai.TextResponseChoice{choice},
|
||||||
@@ -152,15 +88,17 @@ func responseClaude2OpenAI(claudeResponse *Response) *openai.TextResponse {
|
|||||||
return &fullTextResponse
|
return &fullTextResponse
|
||||||
}
|
}
|
||||||
|
|
||||||
func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func StreamHandler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, string) {
|
||||||
|
responseText := ""
|
||||||
|
responseId := fmt.Sprintf("chatcmpl-%s", helper.GetUUID())
|
||||||
createdTime := helper.GetTimestamp()
|
createdTime := helper.GetTimestamp()
|
||||||
scanner := bufio.NewScanner(resp.Body)
|
scanner := bufio.NewScanner(resp.Body)
|
||||||
scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
|
scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
|
||||||
if atEOF && len(data) == 0 {
|
if atEOF && len(data) == 0 {
|
||||||
return 0, nil, nil
|
return 0, nil, nil
|
||||||
}
|
}
|
||||||
if i := strings.Index(string(data), "\n"); i >= 0 {
|
if i := strings.Index(string(data), "\r\n\r\n"); i >= 0 {
|
||||||
return i + 1, data[0:i], nil
|
return i + 4, data[0:i], nil
|
||||||
}
|
}
|
||||||
if atEOF {
|
if atEOF {
|
||||||
return len(data), data, nil
|
return len(data), data, nil
|
||||||
@@ -172,45 +110,29 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
|
|||||||
go func() {
|
go func() {
|
||||||
for scanner.Scan() {
|
for scanner.Scan() {
|
||||||
data := scanner.Text()
|
data := scanner.Text()
|
||||||
if len(data) < 6 {
|
if !strings.HasPrefix(data, "event: completion") {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if !strings.HasPrefix(data, "data: ") {
|
data = strings.TrimPrefix(data, "event: completion\r\ndata: ")
|
||||||
continue
|
|
||||||
}
|
|
||||||
data = strings.TrimPrefix(data, "data: ")
|
|
||||||
dataChan <- data
|
dataChan <- data
|
||||||
}
|
}
|
||||||
stopChan <- true
|
stopChan <- true
|
||||||
}()
|
}()
|
||||||
common.SetEventStreamHeaders(c)
|
common.SetEventStreamHeaders(c)
|
||||||
var usage model.Usage
|
|
||||||
var modelName string
|
|
||||||
var id string
|
|
||||||
c.Stream(func(w io.Writer) bool {
|
c.Stream(func(w io.Writer) bool {
|
||||||
select {
|
select {
|
||||||
case data := <-dataChan:
|
case data := <-dataChan:
|
||||||
// some implementations may add \r at the end of data
|
// some implementations may add \r at the end of data
|
||||||
data = strings.TrimSuffix(data, "\r")
|
data = strings.TrimSuffix(data, "\r")
|
||||||
var claudeResponse StreamResponse
|
var claudeResponse Response
|
||||||
err := json.Unmarshal([]byte(data), &claudeResponse)
|
err := json.Unmarshal([]byte(data), &claudeResponse)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logger.SysError("error unmarshalling stream response: " + err.Error())
|
logger.SysError("error unmarshalling stream response: " + err.Error())
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
response, meta := streamResponseClaude2OpenAI(&claudeResponse)
|
responseText += claudeResponse.Completion
|
||||||
if meta != nil {
|
response := streamResponseClaude2OpenAI(&claudeResponse)
|
||||||
usage.PromptTokens += meta.Usage.InputTokens
|
response.Id = responseId
|
||||||
usage.CompletionTokens += meta.Usage.OutputTokens
|
|
||||||
modelName = meta.Model
|
|
||||||
id = fmt.Sprintf("chatcmpl-%s", meta.Id)
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
if response == nil {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
response.Id = id
|
|
||||||
response.Model = modelName
|
|
||||||
response.Created = createdTime
|
response.Created = createdTime
|
||||||
jsonStr, err := json.Marshal(response)
|
jsonStr, err := json.Marshal(response)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -224,11 +146,14 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
|
|||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
_ = resp.Body.Close()
|
err := resp.Body.Close()
|
||||||
return nil, &usage
|
if err != nil {
|
||||||
|
return openai.ErrorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), ""
|
||||||
|
}
|
||||||
|
return nil, responseText
|
||||||
}
|
}
|
||||||
|
|
||||||
func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName string) (*model.ErrorWithStatusCode, *model.Usage) {
|
func Handler(c *gin.Context, resp *http.Response, promptTokens int, model string) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
responseBody, err := io.ReadAll(resp.Body)
|
responseBody, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return openai.ErrorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
|
||||||
@@ -243,8 +168,8 @@ func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName st
|
|||||||
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
if claudeResponse.Error.Type != "" {
|
if claudeResponse.Error.Type != "" {
|
||||||
return &model.ErrorWithStatusCode{
|
return &openai.ErrorWithStatusCode{
|
||||||
Error: model.Error{
|
Error: openai.Error{
|
||||||
Message: claudeResponse.Error.Message,
|
Message: claudeResponse.Error.Message,
|
||||||
Type: claudeResponse.Error.Type,
|
Type: claudeResponse.Error.Type,
|
||||||
Param: "",
|
Param: "",
|
||||||
@@ -254,11 +179,12 @@ func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName st
|
|||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
fullTextResponse := responseClaude2OpenAI(&claudeResponse)
|
fullTextResponse := responseClaude2OpenAI(&claudeResponse)
|
||||||
fullTextResponse.Model = modelName
|
fullTextResponse.Model = model
|
||||||
usage := model.Usage{
|
completionTokens := openai.CountTokenText(claudeResponse.Completion, model)
|
||||||
PromptTokens: claudeResponse.Usage.InputTokens,
|
usage := openai.Usage{
|
||||||
CompletionTokens: claudeResponse.Usage.OutputTokens,
|
PromptTokens: promptTokens,
|
||||||
TotalTokens: claudeResponse.Usage.InputTokens + claudeResponse.Usage.OutputTokens,
|
CompletionTokens: completionTokens,
|
||||||
|
TotalTokens: promptTokens + completionTokens,
|
||||||
}
|
}
|
||||||
fullTextResponse.Usage = usage
|
fullTextResponse.Usage = usage
|
||||||
jsonResponse, err := json.Marshal(fullTextResponse)
|
jsonResponse, err := json.Marshal(fullTextResponse)
|
||||||
|
|||||||
@@ -1,44 +1,19 @@
|
|||||||
package anthropic
|
package anthropic
|
||||||
|
|
||||||
// https://docs.anthropic.com/claude/reference/messages_post
|
|
||||||
|
|
||||||
type Metadata struct {
|
type Metadata struct {
|
||||||
UserId string `json:"user_id"`
|
UserId string `json:"user_id"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type ImageSource struct {
|
|
||||||
Type string `json:"type"`
|
|
||||||
MediaType string `json:"media_type"`
|
|
||||||
Data string `json:"data"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type Content struct {
|
|
||||||
Type string `json:"type"`
|
|
||||||
Text string `json:"text,omitempty"`
|
|
||||||
Source *ImageSource `json:"source,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type Message struct {
|
|
||||||
Role string `json:"role"`
|
|
||||||
Content []Content `json:"content"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type Request struct {
|
type Request struct {
|
||||||
Model string `json:"model"`
|
Model string `json:"model"`
|
||||||
Messages []Message `json:"messages"`
|
Prompt string `json:"prompt"`
|
||||||
System string `json:"system,omitempty"`
|
MaxTokensToSample int `json:"max_tokens_to_sample"`
|
||||||
MaxTokens int `json:"max_tokens,omitempty"`
|
StopSequences []string `json:"stop_sequences,omitempty"`
|
||||||
StopSequences []string `json:"stop_sequences,omitempty"`
|
Temperature float64 `json:"temperature,omitempty"`
|
||||||
Stream bool `json:"stream,omitempty"`
|
TopP float64 `json:"top_p,omitempty"`
|
||||||
Temperature float64 `json:"temperature,omitempty"`
|
TopK int `json:"top_k,omitempty"`
|
||||||
TopP float64 `json:"top_p,omitempty"`
|
|
||||||
TopK int `json:"top_k,omitempty"`
|
|
||||||
//Metadata `json:"metadata,omitempty"`
|
//Metadata `json:"metadata,omitempty"`
|
||||||
}
|
Stream bool `json:"stream,omitempty"`
|
||||||
|
|
||||||
type Usage struct {
|
|
||||||
InputTokens int `json:"input_tokens"`
|
|
||||||
OutputTokens int `json:"output_tokens"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type Error struct {
|
type Error struct {
|
||||||
@@ -47,29 +22,8 @@ type Error struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type Response struct {
|
type Response struct {
|
||||||
Id string `json:"id"`
|
Completion string `json:"completion"`
|
||||||
Type string `json:"type"`
|
StopReason string `json:"stop_reason"`
|
||||||
Role string `json:"role"`
|
Model string `json:"model"`
|
||||||
Content []Content `json:"content"`
|
Error Error `json:"error"`
|
||||||
Model string `json:"model"`
|
|
||||||
StopReason *string `json:"stop_reason"`
|
|
||||||
StopSequence *string `json:"stop_sequence"`
|
|
||||||
Usage Usage `json:"usage"`
|
|
||||||
Error Error `json:"error"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type Delta struct {
|
|
||||||
Type string `json:"type"`
|
|
||||||
Text string `json:"text"`
|
|
||||||
StopReason *string `json:"stop_reason"`
|
|
||||||
StopSequence *string `json:"stop_sequence"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type StreamResponse struct {
|
|
||||||
Type string `json:"type"`
|
|
||||||
Message *Response `json:"message"`
|
|
||||||
Index int `json:"index"`
|
|
||||||
ContentBlock *Content `json:"content_block"`
|
|
||||||
Delta *Delta `json:"delta"`
|
|
||||||
Usage *Usage `json:"usage"`
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +0,0 @@
|
|||||||
package baichuan
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"Baichuan2-Turbo",
|
|
||||||
"Baichuan2-Turbo-192k",
|
|
||||||
"Baichuan-Text-Embedding",
|
|
||||||
}
|
|
||||||
@@ -1,105 +1,22 @@
|
|||||||
package baidu
|
package baidu
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/relay/channel"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
"strings"
|
"one-api/relay/channel/openai"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Adaptor struct {
|
type Adaptor struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) Init(meta *util.RelayMeta) {
|
func (a *Adaptor) Auth(c *gin.Context) error {
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetRequestURL(meta *util.RelayMeta) (string, error) {
|
|
||||||
// https://cloud.baidu.com/doc/WENXINWORKSHOP/s/clntwmv7t
|
|
||||||
suffix := "chat/"
|
|
||||||
if strings.HasPrefix("Embedding", meta.ActualModelName) {
|
|
||||||
suffix = "embeddings/"
|
|
||||||
}
|
|
||||||
switch meta.ActualModelName {
|
|
||||||
case "ERNIE-4.0":
|
|
||||||
suffix += "completions_pro"
|
|
||||||
case "ERNIE-Bot-4":
|
|
||||||
suffix += "completions_pro"
|
|
||||||
case "ERNIE-3.5-8K":
|
|
||||||
suffix += "completions"
|
|
||||||
case "ERNIE-Bot-8K":
|
|
||||||
suffix += "ernie_bot_8k"
|
|
||||||
case "ERNIE-Bot":
|
|
||||||
suffix += "completions"
|
|
||||||
case "ERNIE-Speed":
|
|
||||||
suffix += "ernie_speed"
|
|
||||||
case "ERNIE-Bot-turbo":
|
|
||||||
suffix += "eb-instant"
|
|
||||||
case "BLOOMZ-7B":
|
|
||||||
suffix += "bloomz_7b1"
|
|
||||||
case "Embedding-V1":
|
|
||||||
suffix += "embedding-v1"
|
|
||||||
default:
|
|
||||||
suffix += meta.ActualModelName
|
|
||||||
}
|
|
||||||
fullRequestURL := fmt.Sprintf("%s/rpc/2.0/ai_custom/v1/wenxinworkshop/%s", meta.BaseURL, suffix)
|
|
||||||
var accessToken string
|
|
||||||
var err error
|
|
||||||
if accessToken, err = GetAccessToken(meta.APIKey); err != nil {
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
fullRequestURL += "?access_token=" + accessToken
|
|
||||||
return fullRequestURL, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) SetupRequestHeader(c *gin.Context, req *http.Request, meta *util.RelayMeta) error {
|
|
||||||
channel.SetupCommonRequestHeader(c, req, meta)
|
|
||||||
req.Header.Set("Authorization", "Bearer "+meta.APIKey)
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) ConvertRequest(c *gin.Context, relayMode int, request *model.GeneralOpenAIRequest) (any, error) {
|
func (a *Adaptor) ConvertRequest(request *openai.GeneralOpenAIRequest) (any, error) {
|
||||||
if request == nil {
|
return nil, nil
|
||||||
return nil, errors.New("request is nil")
|
|
||||||
}
|
|
||||||
switch relayMode {
|
|
||||||
case constant.RelayModeEmbeddings:
|
|
||||||
baiduEmbeddingRequest := ConvertEmbeddingRequest(*request)
|
|
||||||
return baiduEmbeddingRequest, nil
|
|
||||||
default:
|
|
||||||
baiduRequest := ConvertRequest(*request)
|
|
||||||
return baiduRequest, nil
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) DoRequest(c *gin.Context, meta *util.RelayMeta, requestBody io.Reader) (*http.Response, error) {
|
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage, error) {
|
||||||
return channel.DoRequestHelper(a, c, meta, requestBody)
|
return nil, nil, nil
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, meta *util.RelayMeta) (usage *model.Usage, err *model.ErrorWithStatusCode) {
|
|
||||||
if meta.IsStream {
|
|
||||||
err, usage = StreamHandler(c, resp)
|
|
||||||
} else {
|
|
||||||
switch meta.Mode {
|
|
||||||
case constant.RelayModeEmbeddings:
|
|
||||||
err, usage = EmbeddingHandler(c, resp)
|
|
||||||
default:
|
|
||||||
err, usage = Handler(c, resp)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetModelList() []string {
|
|
||||||
return ModelList
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetChannelName() string {
|
|
||||||
return "baidu"
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,10 +0,0 @@
|
|||||||
package baidu
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"ERNIE-Bot-4",
|
|
||||||
"ERNIE-Bot-8K",
|
|
||||||
"ERNIE-Bot",
|
|
||||||
"ERNIE-Speed",
|
|
||||||
"ERNIE-Bot-turbo",
|
|
||||||
"Embedding-V1",
|
|
||||||
}
|
|
||||||
@@ -6,14 +6,13 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
|
"one-api/relay/constant"
|
||||||
|
"one-api/relay/util"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
@@ -44,7 +43,7 @@ type Error struct {
|
|||||||
|
|
||||||
var baiduTokenStore sync.Map
|
var baiduTokenStore sync.Map
|
||||||
|
|
||||||
func ConvertRequest(request model.GeneralOpenAIRequest) *ChatRequest {
|
func ConvertRequest(request openai.GeneralOpenAIRequest) *ChatRequest {
|
||||||
messages := make([]Message, 0, len(request.Messages))
|
messages := make([]Message, 0, len(request.Messages))
|
||||||
for _, message := range request.Messages {
|
for _, message := range request.Messages {
|
||||||
if message.Role == "system" {
|
if message.Role == "system" {
|
||||||
@@ -72,7 +71,7 @@ func ConvertRequest(request model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
func responseBaidu2OpenAI(response *ChatResponse) *openai.TextResponse {
|
func responseBaidu2OpenAI(response *ChatResponse) *openai.TextResponse {
|
||||||
choice := openai.TextResponseChoice{
|
choice := openai.TextResponseChoice{
|
||||||
Index: 0,
|
Index: 0,
|
||||||
Message: model.Message{
|
Message: openai.Message{
|
||||||
Role: "assistant",
|
Role: "assistant",
|
||||||
Content: response.Result,
|
Content: response.Result,
|
||||||
},
|
},
|
||||||
@@ -104,7 +103,7 @@ func streamResponseBaidu2OpenAI(baiduResponse *ChatStreamResponse) *openai.ChatC
|
|||||||
return &response
|
return &response
|
||||||
}
|
}
|
||||||
|
|
||||||
func ConvertEmbeddingRequest(request model.GeneralOpenAIRequest) *EmbeddingRequest {
|
func ConvertEmbeddingRequest(request openai.GeneralOpenAIRequest) *EmbeddingRequest {
|
||||||
return &EmbeddingRequest{
|
return &EmbeddingRequest{
|
||||||
Input: request.ParseInput(),
|
Input: request.ParseInput(),
|
||||||
}
|
}
|
||||||
@@ -127,8 +126,8 @@ func embeddingResponseBaidu2OpenAI(response *EmbeddingResponse) *openai.Embeddin
|
|||||||
return &openAIEmbeddingResponse
|
return &openAIEmbeddingResponse
|
||||||
}
|
}
|
||||||
|
|
||||||
func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func StreamHandler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
var usage model.Usage
|
var usage openai.Usage
|
||||||
scanner := bufio.NewScanner(resp.Body)
|
scanner := bufio.NewScanner(resp.Body)
|
||||||
scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
|
scanner.Split(func(data []byte, atEOF bool) (advance int, token []byte, err error) {
|
||||||
if atEOF && len(data) == 0 {
|
if atEOF && len(data) == 0 {
|
||||||
@@ -190,7 +189,7 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
|
|||||||
return nil, &usage
|
return nil, &usage
|
||||||
}
|
}
|
||||||
|
|
||||||
func Handler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func Handler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
var baiduResponse ChatResponse
|
var baiduResponse ChatResponse
|
||||||
responseBody, err := io.ReadAll(resp.Body)
|
responseBody, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -205,8 +204,8 @@ func Handler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *
|
|||||||
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
if baiduResponse.ErrorMsg != "" {
|
if baiduResponse.ErrorMsg != "" {
|
||||||
return &model.ErrorWithStatusCode{
|
return &openai.ErrorWithStatusCode{
|
||||||
Error: model.Error{
|
Error: openai.Error{
|
||||||
Message: baiduResponse.ErrorMsg,
|
Message: baiduResponse.ErrorMsg,
|
||||||
Type: "baidu_error",
|
Type: "baidu_error",
|
||||||
Param: "",
|
Param: "",
|
||||||
@@ -227,7 +226,7 @@ func Handler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *
|
|||||||
return nil, &fullTextResponse.Usage
|
return nil, &fullTextResponse.Usage
|
||||||
}
|
}
|
||||||
|
|
||||||
func EmbeddingHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, *model.Usage) {
|
func EmbeddingHandler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
var baiduResponse EmbeddingResponse
|
var baiduResponse EmbeddingResponse
|
||||||
responseBody, err := io.ReadAll(resp.Body)
|
responseBody, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -242,8 +241,8 @@ func EmbeddingHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStat
|
|||||||
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
if baiduResponse.ErrorMsg != "" {
|
if baiduResponse.ErrorMsg != "" {
|
||||||
return &model.ErrorWithStatusCode{
|
return &openai.ErrorWithStatusCode{
|
||||||
Error: model.Error{
|
Error: openai.Error{
|
||||||
Message: baiduResponse.ErrorMsg,
|
Message: baiduResponse.ErrorMsg,
|
||||||
Type: "baidu_error",
|
Type: "baidu_error",
|
||||||
Param: "",
|
Param: "",
|
||||||
|
|||||||
@@ -1,18 +1,18 @@
|
|||||||
package baidu
|
package baidu
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
"one-api/relay/channel/openai"
|
||||||
"time"
|
"time"
|
||||||
)
|
)
|
||||||
|
|
||||||
type ChatResponse struct {
|
type ChatResponse struct {
|
||||||
Id string `json:"id"`
|
Id string `json:"id"`
|
||||||
Object string `json:"object"`
|
Object string `json:"object"`
|
||||||
Created int64 `json:"created"`
|
Created int64 `json:"created"`
|
||||||
Result string `json:"result"`
|
Result string `json:"result"`
|
||||||
IsTruncated bool `json:"is_truncated"`
|
IsTruncated bool `json:"is_truncated"`
|
||||||
NeedClearHistory bool `json:"need_clear_history"`
|
NeedClearHistory bool `json:"need_clear_history"`
|
||||||
Usage model.Usage `json:"usage"`
|
Usage openai.Usage `json:"usage"`
|
||||||
Error
|
Error
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -37,7 +37,7 @@ type EmbeddingResponse struct {
|
|||||||
Object string `json:"object"`
|
Object string `json:"object"`
|
||||||
Created int64 `json:"created"`
|
Created int64 `json:"created"`
|
||||||
Data []EmbeddingData `json:"data"`
|
Data []EmbeddingData `json:"data"`
|
||||||
Usage model.Usage `json:"usage"`
|
Usage openai.Usage `json:"usage"`
|
||||||
Error
|
Error
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,51 +0,0 @@
|
|||||||
package channel
|
|
||||||
|
|
||||||
import (
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"github.com/gin-gonic/gin"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
|
||||||
)
|
|
||||||
|
|
||||||
func SetupCommonRequestHeader(c *gin.Context, req *http.Request, meta *util.RelayMeta) {
|
|
||||||
req.Header.Set("Content-Type", c.Request.Header.Get("Content-Type"))
|
|
||||||
req.Header.Set("Accept", c.Request.Header.Get("Accept"))
|
|
||||||
if meta.IsStream && c.Request.Header.Get("Accept") == "" {
|
|
||||||
req.Header.Set("Accept", "text/event-stream")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func DoRequestHelper(a Adaptor, c *gin.Context, meta *util.RelayMeta, requestBody io.Reader) (*http.Response, error) {
|
|
||||||
fullRequestURL, err := a.GetRequestURL(meta)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("get request url failed: %w", err)
|
|
||||||
}
|
|
||||||
req, err := http.NewRequest(c.Request.Method, fullRequestURL, requestBody)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("new request failed: %w", err)
|
|
||||||
}
|
|
||||||
err = a.SetupRequestHeader(c, req, meta)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("setup request header failed: %w", err)
|
|
||||||
}
|
|
||||||
resp, err := DoRequest(c, req)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("do request failed: %w", err)
|
|
||||||
}
|
|
||||||
return resp, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func DoRequest(c *gin.Context, req *http.Request) (*http.Response, error) {
|
|
||||||
resp, err := util.HTTPClient.Do(req)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
if resp == nil {
|
|
||||||
return nil, errors.New("resp is nil")
|
|
||||||
}
|
|
||||||
_ = req.Body.Close()
|
|
||||||
_ = c.Request.Body.Close()
|
|
||||||
return resp, nil
|
|
||||||
}
|
|
||||||
@@ -1,66 +0,0 @@
|
|||||||
package gemini
|
|
||||||
|
|
||||||
import (
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"github.com/gin-gonic/gin"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
channelhelper "github.com/songquanpeng/one-api/relay/channel"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
|
||||||
)
|
|
||||||
|
|
||||||
type Adaptor struct {
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) Init(meta *util.RelayMeta) {
|
|
||||||
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetRequestURL(meta *util.RelayMeta) (string, error) {
|
|
||||||
version := helper.AssignOrDefault(meta.APIVersion, "v1")
|
|
||||||
action := "generateContent"
|
|
||||||
if meta.IsStream {
|
|
||||||
action = "streamGenerateContent"
|
|
||||||
}
|
|
||||||
return fmt.Sprintf("%s/%s/models/%s:%s", meta.BaseURL, version, meta.ActualModelName, action), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) SetupRequestHeader(c *gin.Context, req *http.Request, meta *util.RelayMeta) error {
|
|
||||||
channelhelper.SetupCommonRequestHeader(c, req, meta)
|
|
||||||
req.Header.Set("x-goog-api-key", meta.APIKey)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) ConvertRequest(c *gin.Context, relayMode int, request *model.GeneralOpenAIRequest) (any, error) {
|
|
||||||
if request == nil {
|
|
||||||
return nil, errors.New("request is nil")
|
|
||||||
}
|
|
||||||
return ConvertRequest(*request), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) DoRequest(c *gin.Context, meta *util.RelayMeta, requestBody io.Reader) (*http.Response, error) {
|
|
||||||
return channelhelper.DoRequestHelper(a, c, meta, requestBody)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, meta *util.RelayMeta) (usage *model.Usage, err *model.ErrorWithStatusCode) {
|
|
||||||
if meta.IsStream {
|
|
||||||
var responseText string
|
|
||||||
err, responseText = StreamHandler(c, resp)
|
|
||||||
usage = openai.ResponseText2Usage(responseText, meta.ActualModelName, meta.PromptTokens)
|
|
||||||
} else {
|
|
||||||
err, usage = Handler(c, resp, meta.PromptTokens, meta.ActualModelName)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetModelList() []string {
|
|
||||||
return ModelList
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetChannelName() string {
|
|
||||||
return "google gemini"
|
|
||||||
}
|
|
||||||
@@ -1,6 +0,0 @@
|
|||||||
package gemini
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"gemini-pro", "gemini-1.0-pro-001",
|
|
||||||
"gemini-pro-vision", "gemini-1.0-pro-vision-001",
|
|
||||||
}
|
|
||||||
@@ -1,41 +0,0 @@
|
|||||||
package gemini
|
|
||||||
|
|
||||||
type ChatRequest struct {
|
|
||||||
Contents []ChatContent `json:"contents"`
|
|
||||||
SafetySettings []ChatSafetySettings `json:"safety_settings,omitempty"`
|
|
||||||
GenerationConfig ChatGenerationConfig `json:"generation_config,omitempty"`
|
|
||||||
Tools []ChatTools `json:"tools,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type InlineData struct {
|
|
||||||
MimeType string `json:"mimeType"`
|
|
||||||
Data string `json:"data"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type Part struct {
|
|
||||||
Text string `json:"text,omitempty"`
|
|
||||||
InlineData *InlineData `json:"inlineData,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type ChatContent struct {
|
|
||||||
Role string `json:"role,omitempty"`
|
|
||||||
Parts []Part `json:"parts"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type ChatSafetySettings struct {
|
|
||||||
Category string `json:"category"`
|
|
||||||
Threshold string `json:"threshold"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type ChatTools struct {
|
|
||||||
FunctionDeclarations any `json:"functionDeclarations,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type ChatGenerationConfig struct {
|
|
||||||
Temperature float64 `json:"temperature,omitempty"`
|
|
||||||
TopP float64 `json:"topP,omitempty"`
|
|
||||||
TopK float64 `json:"topK,omitempty"`
|
|
||||||
MaxOutputTokens int `json:"maxOutputTokens,omitempty"`
|
|
||||||
CandidateCount int `json:"candidateCount,omitempty"`
|
|
||||||
StopSequences []string `json:"stopSequences,omitempty"`
|
|
||||||
}
|
|
||||||
22
relay/channel/google/adaptor.go
Normal file
22
relay/channel/google/adaptor.go
Normal file
@@ -0,0 +1,22 @@
|
|||||||
|
package google
|
||||||
|
|
||||||
|
import (
|
||||||
|
"github.com/gin-gonic/gin"
|
||||||
|
"net/http"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
|
)
|
||||||
|
|
||||||
|
type Adaptor struct {
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a *Adaptor) Auth(c *gin.Context) error {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a *Adaptor) ConvertRequest(request *openai.GeneralOpenAIRequest) (any, error) {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage, error) {
|
||||||
|
return nil, nil, nil
|
||||||
|
}
|
||||||
@@ -1,19 +1,18 @@
|
|||||||
package gemini
|
package google
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bufio"
|
"bufio"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/config"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/image"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/config"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/image"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
|
"one-api/relay/constant"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
@@ -22,14 +21,14 @@ import (
|
|||||||
// https://ai.google.dev/docs/gemini_api_overview?hl=zh-cn
|
// https://ai.google.dev/docs/gemini_api_overview?hl=zh-cn
|
||||||
|
|
||||||
const (
|
const (
|
||||||
VisionMaxImageNum = 16
|
GeminiVisionMaxImageNum = 16
|
||||||
)
|
)
|
||||||
|
|
||||||
// Setting safety to the lowest possible values since Gemini is already powerless enough
|
// Setting safety to the lowest possible values since Gemini is already powerless enough
|
||||||
func ConvertRequest(textRequest model.GeneralOpenAIRequest) *ChatRequest {
|
func ConvertGeminiRequest(textRequest openai.GeneralOpenAIRequest) *GeminiChatRequest {
|
||||||
geminiRequest := ChatRequest{
|
geminiRequest := GeminiChatRequest{
|
||||||
Contents: make([]ChatContent, 0, len(textRequest.Messages)),
|
Contents: make([]GeminiChatContent, 0, len(textRequest.Messages)),
|
||||||
SafetySettings: []ChatSafetySettings{
|
SafetySettings: []GeminiChatSafetySettings{
|
||||||
{
|
{
|
||||||
Category: "HARM_CATEGORY_HARASSMENT",
|
Category: "HARM_CATEGORY_HARASSMENT",
|
||||||
Threshold: config.GeminiSafetySetting,
|
Threshold: config.GeminiSafetySetting,
|
||||||
@@ -47,14 +46,14 @@ func ConvertRequest(textRequest model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
Threshold: config.GeminiSafetySetting,
|
Threshold: config.GeminiSafetySetting,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
GenerationConfig: ChatGenerationConfig{
|
GenerationConfig: GeminiChatGenerationConfig{
|
||||||
Temperature: textRequest.Temperature,
|
Temperature: textRequest.Temperature,
|
||||||
TopP: textRequest.TopP,
|
TopP: textRequest.TopP,
|
||||||
MaxOutputTokens: textRequest.MaxTokens,
|
MaxOutputTokens: textRequest.MaxTokens,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
if textRequest.Functions != nil {
|
if textRequest.Functions != nil {
|
||||||
geminiRequest.Tools = []ChatTools{
|
geminiRequest.Tools = []GeminiChatTools{
|
||||||
{
|
{
|
||||||
FunctionDeclarations: textRequest.Functions,
|
FunctionDeclarations: textRequest.Functions,
|
||||||
},
|
},
|
||||||
@@ -62,30 +61,30 @@ func ConvertRequest(textRequest model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
}
|
}
|
||||||
shouldAddDummyModelMessage := false
|
shouldAddDummyModelMessage := false
|
||||||
for _, message := range textRequest.Messages {
|
for _, message := range textRequest.Messages {
|
||||||
content := ChatContent{
|
content := GeminiChatContent{
|
||||||
Role: message.Role,
|
Role: message.Role,
|
||||||
Parts: []Part{
|
Parts: []GeminiPart{
|
||||||
{
|
{
|
||||||
Text: message.StringContent(),
|
Text: message.StringContent(),
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
openaiContent := message.ParseContent()
|
openaiContent := message.ParseContent()
|
||||||
var parts []Part
|
var parts []GeminiPart
|
||||||
imageNum := 0
|
imageNum := 0
|
||||||
for _, part := range openaiContent {
|
for _, part := range openaiContent {
|
||||||
if part.Type == model.ContentTypeText {
|
if part.Type == openai.ContentTypeText {
|
||||||
parts = append(parts, Part{
|
parts = append(parts, GeminiPart{
|
||||||
Text: part.Text,
|
Text: part.Text,
|
||||||
})
|
})
|
||||||
} else if part.Type == model.ContentTypeImageURL {
|
} else if part.Type == openai.ContentTypeImageURL {
|
||||||
imageNum += 1
|
imageNum += 1
|
||||||
if imageNum > VisionMaxImageNum {
|
if imageNum > GeminiVisionMaxImageNum {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
mimeType, data, _ := image.GetImageFromUrl(part.ImageURL.Url)
|
mimeType, data, _ := image.GetImageFromUrl(part.ImageURL.Url)
|
||||||
parts = append(parts, Part{
|
parts = append(parts, GeminiPart{
|
||||||
InlineData: &InlineData{
|
InlineData: &GeminiInlineData{
|
||||||
MimeType: mimeType,
|
MimeType: mimeType,
|
||||||
Data: data,
|
Data: data,
|
||||||
},
|
},
|
||||||
@@ -107,9 +106,9 @@ func ConvertRequest(textRequest model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
|
|
||||||
// If a system message is the last message, we need to add a dummy model message to make gemini happy
|
// If a system message is the last message, we need to add a dummy model message to make gemini happy
|
||||||
if shouldAddDummyModelMessage {
|
if shouldAddDummyModelMessage {
|
||||||
geminiRequest.Contents = append(geminiRequest.Contents, ChatContent{
|
geminiRequest.Contents = append(geminiRequest.Contents, GeminiChatContent{
|
||||||
Role: "model",
|
Role: "model",
|
||||||
Parts: []Part{
|
Parts: []GeminiPart{
|
||||||
{
|
{
|
||||||
Text: "Okay",
|
Text: "Okay",
|
||||||
},
|
},
|
||||||
@@ -122,12 +121,12 @@ func ConvertRequest(textRequest model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
return &geminiRequest
|
return &geminiRequest
|
||||||
}
|
}
|
||||||
|
|
||||||
type ChatResponse struct {
|
type GeminiChatResponse struct {
|
||||||
Candidates []ChatCandidate `json:"candidates"`
|
Candidates []GeminiChatCandidate `json:"candidates"`
|
||||||
PromptFeedback ChatPromptFeedback `json:"promptFeedback"`
|
PromptFeedback GeminiChatPromptFeedback `json:"promptFeedback"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func (g *ChatResponse) GetResponseText() string {
|
func (g *GeminiChatResponse) GetResponseText() string {
|
||||||
if g == nil {
|
if g == nil {
|
||||||
return ""
|
return ""
|
||||||
}
|
}
|
||||||
@@ -137,23 +136,23 @@ func (g *ChatResponse) GetResponseText() string {
|
|||||||
return ""
|
return ""
|
||||||
}
|
}
|
||||||
|
|
||||||
type ChatCandidate struct {
|
type GeminiChatCandidate struct {
|
||||||
Content ChatContent `json:"content"`
|
Content GeminiChatContent `json:"content"`
|
||||||
FinishReason string `json:"finishReason"`
|
FinishReason string `json:"finishReason"`
|
||||||
Index int64 `json:"index"`
|
Index int64 `json:"index"`
|
||||||
SafetyRatings []ChatSafetyRating `json:"safetyRatings"`
|
SafetyRatings []GeminiChatSafetyRating `json:"safetyRatings"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type ChatSafetyRating struct {
|
type GeminiChatSafetyRating struct {
|
||||||
Category string `json:"category"`
|
Category string `json:"category"`
|
||||||
Probability string `json:"probability"`
|
Probability string `json:"probability"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type ChatPromptFeedback struct {
|
type GeminiChatPromptFeedback struct {
|
||||||
SafetyRatings []ChatSafetyRating `json:"safetyRatings"`
|
SafetyRatings []GeminiChatSafetyRating `json:"safetyRatings"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func responseGeminiChat2OpenAI(response *ChatResponse) *openai.TextResponse {
|
func responseGeminiChat2OpenAI(response *GeminiChatResponse) *openai.TextResponse {
|
||||||
fullTextResponse := openai.TextResponse{
|
fullTextResponse := openai.TextResponse{
|
||||||
Id: fmt.Sprintf("chatcmpl-%s", helper.GetUUID()),
|
Id: fmt.Sprintf("chatcmpl-%s", helper.GetUUID()),
|
||||||
Object: "chat.completion",
|
Object: "chat.completion",
|
||||||
@@ -163,7 +162,7 @@ func responseGeminiChat2OpenAI(response *ChatResponse) *openai.TextResponse {
|
|||||||
for i, candidate := range response.Candidates {
|
for i, candidate := range response.Candidates {
|
||||||
choice := openai.TextResponseChoice{
|
choice := openai.TextResponseChoice{
|
||||||
Index: i,
|
Index: i,
|
||||||
Message: model.Message{
|
Message: openai.Message{
|
||||||
Role: "assistant",
|
Role: "assistant",
|
||||||
Content: "",
|
Content: "",
|
||||||
},
|
},
|
||||||
@@ -177,7 +176,7 @@ func responseGeminiChat2OpenAI(response *ChatResponse) *openai.TextResponse {
|
|||||||
return &fullTextResponse
|
return &fullTextResponse
|
||||||
}
|
}
|
||||||
|
|
||||||
func streamResponseGeminiChat2OpenAI(geminiResponse *ChatResponse) *openai.ChatCompletionsStreamResponse {
|
func streamResponseGeminiChat2OpenAI(geminiResponse *GeminiChatResponse) *openai.ChatCompletionsStreamResponse {
|
||||||
var choice openai.ChatCompletionsStreamResponseChoice
|
var choice openai.ChatCompletionsStreamResponseChoice
|
||||||
choice.Delta.Content = geminiResponse.GetResponseText()
|
choice.Delta.Content = geminiResponse.GetResponseText()
|
||||||
choice.FinishReason = &constant.StopFinishReason
|
choice.FinishReason = &constant.StopFinishReason
|
||||||
@@ -188,7 +187,7 @@ func streamResponseGeminiChat2OpenAI(geminiResponse *ChatResponse) *openai.ChatC
|
|||||||
return &response
|
return &response
|
||||||
}
|
}
|
||||||
|
|
||||||
func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, string) {
|
func StreamHandler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, string) {
|
||||||
responseText := ""
|
responseText := ""
|
||||||
dataChan := make(chan string)
|
dataChan := make(chan string)
|
||||||
stopChan := make(chan bool)
|
stopChan := make(chan bool)
|
||||||
@@ -258,7 +257,7 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
|
|||||||
return nil, responseText
|
return nil, responseText
|
||||||
}
|
}
|
||||||
|
|
||||||
func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName string) (*model.ErrorWithStatusCode, *model.Usage) {
|
func GeminiHandler(c *gin.Context, resp *http.Response, promptTokens int, model string) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
responseBody, err := io.ReadAll(resp.Body)
|
responseBody, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return openai.ErrorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
|
||||||
@@ -267,14 +266,14 @@ func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName st
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return openai.ErrorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
var geminiResponse ChatResponse
|
var geminiResponse GeminiChatResponse
|
||||||
err = json.Unmarshal(responseBody, &geminiResponse)
|
err = json.Unmarshal(responseBody, &geminiResponse)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
if len(geminiResponse.Candidates) == 0 {
|
if len(geminiResponse.Candidates) == 0 {
|
||||||
return &model.ErrorWithStatusCode{
|
return &openai.ErrorWithStatusCode{
|
||||||
Error: model.Error{
|
Error: openai.Error{
|
||||||
Message: "No candidates returned",
|
Message: "No candidates returned",
|
||||||
Type: "server_error",
|
Type: "server_error",
|
||||||
Param: "",
|
Param: "",
|
||||||
@@ -284,9 +283,9 @@ func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName st
|
|||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
fullTextResponse := responseGeminiChat2OpenAI(&geminiResponse)
|
fullTextResponse := responseGeminiChat2OpenAI(&geminiResponse)
|
||||||
fullTextResponse.Model = modelName
|
fullTextResponse.Model = model
|
||||||
completionTokens := openai.CountTokenText(geminiResponse.GetResponseText(), modelName)
|
completionTokens := openai.CountTokenText(geminiResponse.GetResponseText(), model)
|
||||||
usage := model.Usage{
|
usage := openai.Usage{
|
||||||
PromptTokens: promptTokens,
|
PromptTokens: promptTokens,
|
||||||
CompletionTokens: completionTokens,
|
CompletionTokens: completionTokens,
|
||||||
TotalTokens: promptTokens + completionTokens,
|
TotalTokens: promptTokens + completionTokens,
|
||||||
80
relay/channel/google/model.go
Normal file
80
relay/channel/google/model.go
Normal file
@@ -0,0 +1,80 @@
|
|||||||
|
package google
|
||||||
|
|
||||||
|
import (
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
|
)
|
||||||
|
|
||||||
|
type GeminiChatRequest struct {
|
||||||
|
Contents []GeminiChatContent `json:"contents"`
|
||||||
|
SafetySettings []GeminiChatSafetySettings `json:"safety_settings,omitempty"`
|
||||||
|
GenerationConfig GeminiChatGenerationConfig `json:"generation_config,omitempty"`
|
||||||
|
Tools []GeminiChatTools `json:"tools,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type GeminiInlineData struct {
|
||||||
|
MimeType string `json:"mimeType"`
|
||||||
|
Data string `json:"data"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type GeminiPart struct {
|
||||||
|
Text string `json:"text,omitempty"`
|
||||||
|
InlineData *GeminiInlineData `json:"inlineData,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type GeminiChatContent struct {
|
||||||
|
Role string `json:"role,omitempty"`
|
||||||
|
Parts []GeminiPart `json:"parts"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type GeminiChatSafetySettings struct {
|
||||||
|
Category string `json:"category"`
|
||||||
|
Threshold string `json:"threshold"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type GeminiChatTools struct {
|
||||||
|
FunctionDeclarations any `json:"functionDeclarations,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type GeminiChatGenerationConfig struct {
|
||||||
|
Temperature float64 `json:"temperature,omitempty"`
|
||||||
|
TopP float64 `json:"topP,omitempty"`
|
||||||
|
TopK float64 `json:"topK,omitempty"`
|
||||||
|
MaxOutputTokens int `json:"maxOutputTokens,omitempty"`
|
||||||
|
CandidateCount int `json:"candidateCount,omitempty"`
|
||||||
|
StopSequences []string `json:"stopSequences,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type PaLMChatMessage struct {
|
||||||
|
Author string `json:"author"`
|
||||||
|
Content string `json:"content"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type PaLMFilter struct {
|
||||||
|
Reason string `json:"reason"`
|
||||||
|
Message string `json:"message"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type PaLMPrompt struct {
|
||||||
|
Messages []PaLMChatMessage `json:"messages"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type PaLMChatRequest struct {
|
||||||
|
Prompt PaLMPrompt `json:"prompt"`
|
||||||
|
Temperature float64 `json:"temperature,omitempty"`
|
||||||
|
CandidateCount int `json:"candidateCount,omitempty"`
|
||||||
|
TopP float64 `json:"topP,omitempty"`
|
||||||
|
TopK int `json:"topK,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type PaLMError struct {
|
||||||
|
Code int `json:"code"`
|
||||||
|
Message string `json:"message"`
|
||||||
|
Status string `json:"status"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type PaLMChatResponse struct {
|
||||||
|
Candidates []PaLMChatMessage `json:"candidates"`
|
||||||
|
Messages []openai.Message `json:"messages"`
|
||||||
|
Filters []PaLMFilter `json:"filters"`
|
||||||
|
Error PaLMError `json:"error"`
|
||||||
|
}
|
||||||
@@ -1,26 +1,25 @@
|
|||||||
package palm
|
package google
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/common/helper"
|
|
||||||
"github.com/songquanpeng/one-api/common/logger"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/openai"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"io"
|
"io"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/common/helper"
|
||||||
|
"one-api/common/logger"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
|
"one-api/relay/constant"
|
||||||
)
|
)
|
||||||
|
|
||||||
// https://developers.generativeai.google/api/rest/generativelanguage/models/generateMessage#request-body
|
// https://developers.generativeai.google/api/rest/generativelanguage/models/generateMessage#request-body
|
||||||
// https://developers.generativeai.google/api/rest/generativelanguage/models/generateMessage#response-body
|
// https://developers.generativeai.google/api/rest/generativelanguage/models/generateMessage#response-body
|
||||||
|
|
||||||
func ConvertRequest(textRequest model.GeneralOpenAIRequest) *ChatRequest {
|
func ConvertPaLMRequest(textRequest openai.GeneralOpenAIRequest) *PaLMChatRequest {
|
||||||
palmRequest := ChatRequest{
|
palmRequest := PaLMChatRequest{
|
||||||
Prompt: Prompt{
|
Prompt: PaLMPrompt{
|
||||||
Messages: make([]ChatMessage, 0, len(textRequest.Messages)),
|
Messages: make([]PaLMChatMessage, 0, len(textRequest.Messages)),
|
||||||
},
|
},
|
||||||
Temperature: textRequest.Temperature,
|
Temperature: textRequest.Temperature,
|
||||||
CandidateCount: textRequest.N,
|
CandidateCount: textRequest.N,
|
||||||
@@ -28,7 +27,7 @@ func ConvertRequest(textRequest model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
TopK: textRequest.MaxTokens,
|
TopK: textRequest.MaxTokens,
|
||||||
}
|
}
|
||||||
for _, message := range textRequest.Messages {
|
for _, message := range textRequest.Messages {
|
||||||
palmMessage := ChatMessage{
|
palmMessage := PaLMChatMessage{
|
||||||
Content: message.StringContent(),
|
Content: message.StringContent(),
|
||||||
}
|
}
|
||||||
if message.Role == "user" {
|
if message.Role == "user" {
|
||||||
@@ -41,14 +40,14 @@ func ConvertRequest(textRequest model.GeneralOpenAIRequest) *ChatRequest {
|
|||||||
return &palmRequest
|
return &palmRequest
|
||||||
}
|
}
|
||||||
|
|
||||||
func responsePaLM2OpenAI(response *ChatResponse) *openai.TextResponse {
|
func responsePaLM2OpenAI(response *PaLMChatResponse) *openai.TextResponse {
|
||||||
fullTextResponse := openai.TextResponse{
|
fullTextResponse := openai.TextResponse{
|
||||||
Choices: make([]openai.TextResponseChoice, 0, len(response.Candidates)),
|
Choices: make([]openai.TextResponseChoice, 0, len(response.Candidates)),
|
||||||
}
|
}
|
||||||
for i, candidate := range response.Candidates {
|
for i, candidate := range response.Candidates {
|
||||||
choice := openai.TextResponseChoice{
|
choice := openai.TextResponseChoice{
|
||||||
Index: i,
|
Index: i,
|
||||||
Message: model.Message{
|
Message: openai.Message{
|
||||||
Role: "assistant",
|
Role: "assistant",
|
||||||
Content: candidate.Content,
|
Content: candidate.Content,
|
||||||
},
|
},
|
||||||
@@ -59,7 +58,7 @@ func responsePaLM2OpenAI(response *ChatResponse) *openai.TextResponse {
|
|||||||
return &fullTextResponse
|
return &fullTextResponse
|
||||||
}
|
}
|
||||||
|
|
||||||
func streamResponsePaLM2OpenAI(palmResponse *ChatResponse) *openai.ChatCompletionsStreamResponse {
|
func streamResponsePaLM2OpenAI(palmResponse *PaLMChatResponse) *openai.ChatCompletionsStreamResponse {
|
||||||
var choice openai.ChatCompletionsStreamResponseChoice
|
var choice openai.ChatCompletionsStreamResponseChoice
|
||||||
if len(palmResponse.Candidates) > 0 {
|
if len(palmResponse.Candidates) > 0 {
|
||||||
choice.Delta.Content = palmResponse.Candidates[0].Content
|
choice.Delta.Content = palmResponse.Candidates[0].Content
|
||||||
@@ -72,7 +71,7 @@ func streamResponsePaLM2OpenAI(palmResponse *ChatResponse) *openai.ChatCompletio
|
|||||||
return &response
|
return &response
|
||||||
}
|
}
|
||||||
|
|
||||||
func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusCode, string) {
|
func PaLMStreamHandler(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, string) {
|
||||||
responseText := ""
|
responseText := ""
|
||||||
responseId := fmt.Sprintf("chatcmpl-%s", helper.GetUUID())
|
responseId := fmt.Sprintf("chatcmpl-%s", helper.GetUUID())
|
||||||
createdTime := helper.GetTimestamp()
|
createdTime := helper.GetTimestamp()
|
||||||
@@ -91,7 +90,7 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
|
|||||||
stopChan <- true
|
stopChan <- true
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
var palmResponse ChatResponse
|
var palmResponse PaLMChatResponse
|
||||||
err = json.Unmarshal(responseBody, &palmResponse)
|
err = json.Unmarshal(responseBody, &palmResponse)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logger.SysError("error unmarshalling stream response: " + err.Error())
|
logger.SysError("error unmarshalling stream response: " + err.Error())
|
||||||
@@ -131,7 +130,7 @@ func StreamHandler(c *gin.Context, resp *http.Response) (*model.ErrorWithStatusC
|
|||||||
return nil, responseText
|
return nil, responseText
|
||||||
}
|
}
|
||||||
|
|
||||||
func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName string) (*model.ErrorWithStatusCode, *model.Usage) {
|
func PaLMHandler(c *gin.Context, resp *http.Response, promptTokens int, model string) (*openai.ErrorWithStatusCode, *openai.Usage) {
|
||||||
responseBody, err := io.ReadAll(resp.Body)
|
responseBody, err := io.ReadAll(resp.Body)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return openai.ErrorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "read_response_body_failed", http.StatusInternalServerError), nil
|
||||||
@@ -140,14 +139,14 @@ func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName st
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return openai.ErrorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "close_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
var palmResponse ChatResponse
|
var palmResponse PaLMChatResponse
|
||||||
err = json.Unmarshal(responseBody, &palmResponse)
|
err = json.Unmarshal(responseBody, &palmResponse)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
return openai.ErrorWrapper(err, "unmarshal_response_body_failed", http.StatusInternalServerError), nil
|
||||||
}
|
}
|
||||||
if palmResponse.Error.Code != 0 || len(palmResponse.Candidates) == 0 {
|
if palmResponse.Error.Code != 0 || len(palmResponse.Candidates) == 0 {
|
||||||
return &model.ErrorWithStatusCode{
|
return &openai.ErrorWithStatusCode{
|
||||||
Error: model.Error{
|
Error: openai.Error{
|
||||||
Message: palmResponse.Error.Message,
|
Message: palmResponse.Error.Message,
|
||||||
Type: palmResponse.Error.Status,
|
Type: palmResponse.Error.Status,
|
||||||
Param: "",
|
Param: "",
|
||||||
@@ -157,9 +156,9 @@ func Handler(c *gin.Context, resp *http.Response, promptTokens int, modelName st
|
|||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
fullTextResponse := responsePaLM2OpenAI(&palmResponse)
|
fullTextResponse := responsePaLM2OpenAI(&palmResponse)
|
||||||
fullTextResponse.Model = modelName
|
fullTextResponse.Model = model
|
||||||
completionTokens := openai.CountTokenText(palmResponse.Candidates[0].Content, modelName)
|
completionTokens := openai.CountTokenText(palmResponse.Candidates[0].Content, model)
|
||||||
usage := model.Usage{
|
usage := openai.Usage{
|
||||||
PromptTokens: promptTokens,
|
PromptTokens: promptTokens,
|
||||||
CompletionTokens: completionTokens,
|
CompletionTokens: completionTokens,
|
||||||
TotalTokens: promptTokens + completionTokens,
|
TotalTokens: promptTokens + completionTokens,
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
package groq
|
|
||||||
|
|
||||||
// https://console.groq.com/docs/models
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"gemma-7b-it",
|
|
||||||
"llama2-7b-2048",
|
|
||||||
"llama2-70b-4096",
|
|
||||||
"mixtral-8x7b-32768",
|
|
||||||
}
|
|
||||||
@@ -2,19 +2,14 @@ package channel
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"one-api/relay/channel/openai"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Adaptor interface {
|
type Adaptor interface {
|
||||||
Init(meta *util.RelayMeta)
|
GetRequestURL() string
|
||||||
GetRequestURL(meta *util.RelayMeta) (string, error)
|
Auth(c *gin.Context) error
|
||||||
SetupRequestHeader(c *gin.Context, req *http.Request, meta *util.RelayMeta) error
|
ConvertRequest(request *openai.GeneralOpenAIRequest) (any, error)
|
||||||
ConvertRequest(c *gin.Context, relayMode int, request *model.GeneralOpenAIRequest) (any, error)
|
DoRequest(request *openai.GeneralOpenAIRequest) error
|
||||||
DoRequest(c *gin.Context, meta *util.RelayMeta, requestBody io.Reader) (*http.Response, error)
|
DoResponse(c *gin.Context, resp *http.Response) (*openai.ErrorWithStatusCode, *openai.Usage, error)
|
||||||
DoResponse(c *gin.Context, resp *http.Response, meta *util.RelayMeta) (usage *model.Usage, err *model.ErrorWithStatusCode)
|
|
||||||
GetModelList() []string
|
|
||||||
GetChannelName() string
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,7 +0,0 @@
|
|||||||
package minimax
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"abab5.5s-chat",
|
|
||||||
"abab5.5-chat",
|
|
||||||
"abab6-chat",
|
|
||||||
}
|
|
||||||
@@ -1,14 +0,0 @@
|
|||||||
package minimax
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
"github.com/songquanpeng/one-api/relay/constant"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
)
|
|
||||||
|
|
||||||
func GetRequestURL(meta *util.RelayMeta) (string, error) {
|
|
||||||
if meta.Mode == constant.RelayModeChatCompletions {
|
|
||||||
return fmt.Sprintf("%s/v1/text/chatcompletion_v2", meta.BaseURL), nil
|
|
||||||
}
|
|
||||||
return "", fmt.Errorf("unsupported relay mode %d for minimax", meta.Mode)
|
|
||||||
}
|
|
||||||
@@ -1,10 +0,0 @@
|
|||||||
package mistral
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"open-mistral-7b",
|
|
||||||
"open-mixtral-8x7b",
|
|
||||||
"mistral-small-latest",
|
|
||||||
"mistral-medium-latest",
|
|
||||||
"mistral-large-latest",
|
|
||||||
"mistral-embed",
|
|
||||||
}
|
|
||||||
@@ -1,7 +0,0 @@
|
|||||||
package moonshot
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"moonshot-v1-8k",
|
|
||||||
"moonshot-v1-32k",
|
|
||||||
"moonshot-v1-128k",
|
|
||||||
}
|
|
||||||
@@ -1,92 +1,21 @@
|
|||||||
package openai
|
package openai
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/minimax"
|
|
||||||
"github.com/songquanpeng/one-api/relay/model"
|
|
||||||
"github.com/songquanpeng/one-api/relay/util"
|
|
||||||
"io"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
"strings"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
type Adaptor struct {
|
type Adaptor struct {
|
||||||
ChannelType int
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) Init(meta *util.RelayMeta) {
|
func (a *Adaptor) Auth(c *gin.Context) error {
|
||||||
a.ChannelType = meta.ChannelType
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetRequestURL(meta *util.RelayMeta) (string, error) {
|
|
||||||
switch meta.ChannelType {
|
|
||||||
case common.ChannelTypeAzure:
|
|
||||||
// https://learn.microsoft.com/en-us/azure/cognitive-services/openai/chatgpt-quickstart?pivots=rest-api&tabs=command-line#rest-api
|
|
||||||
requestURL := strings.Split(meta.RequestURLPath, "?")[0]
|
|
||||||
requestURL = fmt.Sprintf("%s?api-version=%s", requestURL, meta.APIVersion)
|
|
||||||
task := strings.TrimPrefix(requestURL, "/v1/")
|
|
||||||
model_ := meta.ActualModelName
|
|
||||||
model_ = strings.Replace(model_, ".", "", -1)
|
|
||||||
// https://github.com/songquanpeng/one-api/issues/67
|
|
||||||
model_ = strings.TrimSuffix(model_, "-0301")
|
|
||||||
model_ = strings.TrimSuffix(model_, "-0314")
|
|
||||||
model_ = strings.TrimSuffix(model_, "-0613")
|
|
||||||
|
|
||||||
requestURL = fmt.Sprintf("/openai/deployments/%s/%s", model_, task)
|
|
||||||
return util.GetFullRequestURL(meta.BaseURL, requestURL, meta.ChannelType), nil
|
|
||||||
case common.ChannelTypeMinimax:
|
|
||||||
return minimax.GetRequestURL(meta)
|
|
||||||
default:
|
|
||||||
return util.GetFullRequestURL(meta.BaseURL, meta.RequestURLPath, meta.ChannelType), nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) SetupRequestHeader(c *gin.Context, req *http.Request, meta *util.RelayMeta) error {
|
|
||||||
channel.SetupCommonRequestHeader(c, req, meta)
|
|
||||||
if meta.ChannelType == common.ChannelTypeAzure {
|
|
||||||
req.Header.Set("api-key", meta.APIKey)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
req.Header.Set("Authorization", "Bearer "+meta.APIKey)
|
|
||||||
if meta.ChannelType == common.ChannelTypeOpenRouter {
|
|
||||||
req.Header.Set("HTTP-Referer", "https://github.com/songquanpeng/one-api")
|
|
||||||
req.Header.Set("X-Title", "One API")
|
|
||||||
}
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) ConvertRequest(c *gin.Context, relayMode int, request *model.GeneralOpenAIRequest) (any, error) {
|
func (a *Adaptor) ConvertRequest(request *GeneralOpenAIRequest) (any, error) {
|
||||||
if request == nil {
|
return nil, nil
|
||||||
return nil, errors.New("request is nil")
|
|
||||||
}
|
|
||||||
return request, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (a *Adaptor) DoRequest(c *gin.Context, meta *util.RelayMeta, requestBody io.Reader) (*http.Response, error) {
|
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response) (*ErrorWithStatusCode, *Usage, error) {
|
||||||
return channel.DoRequestHelper(a, c, meta, requestBody)
|
return nil, nil, nil
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) DoResponse(c *gin.Context, resp *http.Response, meta *util.RelayMeta) (usage *model.Usage, err *model.ErrorWithStatusCode) {
|
|
||||||
if meta.IsStream {
|
|
||||||
var responseText string
|
|
||||||
err, responseText, _ = StreamHandler(c, resp, meta.Mode)
|
|
||||||
usage = ResponseText2Usage(responseText, meta.ActualModelName, meta.PromptTokens)
|
|
||||||
} else {
|
|
||||||
err, usage = Handler(c, resp, meta.PromptTokens, meta.ActualModelName)
|
|
||||||
}
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetModelList() []string {
|
|
||||||
_, modelList := GetCompatibleChannelMeta(a.ChannelType)
|
|
||||||
return modelList
|
|
||||||
}
|
|
||||||
|
|
||||||
func (a *Adaptor) GetChannelName() string {
|
|
||||||
channelName, _ := GetCompatibleChannelMeta(a.ChannelType)
|
|
||||||
return channelName
|
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,42 +0,0 @@
|
|||||||
package openai
|
|
||||||
|
|
||||||
import (
|
|
||||||
"github.com/songquanpeng/one-api/common"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/ai360"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/baichuan"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/groq"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/minimax"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/mistral"
|
|
||||||
"github.com/songquanpeng/one-api/relay/channel/moonshot"
|
|
||||||
)
|
|
||||||
|
|
||||||
var CompatibleChannels = []int{
|
|
||||||
common.ChannelTypeAzure,
|
|
||||||
common.ChannelType360,
|
|
||||||
common.ChannelTypeMoonshot,
|
|
||||||
common.ChannelTypeBaichuan,
|
|
||||||
common.ChannelTypeMinimax,
|
|
||||||
common.ChannelTypeMistral,
|
|
||||||
common.ChannelTypeGroq,
|
|
||||||
}
|
|
||||||
|
|
||||||
func GetCompatibleChannelMeta(channelType int) (string, []string) {
|
|
||||||
switch channelType {
|
|
||||||
case common.ChannelTypeAzure:
|
|
||||||
return "azure", ModelList
|
|
||||||
case common.ChannelType360:
|
|
||||||
return "360", ai360.ModelList
|
|
||||||
case common.ChannelTypeMoonshot:
|
|
||||||
return "moonshot", moonshot.ModelList
|
|
||||||
case common.ChannelTypeBaichuan:
|
|
||||||
return "baichuan", baichuan.ModelList
|
|
||||||
case common.ChannelTypeMinimax:
|
|
||||||
return "minimax", minimax.ModelList
|
|
||||||
case common.ChannelTypeMistral:
|
|
||||||
return "mistralai", mistral.ModelList
|
|
||||||
case common.ChannelTypeGroq:
|
|
||||||
return "groq", groq.ModelList
|
|
||||||
default:
|
|
||||||
return "openai", ModelList
|
|
||||||
}
|
|
||||||
}
|
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
package model
|
package openai
|
||||||
|
|
||||||
const (
|
const (
|
||||||
ContentTypeText = "text"
|
ContentTypeText = "text"
|
||||||
@@ -1,19 +0,0 @@
|
|||||||
package openai
|
|
||||||
|
|
||||||
var ModelList = []string{
|
|
||||||
"gpt-3.5-turbo", "gpt-3.5-turbo-0301", "gpt-3.5-turbo-0613", "gpt-3.5-turbo-1106", "gpt-3.5-turbo-0125",
|
|
||||||
"gpt-3.5-turbo-16k", "gpt-3.5-turbo-16k-0613",
|
|
||||||
"gpt-3.5-turbo-instruct",
|
|
||||||
"gpt-4", "gpt-4-0314", "gpt-4-0613", "gpt-4-1106-preview", "gpt-4-0125-preview",
|
|
||||||
"gpt-4-32k", "gpt-4-32k-0314", "gpt-4-32k-0613",
|
|
||||||
"gpt-4-turbo-preview",
|
|
||||||
"gpt-4-vision-preview",
|
|
||||||
"text-embedding-ada-002", "text-embedding-3-small", "text-embedding-3-large",
|
|
||||||
"text-curie-001", "text-babbage-001", "text-ada-001", "text-davinci-002", "text-davinci-003",
|
|
||||||
"text-moderation-latest", "text-moderation-stable",
|
|
||||||
"text-davinci-edit-001",
|
|
||||||
"davinci-002", "babbage-002",
|
|
||||||
"dall-e-2", "dall-e-3",
|
|
||||||
"whisper-1",
|
|
||||||
"tts-1", "tts-1-1106", "tts-1-hd", "tts-1-hd-1106",
|
|
||||||
}
|
|
||||||
@@ -1,11 +0,0 @@
|
|||||||
package openai
|
|
||||||
|
|
||||||
import "github.com/songquanpeng/one-api/relay/model"
|
|
||||||
|
|
||||||
func ResponseText2Usage(responseText string, modeName string, promptTokens int) *model.Usage {
|
|
||||||
usage := &model.Usage{}
|
|
||||||
usage.PromptTokens = promptTokens
|
|
||||||
usage.CompletionTokens = CountTokenText(responseText, modeName)
|
|
||||||
usage.TotalTokens = usage.PromptTokens + usage.CompletionTokens
|
|
||||||
return usage
|
|
||||||
}
|
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user