Compare commits
9 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
fc284cc1f0 | ||
|
|
9c0a49b97a | ||
|
|
61c47a3b08 | ||
|
|
c5aa59e297 | ||
|
|
211a862d54 | ||
|
|
c4c89e8e1b | ||
|
|
72983ac734 | ||
|
|
4d43dce64b | ||
|
|
0fa94d3c94 |
@@ -189,6 +189,7 @@ const (
|
|||||||
ChannelTypeTencent = 23
|
ChannelTypeTencent = 23
|
||||||
ChannelTypeAzureSpeech = 24
|
ChannelTypeAzureSpeech = 24
|
||||||
ChannelTypeGemini = 25
|
ChannelTypeGemini = 25
|
||||||
|
ChannelTypeBaichuan = 26
|
||||||
)
|
)
|
||||||
|
|
||||||
var ChannelBaseURLs = []string{
|
var ChannelBaseURLs = []string{
|
||||||
@@ -218,6 +219,7 @@ var ChannelBaseURLs = []string{
|
|||||||
"https://hunyuan.cloud.tencent.com", //23
|
"https://hunyuan.cloud.tencent.com", //23
|
||||||
"", //24
|
"", //24
|
||||||
"", //25
|
"", //25
|
||||||
|
"https://api.baichuan-ai.com", //26
|
||||||
}
|
}
|
||||||
|
|
||||||
const (
|
const (
|
||||||
|
|||||||
@@ -3,6 +3,7 @@ package image
|
|||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"encoding/base64"
|
"encoding/base64"
|
||||||
|
"errors"
|
||||||
"image"
|
"image"
|
||||||
_ "image/gif"
|
_ "image/gif"
|
||||||
_ "image/jpeg"
|
_ "image/jpeg"
|
||||||
@@ -44,8 +45,26 @@ func GetImageSizeFromUrl(url string) (width int, height int, err error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func GetImageFromUrl(url string) (mimeType string, data string, err error) {
|
func GetImageFromUrl(url string) (mimeType string, data string, err error) {
|
||||||
|
|
||||||
|
if strings.HasPrefix(url, "data:image/") {
|
||||||
|
dataURLPattern := regexp.MustCompile(`data:image/([^;]+);base64,(.*)`)
|
||||||
|
|
||||||
|
matches := dataURLPattern.FindStringSubmatch(url)
|
||||||
|
if len(matches) == 3 && matches[2] != "" {
|
||||||
|
mimeType = "image/" + matches[1]
|
||||||
|
data = matches[2]
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
err = errors.New("image base64 decode failed")
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
isImage, err := IsImageUrl(url)
|
isImage, err := IsImageUrl(url)
|
||||||
if !isImage {
|
if !isImage {
|
||||||
|
if err == nil {
|
||||||
|
err = errors.New("invalid image link")
|
||||||
|
}
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
resp, err := http.Get(url)
|
resp, err := http.Get(url)
|
||||||
|
|||||||
@@ -169,3 +169,34 @@ func TestGetImageSizeFromBase64(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestGetImageFromUrl(t *testing.T) {
|
||||||
|
for i, c := range cases {
|
||||||
|
t.Run("Decode:"+strconv.Itoa(i), func(t *testing.T) {
|
||||||
|
resp, err := http.Get(c.url)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
defer resp.Body.Close()
|
||||||
|
data, err := io.ReadAll(resp.Body)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
encoded := base64.StdEncoding.EncodeToString(data)
|
||||||
|
|
||||||
|
mimeType, base64Data, err := img.GetImageFromUrl(c.url)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, encoded, base64Data)
|
||||||
|
assert.Equal(t, "image/"+c.format, mimeType)
|
||||||
|
|
||||||
|
encodedBase64 := "data:image/" + c.format + ";base64," + encoded
|
||||||
|
mimeType, base64Data, err = img.GetImageFromUrl(encodedBase64)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, encoded, base64Data)
|
||||||
|
assert.Equal(t, "image/"+c.format, mimeType)
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
url := "https://raw.githubusercontent.com/songquanpeng/one-api/main/README.md"
|
||||||
|
_, _, err := img.GetImageFromUrl(url)
|
||||||
|
assert.Error(t, err)
|
||||||
|
encodedBase64 := "data:image/text;base64,"
|
||||||
|
_, _, err = img.GetImageFromUrl(encodedBase64)
|
||||||
|
assert.Error(t, err)
|
||||||
|
}
|
||||||
|
|||||||
@@ -93,6 +93,7 @@ var ModelRatio = map[string]float64{
|
|||||||
"qwen-plus": 1.4286, // ¥0.02 / 1k tokens
|
"qwen-plus": 1.4286, // ¥0.02 / 1k tokens
|
||||||
"qwen-max": 1.4286, // ¥0.02 / 1k tokens
|
"qwen-max": 1.4286, // ¥0.02 / 1k tokens
|
||||||
"qwen-max-longcontext": 1.4286, // ¥0.02 / 1k tokens
|
"qwen-max-longcontext": 1.4286, // ¥0.02 / 1k tokens
|
||||||
|
"qwen-vl-plus": 0.5715, // ¥0.008 / 1k tokens
|
||||||
"text-embedding-v1": 0.05, // ¥0.0007 / 1k tokens
|
"text-embedding-v1": 0.05, // ¥0.0007 / 1k tokens
|
||||||
"SparkDesk": 1.2858, // ¥0.018 / 1k tokens
|
"SparkDesk": 1.2858, // ¥0.018 / 1k tokens
|
||||||
"360GPT_S2_V9": 0.8572, // ¥0.012 / 1k tokens
|
"360GPT_S2_V9": 0.8572, // ¥0.012 / 1k tokens
|
||||||
@@ -100,6 +101,10 @@ var ModelRatio = map[string]float64{
|
|||||||
"embedding_s1_v1": 0.0715, // ¥0.001 / 1k tokens
|
"embedding_s1_v1": 0.0715, // ¥0.001 / 1k tokens
|
||||||
"semantic_similarity_s1_v1": 0.0715, // ¥0.001 / 1k tokens
|
"semantic_similarity_s1_v1": 0.0715, // ¥0.001 / 1k tokens
|
||||||
"hunyuan": 7.143, // ¥0.1 / 1k tokens // https://cloud.tencent.com/document/product/1729/97731#e0e6be58-60c8-469f-bdeb-6c264ce3b4d0
|
"hunyuan": 7.143, // ¥0.1 / 1k tokens // https://cloud.tencent.com/document/product/1729/97731#e0e6be58-60c8-469f-bdeb-6c264ce3b4d0
|
||||||
|
"Baichuan2-Turbo": 0.5715, // ¥0.008 / 1k tokens
|
||||||
|
"Baichuan2-Turbo-192k": 1.143, // ¥0.016 / 1k tokens
|
||||||
|
"Baichuan2-53B": 1.4286, // ¥0.02 / 1k tokens
|
||||||
|
"Baichuan-Text-Embedding": 0.0357, // ¥0.0005 / 1k tokens
|
||||||
}
|
}
|
||||||
|
|
||||||
func ModelRatio2JSONString() string {
|
func ModelRatio2JSONString() string {
|
||||||
|
|||||||
@@ -190,13 +190,13 @@ func countImageTokens(url string, detail string) (_ int, err error) {
|
|||||||
func CountTokenInput(input any, model string) int {
|
func CountTokenInput(input any, model string) int {
|
||||||
switch v := input.(type) {
|
switch v := input.(type) {
|
||||||
case string:
|
case string:
|
||||||
return CountTokenInput(v, model)
|
return CountTokenText(v, model)
|
||||||
case []string:
|
case []string:
|
||||||
text := ""
|
text := ""
|
||||||
for _, s := range v {
|
for _, s := range v {
|
||||||
text += s
|
text += s
|
||||||
}
|
}
|
||||||
return CountTokenInput(text, model)
|
return CountTokenText(text, model)
|
||||||
}
|
}
|
||||||
return 0
|
return 0
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -18,6 +18,10 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
func testChannel(channel *model.Channel, request types.ChatCompletionRequest) (err error, openaiErr *types.OpenAIError) {
|
func testChannel(channel *model.Channel, request types.ChatCompletionRequest) (err error, openaiErr *types.OpenAIError) {
|
||||||
|
if channel.TestModel == "" {
|
||||||
|
return errors.New("请填写测速模型后再试"), nil
|
||||||
|
}
|
||||||
|
|
||||||
// 创建一个 http.Request
|
// 创建一个 http.Request
|
||||||
req, err := http.NewRequest("POST", "/v1/chat/completions", nil)
|
req, err := http.NewRequest("POST", "/v1/chat/completions", nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -28,26 +32,7 @@ func testChannel(channel *model.Channel, request types.ChatCompletionRequest) (e
|
|||||||
w := httptest.NewRecorder()
|
w := httptest.NewRecorder()
|
||||||
c, _ := gin.CreateTestContext(w)
|
c, _ := gin.CreateTestContext(w)
|
||||||
c.Request = req
|
c.Request = req
|
||||||
|
request.Model = channel.TestModel
|
||||||
// 创建映射
|
|
||||||
channelTypeToModel := map[int]string{
|
|
||||||
common.ChannelTypePaLM: "PaLM-2",
|
|
||||||
common.ChannelTypeAnthropic: "claude-2",
|
|
||||||
common.ChannelTypeBaidu: "ERNIE-Bot",
|
|
||||||
common.ChannelTypeZhipu: "chatglm_lite",
|
|
||||||
common.ChannelTypeAli: "qwen-turbo",
|
|
||||||
common.ChannelType360: "360GPT_S2_V9",
|
|
||||||
common.ChannelTypeXunfei: "SparkDesk",
|
|
||||||
common.ChannelTypeTencent: "hunyuan",
|
|
||||||
common.ChannelTypeAzure: "gpt-3.5-turbo",
|
|
||||||
}
|
|
||||||
|
|
||||||
// 从映射中获取模型名称
|
|
||||||
model, ok := channelTypeToModel[channel.Type]
|
|
||||||
if !ok {
|
|
||||||
model = "gpt-3.5-turbo" // 默认值
|
|
||||||
}
|
|
||||||
request.Model = model
|
|
||||||
|
|
||||||
provider := providers.GetProvider(channel, c)
|
provider := providers.GetProvider(channel, c)
|
||||||
if provider == nil {
|
if provider == nil {
|
||||||
@@ -69,13 +54,15 @@ func testChannel(channel *model.Channel, request types.ChatCompletionRequest) (e
|
|||||||
promptTokens := common.CountTokenMessages(request.Messages, request.Model)
|
promptTokens := common.CountTokenMessages(request.Messages, request.Model)
|
||||||
Usage, openAIErrorWithStatusCode := chatProvider.ChatAction(&request, true, promptTokens)
|
Usage, openAIErrorWithStatusCode := chatProvider.ChatAction(&request, true, promptTokens)
|
||||||
if openAIErrorWithStatusCode != nil {
|
if openAIErrorWithStatusCode != nil {
|
||||||
return nil, &openAIErrorWithStatusCode.OpenAIError
|
return errors.New(openAIErrorWithStatusCode.Message), &openAIErrorWithStatusCode.OpenAIError
|
||||||
}
|
}
|
||||||
|
|
||||||
if Usage.CompletionTokens == 0 {
|
if Usage.CompletionTokens == 0 {
|
||||||
return fmt.Errorf("channel %s, message 补全 tokens 非预期返回 0", channel.Name), nil
|
return fmt.Errorf("channel %s, message 补全 tokens 非预期返回 0", channel.Name), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
common.SysLog(fmt.Sprintf("测试模型 %s 返回内容为:%s", channel.Name, w.Body.String()))
|
||||||
|
|
||||||
return nil, nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -2,7 +2,11 @@ package controller
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/model"
|
||||||
"one-api/types"
|
"one-api/types"
|
||||||
|
"sort"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
)
|
)
|
||||||
@@ -25,559 +29,38 @@ type OpenAIModelPermission struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type OpenAIModels struct {
|
type OpenAIModels struct {
|
||||||
Id string `json:"id"`
|
Id string `json:"id"`
|
||||||
Object string `json:"object"`
|
Object string `json:"object"`
|
||||||
Created int `json:"created"`
|
Created int `json:"created"`
|
||||||
OwnedBy string `json:"owned_by"`
|
OwnedBy *string `json:"owned_by"`
|
||||||
Permission []OpenAIModelPermission `json:"permission"`
|
Permission *[]OpenAIModelPermission `json:"permission"`
|
||||||
Root string `json:"root"`
|
Root *string `json:"root"`
|
||||||
Parent *string `json:"parent"`
|
Parent *string `json:"parent"`
|
||||||
}
|
}
|
||||||
|
|
||||||
var openAIModels []OpenAIModels
|
var openAIModels []OpenAIModels
|
||||||
var openAIModelsMap map[string]OpenAIModels
|
var openAIModelsMap map[string]OpenAIModels
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
var permission []OpenAIModelPermission
|
|
||||||
permission = append(permission, OpenAIModelPermission{
|
|
||||||
Id: "modelperm-LwHkVFn8AcMItP432fKKDIKJ",
|
|
||||||
Object: "model_permission",
|
|
||||||
Created: 1626777600,
|
|
||||||
AllowCreateEngine: true,
|
|
||||||
AllowSampling: true,
|
|
||||||
AllowLogprobs: true,
|
|
||||||
AllowSearchIndices: false,
|
|
||||||
AllowView: true,
|
|
||||||
AllowFineTuning: false,
|
|
||||||
Organization: "*",
|
|
||||||
Group: nil,
|
|
||||||
IsBlocking: false,
|
|
||||||
})
|
|
||||||
// https://platform.openai.com/docs/models/model-endpoint-compatibility
|
// https://platform.openai.com/docs/models/model-endpoint-compatibility
|
||||||
openAIModels = []OpenAIModels{
|
keys := make([]string, 0, len(common.ModelRatio))
|
||||||
{
|
for k := range common.ModelRatio {
|
||||||
Id: "dall-e-2",
|
keys = append(keys, k)
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "dall-e-2",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "dall-e-3",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "dall-e-3",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "whisper-1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "whisper-1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "tts-1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "tts-1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "tts-1-1106",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "tts-1-1106",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "tts-1-hd",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "tts-1-hd",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "tts-1-hd-1106",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "tts-1-hd-1106",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-3.5-turbo",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-3.5-turbo",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-3.5-turbo-0301",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-3.5-turbo-0301",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-3.5-turbo-0613",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-3.5-turbo-0613",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-3.5-turbo-16k",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-3.5-turbo-16k",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-3.5-turbo-16k-0613",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-3.5-turbo-16k-0613",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-3.5-turbo-1106",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1699593571,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-3.5-turbo-1106",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-3.5-turbo-instruct",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-3.5-turbo-instruct",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-4",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-4",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-4-0314",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-4-0314",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-4-0613",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-4-0613",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-4-32k",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-4-32k",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-4-32k-0314",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-4-32k-0314",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-4-32k-0613",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-4-32k-0613",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-4-1106-preview",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1699593571,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-4-1106-preview",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gpt-4-vision-preview",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1699593571,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gpt-4-vision-preview",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-embedding-ada-002",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-embedding-ada-002",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-davinci-003",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-davinci-003",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-davinci-002",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-davinci-002",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-curie-001",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-curie-001",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-babbage-001",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-babbage-001",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-ada-001",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-ada-001",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-moderation-latest",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-moderation-latest",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-moderation-stable",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-moderation-stable",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-davinci-edit-001",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-davinci-edit-001",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "code-davinci-edit-001",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "openai",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "code-davinci-edit-001",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "claude-instant-1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "anthropic",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "claude-instant-1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "claude-2",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "anthropic",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "claude-2",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "claude-2.1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "anthropic",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "claude-2.1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "claude-2.0",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "anthropic",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "claude-2.0",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "ERNIE-Bot",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "baidu",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "ERNIE-Bot",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "ERNIE-Bot-turbo",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "baidu",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "ERNIE-Bot-turbo",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "ERNIE-Bot-4",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "baidu",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "ERNIE-Bot-4",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "Embedding-V1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "baidu",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "Embedding-V1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "PaLM-2",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "google",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "PaLM-2",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gemini-pro",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "google",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gemini-pro",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "gemini-pro-vision",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "google",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "gemini-pro-vision",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "chatglm_turbo",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "zhipu",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "chatglm_turbo",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "chatglm_pro",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "zhipu",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "chatglm_pro",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "chatglm_std",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "zhipu",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "chatglm_std",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "chatglm_lite",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "zhipu",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "chatglm_lite",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "qwen-turbo",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "ali",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "qwen-turbo",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "qwen-plus",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "ali",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "qwen-plus",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "qwen-max",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "ali",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "qwen-max",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "qwen-max-longcontext",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "ali",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "qwen-max-longcontext",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "text-embedding-v1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "ali",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "text-embedding-v1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "SparkDesk",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "xunfei",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "SparkDesk",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "360GPT_S2_V9",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "360",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "360GPT_S2_V9",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "embedding-bert-512-v1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "360",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "embedding-bert-512-v1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "embedding_s1_v1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "360",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "embedding_s1_v1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "semantic_similarity_s1_v1",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "360",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "semantic_similarity_s1_v1",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Id: "hunyuan",
|
|
||||||
Object: "model",
|
|
||||||
Created: 1677649963,
|
|
||||||
OwnedBy: "tencent",
|
|
||||||
Permission: permission,
|
|
||||||
Root: "hunyuan",
|
|
||||||
Parent: nil,
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
|
sort.Strings(keys)
|
||||||
|
|
||||||
|
for _, modelId := range keys {
|
||||||
|
openAIModels = append(openAIModels, OpenAIModels{
|
||||||
|
Id: modelId,
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: nil,
|
||||||
|
Permission: nil,
|
||||||
|
Root: nil,
|
||||||
|
Parent: nil,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
openAIModelsMap = make(map[string]OpenAIModels)
|
openAIModelsMap = make(map[string]OpenAIModels)
|
||||||
for _, model := range openAIModels {
|
for _, model := range openAIModels {
|
||||||
openAIModelsMap[model.Id] = model
|
openAIModelsMap[model.Id] = model
|
||||||
@@ -585,6 +68,35 @@ func init() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func ListModels(c *gin.Context) {
|
func ListModels(c *gin.Context) {
|
||||||
|
groupName := c.GetString("group")
|
||||||
|
|
||||||
|
models, err := model.CacheGetGroupModels(groupName)
|
||||||
|
if err != nil {
|
||||||
|
common.AbortWithMessage(c, http.StatusServiceUnavailable, err.Error())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
sort.Strings(models)
|
||||||
|
|
||||||
|
groupOpenAIModels := make([]OpenAIModels, 0, len(models))
|
||||||
|
for _, modelId := range models {
|
||||||
|
groupOpenAIModels = append(groupOpenAIModels, OpenAIModels{
|
||||||
|
Id: modelId,
|
||||||
|
Object: "model",
|
||||||
|
Created: 1677649963,
|
||||||
|
OwnedBy: nil,
|
||||||
|
Permission: nil,
|
||||||
|
Root: nil,
|
||||||
|
Parent: nil,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
c.JSON(200, gin.H{
|
||||||
|
"object": "list",
|
||||||
|
"data": groupOpenAIModels,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func ListModelsForAdmin(c *gin.Context) {
|
||||||
c.JSON(200, gin.H{
|
c.JSON(200, gin.H{
|
||||||
"object": "list",
|
"object": "list",
|
||||||
"data": openAIModels,
|
"data": openAIModels,
|
||||||
|
|||||||
@@ -45,6 +45,8 @@ func fetchChannel(c *gin.Context, modelName string) (channel *model.Channel, pas
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
c.Set("channel_id", channel.Id)
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -39,6 +39,22 @@ func GetRandomSatisfiedChannel(group string, model string) (*Channel, error) {
|
|||||||
return &channel, err
|
return &channel, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func GetGroupModels(group string) ([]string, error) {
|
||||||
|
var models []string
|
||||||
|
groupCol := "`group`"
|
||||||
|
trueVal := "1"
|
||||||
|
if common.UsingPostgreSQL {
|
||||||
|
groupCol = `"group"`
|
||||||
|
trueVal = "true"
|
||||||
|
}
|
||||||
|
|
||||||
|
err := DB.Model(&Ability{}).Where(groupCol+" = ? and enabled = ? ", group, trueVal).Distinct("model").Pluck("model", &models).Error
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return models, nil
|
||||||
|
}
|
||||||
|
|
||||||
func (channel *Channel) AddAbilities() error {
|
func (channel *Channel) AddAbilities() error {
|
||||||
models_ := strings.Split(channel.Models, ",")
|
models_ := strings.Split(channel.Models, ",")
|
||||||
groups_ := strings.Split(channel.Group, ",")
|
groups_ := strings.Split(channel.Group, ",")
|
||||||
|
|||||||
@@ -213,3 +213,22 @@ func CacheGetRandomSatisfiedChannel(group string, model string) (*Channel, error
|
|||||||
idx := rand.Intn(endIdx)
|
idx := rand.Intn(endIdx)
|
||||||
return channels[idx], nil
|
return channels[idx], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func CacheGetGroupModels(group string) ([]string, error) {
|
||||||
|
if !common.MemoryCacheEnabled {
|
||||||
|
return GetGroupModels(group)
|
||||||
|
}
|
||||||
|
channelSyncLock.RLock()
|
||||||
|
defer channelSyncLock.RUnlock()
|
||||||
|
|
||||||
|
groupModels := group2model2channels[group]
|
||||||
|
if groupModels == nil {
|
||||||
|
return nil, errors.New("group not found")
|
||||||
|
}
|
||||||
|
|
||||||
|
models := make([]string, 0)
|
||||||
|
for model := range groupModels {
|
||||||
|
models = append(models, model)
|
||||||
|
}
|
||||||
|
return models, nil
|
||||||
|
}
|
||||||
|
|||||||
@@ -26,6 +26,7 @@ type Channel struct {
|
|||||||
ModelMapping *string `json:"model_mapping" gorm:"type:varchar(1024);default:''"`
|
ModelMapping *string `json:"model_mapping" gorm:"type:varchar(1024);default:''"`
|
||||||
Priority *int64 `json:"priority" gorm:"bigint;default:0"`
|
Priority *int64 `json:"priority" gorm:"bigint;default:0"`
|
||||||
Proxy string `json:"proxy" gorm:"type:varchar(255);default:''"`
|
Proxy string `json:"proxy" gorm:"type:varchar(255);default:''"`
|
||||||
|
TestModel string `json:"test_model" gorm:"type:varchar(50);default:''"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func GetAllChannels(startIdx int, num int, selectAll bool) ([]*Channel, error) {
|
func GetAllChannels(startIdx int, num int, selectAll bool) ([]*Channel, error) {
|
||||||
|
|||||||
@@ -2,6 +2,7 @@ package ali
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"strings"
|
||||||
|
|
||||||
"one-api/providers/base"
|
"one-api/providers/base"
|
||||||
|
|
||||||
@@ -28,6 +29,16 @@ type AliProvider struct {
|
|||||||
base.BaseProvider
|
base.BaseProvider
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (p *AliProvider) GetFullRequestURL(requestURL string, modelName string) string {
|
||||||
|
baseURL := strings.TrimSuffix(p.GetBaseURL(), "/")
|
||||||
|
|
||||||
|
if modelName == "qwen-vl-plus" {
|
||||||
|
requestURL = "/api/v1/services/aigc/multimodal-generation/generation"
|
||||||
|
}
|
||||||
|
|
||||||
|
return fmt.Sprintf("%s%s", baseURL, requestURL)
|
||||||
|
}
|
||||||
|
|
||||||
// 获取请求头
|
// 获取请求头
|
||||||
func (p *AliProvider) GetRequestHeaders() (headers map[string]string) {
|
func (p *AliProvider) GetRequestHeaders() (headers map[string]string) {
|
||||||
headers = make(map[string]string)
|
headers = make(map[string]string)
|
||||||
|
|||||||
@@ -26,21 +26,12 @@ func (aliResponse *AliChatResponse) ResponseHandler(resp *http.Response) (OpenAI
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
choice := types.ChatCompletionChoice{
|
|
||||||
Index: 0,
|
|
||||||
Message: types.ChatCompletionMessage{
|
|
||||||
Role: "assistant",
|
|
||||||
Content: aliResponse.Output.Text,
|
|
||||||
},
|
|
||||||
FinishReason: aliResponse.Output.FinishReason,
|
|
||||||
}
|
|
||||||
|
|
||||||
OpenAIResponse = types.ChatCompletionResponse{
|
OpenAIResponse = types.ChatCompletionResponse{
|
||||||
ID: aliResponse.RequestId,
|
ID: aliResponse.RequestId,
|
||||||
Object: "chat.completion",
|
Object: "chat.completion",
|
||||||
Created: common.GetTimestamp(),
|
Created: common.GetTimestamp(),
|
||||||
Model: aliResponse.Model,
|
Model: aliResponse.Model,
|
||||||
Choices: []types.ChatCompletionChoice{choice},
|
Choices: aliResponse.Output.ToChatCompletionChoices(),
|
||||||
Usage: &types.Usage{
|
Usage: &types.Usage{
|
||||||
PromptTokens: aliResponse.Usage.InputTokens,
|
PromptTokens: aliResponse.Usage.InputTokens,
|
||||||
CompletionTokens: aliResponse.Usage.OutputTokens,
|
CompletionTokens: aliResponse.Usage.OutputTokens,
|
||||||
@@ -58,10 +49,31 @@ func (p *AliProvider) getChatRequestBody(request *types.ChatCompletionRequest) *
|
|||||||
messages := make([]AliMessage, 0, len(request.Messages))
|
messages := make([]AliMessage, 0, len(request.Messages))
|
||||||
for i := 0; i < len(request.Messages); i++ {
|
for i := 0; i < len(request.Messages); i++ {
|
||||||
message := request.Messages[i]
|
message := request.Messages[i]
|
||||||
messages = append(messages, AliMessage{
|
if request.Model != "qwen-vl-plus" {
|
||||||
Content: message.StringContent(),
|
messages = append(messages, AliMessage{
|
||||||
Role: strings.ToLower(message.Role),
|
Content: message.StringContent(),
|
||||||
})
|
Role: strings.ToLower(message.Role),
|
||||||
|
})
|
||||||
|
} else {
|
||||||
|
openaiContent := message.ParseContent()
|
||||||
|
var parts []AliMessagePart
|
||||||
|
for _, part := range openaiContent {
|
||||||
|
if part.Type == types.ContentTypeText {
|
||||||
|
parts = append(parts, AliMessagePart{
|
||||||
|
Text: part.Text,
|
||||||
|
})
|
||||||
|
} else if part.Type == types.ContentTypeImageURL {
|
||||||
|
parts = append(parts, AliMessagePart{
|
||||||
|
Image: part.ImageURL.URL,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
messages = append(messages, AliMessage{
|
||||||
|
Content: parts,
|
||||||
|
Role: strings.ToLower(message.Role),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
enableSearch := false
|
enableSearch := false
|
||||||
@@ -77,6 +89,7 @@ func (p *AliProvider) getChatRequestBody(request *types.ChatCompletionRequest) *
|
|||||||
Messages: messages,
|
Messages: messages,
|
||||||
},
|
},
|
||||||
Parameters: AliParameters{
|
Parameters: AliParameters{
|
||||||
|
ResultFormat: "message",
|
||||||
EnableSearch: enableSearch,
|
EnableSearch: enableSearch,
|
||||||
IncrementalOutput: request.Stream,
|
IncrementalOutput: request.Stream,
|
||||||
},
|
},
|
||||||
@@ -87,6 +100,7 @@ func (p *AliProvider) getChatRequestBody(request *types.ChatCompletionRequest) *
|
|||||||
func (p *AliProvider) ChatAction(request *types.ChatCompletionRequest, isModelMapped bool, promptTokens int) (usage *types.Usage, errWithCode *types.OpenAIErrorWithStatusCode) {
|
func (p *AliProvider) ChatAction(request *types.ChatCompletionRequest, isModelMapped bool, promptTokens int) (usage *types.Usage, errWithCode *types.OpenAIErrorWithStatusCode) {
|
||||||
|
|
||||||
requestBody := p.getChatRequestBody(request)
|
requestBody := p.getChatRequestBody(request)
|
||||||
|
|
||||||
fullRequestURL := p.GetFullRequestURL(p.ChatCompletions, request.Model)
|
fullRequestURL := p.GetFullRequestURL(p.ChatCompletions, request.Model)
|
||||||
headers := p.GetRequestHeaders()
|
headers := p.GetRequestHeaders()
|
||||||
if request.Stream {
|
if request.Stream {
|
||||||
@@ -134,10 +148,15 @@ func (p *AliProvider) ChatAction(request *types.ChatCompletionRequest, isModelMa
|
|||||||
|
|
||||||
// 阿里云响应转OpenAI响应
|
// 阿里云响应转OpenAI响应
|
||||||
func (p *AliProvider) streamResponseAli2OpenAI(aliResponse *AliChatResponse) *types.ChatCompletionStreamResponse {
|
func (p *AliProvider) streamResponseAli2OpenAI(aliResponse *AliChatResponse) *types.ChatCompletionStreamResponse {
|
||||||
|
// chatChoice := aliResponse.Output.ToChatCompletionChoices()
|
||||||
|
// jsonBody, _ := json.MarshalIndent(chatChoice, "", " ")
|
||||||
|
// fmt.Println("requestBody:", string(jsonBody))
|
||||||
var choice types.ChatCompletionStreamChoice
|
var choice types.ChatCompletionStreamChoice
|
||||||
choice.Delta.Content = aliResponse.Output.Text
|
choice.Index = aliResponse.Output.Choices[0].Index
|
||||||
if aliResponse.Output.FinishReason != "null" {
|
choice.Delta.Content = aliResponse.Output.Choices[0].Message.StringContent()
|
||||||
finishReason := aliResponse.Output.FinishReason
|
// fmt.Println("choice.Delta.Content:", chatChoice[0].Message)
|
||||||
|
if aliResponse.Output.Choices[0].FinishReason != "null" {
|
||||||
|
finishReason := aliResponse.Output.Choices[0].FinishReason
|
||||||
choice.FinishReason = &finishReason
|
choice.FinishReason = &finishReason
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -200,7 +219,8 @@ func (p *AliProvider) sendStreamRequest(req *http.Request, model string) (usage
|
|||||||
stopChan <- true
|
stopChan <- true
|
||||||
}()
|
}()
|
||||||
common.SetEventStreamHeaders(p.Context)
|
common.SetEventStreamHeaders(p.Context)
|
||||||
// lastResponseText := ""
|
lastResponseText := ""
|
||||||
|
index := 0
|
||||||
p.Context.Stream(func(w io.Writer) bool {
|
p.Context.Stream(func(w io.Writer) bool {
|
||||||
select {
|
select {
|
||||||
case data := <-dataChan:
|
case data := <-dataChan:
|
||||||
@@ -216,9 +236,11 @@ func (p *AliProvider) sendStreamRequest(req *http.Request, model string) (usage
|
|||||||
usage.TotalTokens = aliResponse.Usage.InputTokens + aliResponse.Usage.OutputTokens
|
usage.TotalTokens = aliResponse.Usage.InputTokens + aliResponse.Usage.OutputTokens
|
||||||
}
|
}
|
||||||
aliResponse.Model = model
|
aliResponse.Model = model
|
||||||
|
aliResponse.Output.Choices[0].Index = index
|
||||||
|
index++
|
||||||
response := p.streamResponseAli2OpenAI(&aliResponse)
|
response := p.streamResponseAli2OpenAI(&aliResponse)
|
||||||
// response.Choices[0].Delta.Content = strings.TrimPrefix(response.Choices[0].Delta.Content, lastResponseText)
|
response.Choices[0].Delta.Content = strings.TrimPrefix(response.Choices[0].Delta.Content, lastResponseText)
|
||||||
// lastResponseText = aliResponse.Output.Text
|
lastResponseText = aliResponse.Output.Choices[0].Message.StringContent()
|
||||||
jsonResponse, err := json.Marshal(response)
|
jsonResponse, err := json.Marshal(response)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
common.SysError("error marshalling stream response: " + err.Error())
|
common.SysError("error marshalling stream response: " + err.Error())
|
||||||
|
|||||||
@@ -1,5 +1,9 @@
|
|||||||
package ali
|
package ali
|
||||||
|
|
||||||
|
import (
|
||||||
|
"one-api/types"
|
||||||
|
)
|
||||||
|
|
||||||
type AliError struct {
|
type AliError struct {
|
||||||
Code string `json:"code"`
|
Code string `json:"code"`
|
||||||
Message string `json:"message"`
|
Message string `json:"message"`
|
||||||
@@ -13,10 +17,15 @@ type AliUsage struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type AliMessage struct {
|
type AliMessage struct {
|
||||||
Content string `json:"content"`
|
Content any `json:"content"`
|
||||||
Role string `json:"role"`
|
Role string `json:"role"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type AliMessagePart struct {
|
||||||
|
Text string `json:"text,omitempty"`
|
||||||
|
Image string `json:"image,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
type AliInput struct {
|
type AliInput struct {
|
||||||
// Prompt string `json:"prompt"`
|
// Prompt string `json:"prompt"`
|
||||||
Messages []AliMessage `json:"messages"`
|
Messages []AliMessage `json:"messages"`
|
||||||
@@ -28,6 +37,7 @@ type AliParameters struct {
|
|||||||
Seed uint64 `json:"seed,omitempty"`
|
Seed uint64 `json:"seed,omitempty"`
|
||||||
EnableSearch bool `json:"enable_search,omitempty"`
|
EnableSearch bool `json:"enable_search,omitempty"`
|
||||||
IncrementalOutput bool `json:"incremental_output,omitempty"`
|
IncrementalOutput bool `json:"incremental_output,omitempty"`
|
||||||
|
ResultFormat string `json:"result_format,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type AliChatRequest struct {
|
type AliChatRequest struct {
|
||||||
@@ -36,9 +46,25 @@ type AliChatRequest struct {
|
|||||||
Parameters AliParameters `json:"parameters,omitempty"`
|
Parameters AliParameters `json:"parameters,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type AliChoice struct {
|
||||||
|
FinishReason string `json:"finish_reason"`
|
||||||
|
Message types.ChatCompletionMessage `json:"message"`
|
||||||
|
}
|
||||||
|
|
||||||
type AliOutput struct {
|
type AliOutput struct {
|
||||||
Text string `json:"text"`
|
Choices []types.ChatCompletionChoice `json:"choices"`
|
||||||
FinishReason string `json:"finish_reason"`
|
}
|
||||||
|
|
||||||
|
func (o *AliOutput) ToChatCompletionChoices() []types.ChatCompletionChoice {
|
||||||
|
for i := range o.Choices {
|
||||||
|
_, ok := o.Choices[i].Message.Content.(string)
|
||||||
|
if ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
o.Choices[i].Message.Content = o.Choices[i].Message.ParseContent()
|
||||||
|
}
|
||||||
|
return o.Choices
|
||||||
}
|
}
|
||||||
|
|
||||||
type AliChatResponse struct {
|
type AliChatResponse struct {
|
||||||
|
|||||||
30
providers/baichuan/base.go
Normal file
30
providers/baichuan/base.go
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
package baichuan
|
||||||
|
|
||||||
|
import (
|
||||||
|
"one-api/providers/base"
|
||||||
|
"one-api/providers/openai"
|
||||||
|
|
||||||
|
"github.com/gin-gonic/gin"
|
||||||
|
)
|
||||||
|
|
||||||
|
// 定义供应商工厂
|
||||||
|
type BaichuanProviderFactory struct{}
|
||||||
|
|
||||||
|
// 创建 BaichuanProvider
|
||||||
|
// https://platform.baichuan-ai.com/docs/api
|
||||||
|
func (f BaichuanProviderFactory) Create(c *gin.Context) base.ProviderInterface {
|
||||||
|
return &BaichuanProvider{
|
||||||
|
OpenAIProvider: openai.OpenAIProvider{
|
||||||
|
BaseProvider: base.BaseProvider{
|
||||||
|
BaseURL: "https://api.baichuan-ai.com",
|
||||||
|
ChatCompletions: "/v1/chat/completions",
|
||||||
|
Embeddings: "/v1/embeddings",
|
||||||
|
Context: c,
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
type BaichuanProvider struct {
|
||||||
|
openai.OpenAIProvider
|
||||||
|
}
|
||||||
100
providers/baichuan/chat.go
Normal file
100
providers/baichuan/chat.go
Normal file
@@ -0,0 +1,100 @@
|
|||||||
|
package baichuan
|
||||||
|
|
||||||
|
import (
|
||||||
|
"net/http"
|
||||||
|
"one-api/common"
|
||||||
|
"one-api/providers/openai"
|
||||||
|
"one-api/types"
|
||||||
|
"strings"
|
||||||
|
)
|
||||||
|
|
||||||
|
func (baichuanResponse *BaichuanChatResponse) ResponseHandler(resp *http.Response) (OpenAIResponse any, errWithCode *types.OpenAIErrorWithStatusCode) {
|
||||||
|
if baichuanResponse.Error.Message != "" {
|
||||||
|
errWithCode = &types.OpenAIErrorWithStatusCode{
|
||||||
|
OpenAIError: baichuanResponse.Error,
|
||||||
|
StatusCode: resp.StatusCode,
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
OpenAIResponse = types.ChatCompletionResponse{
|
||||||
|
ID: baichuanResponse.ID,
|
||||||
|
Object: baichuanResponse.Object,
|
||||||
|
Created: baichuanResponse.Created,
|
||||||
|
Model: baichuanResponse.Model,
|
||||||
|
Choices: baichuanResponse.Choices,
|
||||||
|
Usage: baichuanResponse.Usage,
|
||||||
|
}
|
||||||
|
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// 获取聊天请求体
|
||||||
|
func (p *BaichuanProvider) getChatRequestBody(request *types.ChatCompletionRequest) *BaichuanChatRequest {
|
||||||
|
messages := make([]BaichuanMessage, 0, len(request.Messages))
|
||||||
|
for i := 0; i < len(request.Messages); i++ {
|
||||||
|
message := request.Messages[i]
|
||||||
|
if message.Role == "system" || message.Role == "assistant" {
|
||||||
|
message.Role = "assistant"
|
||||||
|
} else {
|
||||||
|
message.Role = "user"
|
||||||
|
}
|
||||||
|
messages = append(messages, BaichuanMessage{
|
||||||
|
Content: message.StringContent(),
|
||||||
|
Role: strings.ToLower(message.Role),
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
return &BaichuanChatRequest{
|
||||||
|
Model: request.Model,
|
||||||
|
Messages: messages,
|
||||||
|
Stream: request.Stream,
|
||||||
|
Temperature: request.Temperature,
|
||||||
|
TopP: request.TopP,
|
||||||
|
TopK: request.N,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// 聊天
|
||||||
|
func (p *BaichuanProvider) ChatAction(request *types.ChatCompletionRequest, isModelMapped bool, promptTokens int) (usage *types.Usage, errWithCode *types.OpenAIErrorWithStatusCode) {
|
||||||
|
|
||||||
|
requestBody := p.getChatRequestBody(request)
|
||||||
|
|
||||||
|
fullRequestURL := p.GetFullRequestURL(p.ChatCompletions, request.Model)
|
||||||
|
headers := p.GetRequestHeaders()
|
||||||
|
if request.Stream {
|
||||||
|
headers["Accept"] = "text/event-stream"
|
||||||
|
}
|
||||||
|
|
||||||
|
client := common.NewClient()
|
||||||
|
req, err := client.NewRequest(p.Context.Request.Method, fullRequestURL, common.WithBody(requestBody), common.WithHeader(headers))
|
||||||
|
if err != nil {
|
||||||
|
return nil, common.ErrorWrapper(err, "new_request_failed", http.StatusInternalServerError)
|
||||||
|
}
|
||||||
|
|
||||||
|
if request.Stream {
|
||||||
|
openAIProviderChatStreamResponse := &openai.OpenAIProviderChatStreamResponse{}
|
||||||
|
var textResponse string
|
||||||
|
errWithCode, textResponse = p.SendStreamRequest(req, openAIProviderChatStreamResponse)
|
||||||
|
if errWithCode != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
usage = &types.Usage{
|
||||||
|
PromptTokens: promptTokens,
|
||||||
|
CompletionTokens: common.CountTokenText(textResponse, request.Model),
|
||||||
|
TotalTokens: promptTokens + common.CountTokenText(textResponse, request.Model),
|
||||||
|
}
|
||||||
|
|
||||||
|
} else {
|
||||||
|
baichuanResponse := &BaichuanChatResponse{}
|
||||||
|
errWithCode = p.SendRequest(req, baichuanResponse, false)
|
||||||
|
if errWithCode != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
usage = baichuanResponse.Usage
|
||||||
|
}
|
||||||
|
return
|
||||||
|
}
|
||||||
36
providers/baichuan/type.go
Normal file
36
providers/baichuan/type.go
Normal file
@@ -0,0 +1,36 @@
|
|||||||
|
package baichuan
|
||||||
|
|
||||||
|
import "one-api/providers/openai"
|
||||||
|
|
||||||
|
type BaichuanMessage struct {
|
||||||
|
Role string `json:"role"`
|
||||||
|
Content string `json:"content"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type BaichuanKnowledgeBase struct {
|
||||||
|
Ids []string `json:"id"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type BaichuanChatRequest struct {
|
||||||
|
Model string `json:"model"`
|
||||||
|
Messages []BaichuanMessage `json:"messages"`
|
||||||
|
Stream bool `json:"stream,omitempty"`
|
||||||
|
Temperature float64 `json:"temperature,omitempty"`
|
||||||
|
TopP float64 `json:"top_p,omitempty"`
|
||||||
|
TopK int `json:"top_k,omitempty"`
|
||||||
|
WithSearchEnhance bool `json:"with_search_enhance,omitempty"`
|
||||||
|
KnowledgeBase BaichuanKnowledgeBase `json:"knowledge_base,omitempty"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type BaichuanKnowledgeBaseResponse struct {
|
||||||
|
Cites []struct {
|
||||||
|
Title string `json:"title"`
|
||||||
|
Content string `json:"content"`
|
||||||
|
FileId string `json:"file_id"`
|
||||||
|
} `json:"cites"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type BaichuanChatResponse struct {
|
||||||
|
openai.OpenAIProviderChatResponse
|
||||||
|
KnowledgeBase BaichuanKnowledgeBaseResponse `json:"knowledge_base,omitempty"`
|
||||||
|
}
|
||||||
@@ -60,27 +60,27 @@ func (response *GeminiChatResponse) ResponseHandler(resp *http.Response) (OpenAI
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Setting safety to the lowest possible values since Gemini is already powerless enough
|
// Setting safety to the lowest possible values since Gemini is already powerless enough
|
||||||
func (p *GeminiProvider) getChatRequestBody(request *types.ChatCompletionRequest) (requestBody *GeminiChatRequest) {
|
func (p *GeminiProvider) getChatRequestBody(request *types.ChatCompletionRequest) (requestBody *GeminiChatRequest, errWithCode *types.OpenAIErrorWithStatusCode) {
|
||||||
geminiRequest := GeminiChatRequest{
|
geminiRequest := GeminiChatRequest{
|
||||||
Contents: make([]GeminiChatContent, 0, len(request.Messages)),
|
Contents: make([]GeminiChatContent, 0, len(request.Messages)),
|
||||||
//SafetySettings: []GeminiChatSafetySettings{
|
SafetySettings: []GeminiChatSafetySettings{
|
||||||
// {
|
{
|
||||||
// Category: "HARM_CATEGORY_HARASSMENT",
|
Category: "HARM_CATEGORY_HARASSMENT",
|
||||||
// Threshold: "BLOCK_ONLY_HIGH",
|
Threshold: "BLOCK_NONE",
|
||||||
// },
|
},
|
||||||
// {
|
{
|
||||||
// Category: "HARM_CATEGORY_HATE_SPEECH",
|
Category: "HARM_CATEGORY_HATE_SPEECH",
|
||||||
// Threshold: "BLOCK_ONLY_HIGH",
|
Threshold: "BLOCK_NONE",
|
||||||
// },
|
},
|
||||||
// {
|
{
|
||||||
// Category: "HARM_CATEGORY_SEXUALLY_EXPLICIT",
|
Category: "HARM_CATEGORY_SEXUALLY_EXPLICIT",
|
||||||
// Threshold: "BLOCK_ONLY_HIGH",
|
Threshold: "BLOCK_NONE",
|
||||||
// },
|
},
|
||||||
// {
|
{
|
||||||
// Category: "HARM_CATEGORY_DANGEROUS_CONTENT",
|
Category: "HARM_CATEGORY_DANGEROUS_CONTENT",
|
||||||
// Threshold: "BLOCK_ONLY_HIGH",
|
Threshold: "BLOCK_NONE",
|
||||||
// },
|
},
|
||||||
//},
|
},
|
||||||
GenerationConfig: GeminiChatGenerationConfig{
|
GenerationConfig: GeminiChatGenerationConfig{
|
||||||
Temperature: request.Temperature,
|
Temperature: request.Temperature,
|
||||||
TopP: request.TopP,
|
TopP: request.TopP,
|
||||||
@@ -118,7 +118,10 @@ func (p *GeminiProvider) getChatRequestBody(request *types.ChatCompletionRequest
|
|||||||
if imageNum > GeminiVisionMaxImageNum {
|
if imageNum > GeminiVisionMaxImageNum {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
mimeType, data, _ := image.GetImageFromUrl(part.ImageURL.URL)
|
mimeType, data, err := image.GetImageFromUrl(part.ImageURL.URL)
|
||||||
|
if err != nil {
|
||||||
|
return nil, common.ErrorWrapper(err, "image_url_invalid", http.StatusBadRequest)
|
||||||
|
}
|
||||||
parts = append(parts, GeminiPart{
|
parts = append(parts, GeminiPart{
|
||||||
InlineData: &GeminiInlineData{
|
InlineData: &GeminiInlineData{
|
||||||
MimeType: mimeType,
|
MimeType: mimeType,
|
||||||
@@ -154,11 +157,14 @@ func (p *GeminiProvider) getChatRequestBody(request *types.ChatCompletionRequest
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return &geminiRequest
|
return &geminiRequest, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (p *GeminiProvider) ChatAction(request *types.ChatCompletionRequest, isModelMapped bool, promptTokens int) (usage *types.Usage, errWithCode *types.OpenAIErrorWithStatusCode) {
|
func (p *GeminiProvider) ChatAction(request *types.ChatCompletionRequest, isModelMapped bool, promptTokens int) (usage *types.Usage, errWithCode *types.OpenAIErrorWithStatusCode) {
|
||||||
requestBody := p.getChatRequestBody(request)
|
requestBody, errWithCode := p.getChatRequestBody(request)
|
||||||
|
if errWithCode != nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
fullRequestURL := p.GetFullRequestURL("generateContent", request.Model)
|
fullRequestURL := p.GetFullRequestURL("generateContent", request.Model)
|
||||||
headers := p.GetRequestHeaders()
|
headers := p.GetRequestHeaders()
|
||||||
if request.Stream {
|
if request.Stream {
|
||||||
|
|||||||
@@ -108,7 +108,7 @@ func (p *OpenAIProvider) GetRequestBody(request any, isModelMapped bool) (reques
|
|||||||
}
|
}
|
||||||
|
|
||||||
// 发送流式请求
|
// 发送流式请求
|
||||||
func (p *OpenAIProvider) sendStreamRequest(req *http.Request, response OpenAIProviderStreamResponseHandler) (openAIErrorWithStatusCode *types.OpenAIErrorWithStatusCode, responseText string) {
|
func (p *OpenAIProvider) SendStreamRequest(req *http.Request, response OpenAIProviderStreamResponseHandler) (openAIErrorWithStatusCode *types.OpenAIErrorWithStatusCode, responseText string) {
|
||||||
defer req.Body.Close()
|
defer req.Body.Close()
|
||||||
|
|
||||||
client := common.GetHttpClient(p.Channel.Proxy)
|
client := common.GetHttpClient(p.Channel.Proxy)
|
||||||
|
|||||||
@@ -46,7 +46,7 @@ func (p *OpenAIProvider) ChatAction(request *types.ChatCompletionRequest, isMode
|
|||||||
if request.Stream {
|
if request.Stream {
|
||||||
openAIProviderChatStreamResponse := &OpenAIProviderChatStreamResponse{}
|
openAIProviderChatStreamResponse := &OpenAIProviderChatStreamResponse{}
|
||||||
var textResponse string
|
var textResponse string
|
||||||
errWithCode, textResponse = p.sendStreamRequest(req, openAIProviderChatStreamResponse)
|
errWithCode, textResponse = p.SendStreamRequest(req, openAIProviderChatStreamResponse)
|
||||||
if errWithCode != nil {
|
if errWithCode != nil {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -47,7 +47,7 @@ func (p *OpenAIProvider) CompleteAction(request *types.CompletionRequest, isMode
|
|||||||
if request.Stream {
|
if request.Stream {
|
||||||
// TODO
|
// TODO
|
||||||
var textResponse string
|
var textResponse string
|
||||||
errWithCode, textResponse = p.sendStreamRequest(req, openAIProviderCompletionResponse)
|
errWithCode, textResponse = p.SendStreamRequest(req, openAIProviderCompletionResponse)
|
||||||
if errWithCode != nil {
|
if errWithCode != nil {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -10,6 +10,7 @@ import (
|
|||||||
"one-api/providers/api2gpt"
|
"one-api/providers/api2gpt"
|
||||||
"one-api/providers/azure"
|
"one-api/providers/azure"
|
||||||
azurespeech "one-api/providers/azureSpeech"
|
azurespeech "one-api/providers/azureSpeech"
|
||||||
|
"one-api/providers/baichuan"
|
||||||
"one-api/providers/baidu"
|
"one-api/providers/baidu"
|
||||||
"one-api/providers/base"
|
"one-api/providers/base"
|
||||||
"one-api/providers/claude"
|
"one-api/providers/claude"
|
||||||
@@ -52,6 +53,7 @@ func init() {
|
|||||||
providerFactories[common.ChannelTypeAPI2GPT] = api2gpt.Api2gptProviderFactory{}
|
providerFactories[common.ChannelTypeAPI2GPT] = api2gpt.Api2gptProviderFactory{}
|
||||||
providerFactories[common.ChannelTypeAzureSpeech] = azurespeech.AzureSpeechProviderFactory{}
|
providerFactories[common.ChannelTypeAzureSpeech] = azurespeech.AzureSpeechProviderFactory{}
|
||||||
providerFactories[common.ChannelTypeGemini] = gemini.GeminiProviderFactory{}
|
providerFactories[common.ChannelTypeGemini] = gemini.GeminiProviderFactory{}
|
||||||
|
providerFactories[common.ChannelTypeBaichuan] = baichuan.BaichuanProviderFactory{}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -58,7 +58,7 @@ func (p *XunfeiProvider) getXunfeiAuthUrl(apiKey string, apiSecret string) (stri
|
|||||||
query := p.Context.Request.URL.Query()
|
query := p.Context.Request.URL.Query()
|
||||||
apiVersion := query.Get("api-version")
|
apiVersion := query.Get("api-version")
|
||||||
if apiVersion == "" {
|
if apiVersion == "" {
|
||||||
apiVersion = p.Channel.Key
|
apiVersion = p.Channel.Other
|
||||||
}
|
}
|
||||||
if apiVersion == "" {
|
if apiVersion == "" {
|
||||||
apiVersion = "v1.1"
|
apiVersion = "v1.1"
|
||||||
|
|||||||
@@ -67,7 +67,7 @@ func SetApiRouter(router *gin.Engine) {
|
|||||||
{
|
{
|
||||||
channelRoute.GET("/", controller.GetAllChannels)
|
channelRoute.GET("/", controller.GetAllChannels)
|
||||||
channelRoute.GET("/search", controller.SearchChannels)
|
channelRoute.GET("/search", controller.SearchChannels)
|
||||||
channelRoute.GET("/models", controller.ListModels)
|
channelRoute.GET("/models", controller.ListModelsForAdmin)
|
||||||
channelRoute.GET("/:id", controller.GetChannel)
|
channelRoute.GET("/:id", controller.GetChannel)
|
||||||
channelRoute.GET("/test", controller.TestAllChannels)
|
channelRoute.GET("/test", controller.TestAllChannels)
|
||||||
channelRoute.GET("/test/:id", controller.TestChannel)
|
channelRoute.GET("/test/:id", controller.TestChannel)
|
||||||
|
|||||||
@@ -11,7 +11,7 @@ func SetRelayRouter(router *gin.Engine) {
|
|||||||
router.Use(middleware.CORS())
|
router.Use(middleware.CORS())
|
||||||
// https://platform.openai.com/docs/api-reference/introduction
|
// https://platform.openai.com/docs/api-reference/introduction
|
||||||
modelsRouter := router.Group("/v1/models")
|
modelsRouter := router.Group("/v1/models")
|
||||||
modelsRouter.Use(middleware.TokenAuth())
|
modelsRouter.Use(middleware.TokenAuth(), middleware.Distribute())
|
||||||
{
|
{
|
||||||
modelsRouter.GET("", controller.ListModels)
|
modelsRouter.GET("", controller.ListModels)
|
||||||
modelsRouter.GET("/:model", controller.RetrieveModel)
|
modelsRouter.GET("/:model", controller.RetrieveModel)
|
||||||
|
|||||||
@@ -27,11 +27,11 @@ func (m ChatCompletionMessage) StringContent() string {
|
|||||||
if !ok {
|
if !ok {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if contentMap["type"] == "text" {
|
|
||||||
if subStr, ok := contentMap["text"].(string); ok {
|
if subStr, ok := contentMap["text"].(string); ok && subStr != "" {
|
||||||
contentStr += subStr
|
contentStr += subStr
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
return contentStr
|
return contentStr
|
||||||
}
|
}
|
||||||
@@ -55,23 +55,26 @@ func (m ChatCompletionMessage) ParseContent() []ChatMessagePart {
|
|||||||
if !ok {
|
if !ok {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
switch contentMap["type"] {
|
|
||||||
case ContentTypeText:
|
if subStr, ok := contentMap["text"].(string); ok && subStr != "" {
|
||||||
if subStr, ok := contentMap["text"].(string); ok {
|
contentList = append(contentList, ChatMessagePart{
|
||||||
contentList = append(contentList, ChatMessagePart{
|
Type: ContentTypeText,
|
||||||
Type: ContentTypeText,
|
Text: subStr,
|
||||||
Text: subStr,
|
})
|
||||||
})
|
} else if subObj, ok := contentMap["image_url"].(map[string]any); ok {
|
||||||
}
|
contentList = append(contentList, ChatMessagePart{
|
||||||
case ContentTypeImageURL:
|
Type: ContentTypeImageURL,
|
||||||
if subObj, ok := contentMap["image_url"].(map[string]any); ok {
|
ImageURL: &ChatMessageImageURL{
|
||||||
contentList = append(contentList, ChatMessagePart{
|
URL: subObj["url"].(string),
|
||||||
Type: ContentTypeImageURL,
|
},
|
||||||
ImageURL: &ChatMessageImageURL{
|
})
|
||||||
URL: subObj["url"].(string),
|
} else if subObj, ok := contentMap["image"].(string); ok {
|
||||||
},
|
contentList = append(contentList, ChatMessagePart{
|
||||||
})
|
Type: ContentTypeImageURL,
|
||||||
}
|
ImageURL: &ChatMessageImageURL{
|
||||||
|
URL: subObj,
|
||||||
|
},
|
||||||
|
})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return contentList
|
return contentList
|
||||||
|
|||||||
@@ -65,6 +65,12 @@ export const CHANNEL_OPTIONS = {
|
|||||||
value: 23,
|
value: 23,
|
||||||
color: 'default'
|
color: 'default'
|
||||||
},
|
},
|
||||||
|
26: {
|
||||||
|
key: 26,
|
||||||
|
text: '百川',
|
||||||
|
value: 26,
|
||||||
|
color: 'orange'
|
||||||
|
},
|
||||||
24: {
|
24: {
|
||||||
key: 24,
|
key: 24,
|
||||||
text: 'Azure Speech',
|
text: 'Azure Speech',
|
||||||
|
|||||||
@@ -36,6 +36,7 @@ const validationSchema = Yup.object().shape({
|
|||||||
key: Yup.string().when('is_edit', { is: false, then: Yup.string().required('密钥 不能为空') }),
|
key: Yup.string().when('is_edit', { is: false, then: Yup.string().required('密钥 不能为空') }),
|
||||||
other: Yup.string(),
|
other: Yup.string(),
|
||||||
proxy: Yup.string(),
|
proxy: Yup.string(),
|
||||||
|
test_model: Yup.string(),
|
||||||
models: Yup.array().min(1, '模型 不能为空'),
|
models: Yup.array().min(1, '模型 不能为空'),
|
||||||
groups: Yup.array().min(1, '用户组 不能为空'),
|
groups: Yup.array().min(1, '用户组 不能为空'),
|
||||||
base_url: Yup.string().when('type', {
|
base_url: Yup.string().when('type', {
|
||||||
@@ -90,7 +91,7 @@ const EditModal = ({ open, channelId, onCancel, onOk }) => {
|
|||||||
if (newInput) {
|
if (newInput) {
|
||||||
Object.keys(newInput).forEach((key) => {
|
Object.keys(newInput).forEach((key) => {
|
||||||
if (
|
if (
|
||||||
(!Array.isArray(values[key]) && values[key] !== null && values[key] !== undefined) ||
|
(!Array.isArray(values[key]) && values[key] !== null && values[key] !== undefined && values[key] !== '') ||
|
||||||
(Array.isArray(values[key]) && values[key].length > 0)
|
(Array.isArray(values[key]) && values[key].length > 0)
|
||||||
) {
|
) {
|
||||||
return;
|
return;
|
||||||
@@ -464,6 +465,29 @@ const EditModal = ({ open, channelId, onCancel, onOk }) => {
|
|||||||
<FormHelperText id="helper-tex-channel-proxy-label"> {inputPrompt.proxy} </FormHelperText>
|
<FormHelperText id="helper-tex-channel-proxy-label"> {inputPrompt.proxy} </FormHelperText>
|
||||||
)}
|
)}
|
||||||
</FormControl>
|
</FormControl>
|
||||||
|
{inputPrompt.test_model && (
|
||||||
|
<FormControl fullWidth error={Boolean(touched.test_model && errors.test_model)} sx={{ ...theme.typography.otherInput }}>
|
||||||
|
<InputLabel htmlFor="channel-test_model-label">{inputLabel.test_model}</InputLabel>
|
||||||
|
<OutlinedInput
|
||||||
|
id="channel-test_model-label"
|
||||||
|
label={inputLabel.test_model}
|
||||||
|
type="text"
|
||||||
|
value={values.test_model}
|
||||||
|
name="test_model"
|
||||||
|
onBlur={handleBlur}
|
||||||
|
onChange={handleChange}
|
||||||
|
inputProps={{}}
|
||||||
|
aria-describedby="helper-text-channel-test_model-label"
|
||||||
|
/>
|
||||||
|
{touched.test_model && errors.test_model ? (
|
||||||
|
<FormHelperText error id="helper-tex-channel-test_model-label">
|
||||||
|
{errors.test_model}
|
||||||
|
</FormHelperText>
|
||||||
|
) : (
|
||||||
|
<FormHelperText id="helper-tex-channel-test_model-label"> {inputPrompt.test_model} </FormHelperText>
|
||||||
|
)}
|
||||||
|
</FormControl>
|
||||||
|
)}
|
||||||
<DialogActions>
|
<DialogActions>
|
||||||
<Button onClick={onCancel}>取消</Button>
|
<Button onClick={onCancel}>取消</Button>
|
||||||
<Button disableElevation disabled={isSubmitting} type="submit" variant="contained" color="primary">
|
<Button disableElevation disabled={isSubmitting} type="submit" variant="contained" color="primary">
|
||||||
|
|||||||
@@ -116,7 +116,7 @@ export default function ChannelPage() {
|
|||||||
if (success) {
|
if (success) {
|
||||||
showSuccess('操作成功完成!');
|
showSuccess('操作成功完成!');
|
||||||
if (action === 'delete') {
|
if (action === 'delete') {
|
||||||
await loadChannels(0);
|
await handleRefresh();
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
showError(message);
|
showError(message);
|
||||||
@@ -127,9 +127,7 @@ export default function ChannelPage() {
|
|||||||
|
|
||||||
// 处理刷新
|
// 处理刷新
|
||||||
const handleRefresh = async () => {
|
const handleRefresh = async () => {
|
||||||
await loadChannels(0);
|
await loadChannels(activePage);
|
||||||
setActivePage(0);
|
|
||||||
setSearchKeyword('');
|
|
||||||
};
|
};
|
||||||
|
|
||||||
// 处理测试所有启用渠道
|
// 处理测试所有启用渠道
|
||||||
|
|||||||
@@ -6,6 +6,7 @@ const defaultConfig = {
|
|||||||
base_url: '',
|
base_url: '',
|
||||||
other: '',
|
other: '',
|
||||||
proxy: '',
|
proxy: '',
|
||||||
|
test_model: '',
|
||||||
model_mapping: '',
|
model_mapping: '',
|
||||||
models: [],
|
models: [],
|
||||||
groups: ['default']
|
groups: ['default']
|
||||||
@@ -17,6 +18,7 @@ const defaultConfig = {
|
|||||||
key: '密钥',
|
key: '密钥',
|
||||||
other: '其他参数',
|
other: '其他参数',
|
||||||
proxy: '代理地址',
|
proxy: '代理地址',
|
||||||
|
test_model: '测速模型',
|
||||||
models: '模型',
|
models: '模型',
|
||||||
model_mapping: '模型映射关系',
|
model_mapping: '模型映射关系',
|
||||||
groups: '用户组'
|
groups: '用户组'
|
||||||
@@ -28,6 +30,7 @@ const defaultConfig = {
|
|||||||
key: '请输入渠道对应的鉴权密钥',
|
key: '请输入渠道对应的鉴权密钥',
|
||||||
other: '',
|
other: '',
|
||||||
proxy: '单独设置代理地址,支持http和socks5,例如:http://127.0.0.1:1080',
|
proxy: '单独设置代理地址,支持http和socks5,例如:http://127.0.0.1:1080',
|
||||||
|
test_model: '用于测试使用的模型,为空时无法测速,如:gpt-3.5-turbo',
|
||||||
models: '请选择该渠道所支持的模型',
|
models: '请选择该渠道所支持的模型',
|
||||||
model_mapping:
|
model_mapping:
|
||||||
'请输入要修改的模型映射关系,格式为:api请求模型ID:实际转发给渠道的模型ID,使用JSON数组表示,例如:{"gpt-3.5": "gpt-35"}',
|
'请输入要修改的模型映射关系,格式为:api请求模型ID:实际转发给渠道的模型ID,使用JSON数组表示,例如:{"gpt-3.5": "gpt-35"}',
|
||||||
@@ -48,17 +51,20 @@ const typeConfig = {
|
|||||||
},
|
},
|
||||||
11: {
|
11: {
|
||||||
input: {
|
input: {
|
||||||
models: ['PaLM-2']
|
models: ['PaLM-2'],
|
||||||
|
test_model: 'PaLM-2'
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
14: {
|
14: {
|
||||||
input: {
|
input: {
|
||||||
models: ['claude-instant-1', 'claude-2', 'claude-2.0', 'claude-2.1']
|
models: ['claude-instant-1', 'claude-2', 'claude-2.0', 'claude-2.1'],
|
||||||
|
test_model: 'claude-2'
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
15: {
|
15: {
|
||||||
input: {
|
input: {
|
||||||
models: ['ERNIE-Bot', 'ERNIE-Bot-turbo', 'ERNIE-Bot-4', 'Embedding-V1']
|
models: ['ERNIE-Bot', 'ERNIE-Bot-turbo', 'ERNIE-Bot-4', 'Embedding-V1'],
|
||||||
|
test_model: 'ERNIE-Bot'
|
||||||
},
|
},
|
||||||
prompt: {
|
prompt: {
|
||||||
key: '按照如下格式输入:APIKey|SecretKey'
|
key: '按照如下格式输入:APIKey|SecretKey'
|
||||||
@@ -66,7 +72,8 @@ const typeConfig = {
|
|||||||
},
|
},
|
||||||
16: {
|
16: {
|
||||||
input: {
|
input: {
|
||||||
models: ['chatglm_turbo', 'chatglm_pro', 'chatglm_std', 'chatglm_lite']
|
models: ['chatglm_turbo', 'chatglm_pro', 'chatglm_std', 'chatglm_lite'],
|
||||||
|
test_model: 'chatglm_lite'
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
17: {
|
17: {
|
||||||
@@ -84,7 +91,8 @@ const typeConfig = {
|
|||||||
'qwen-plus-internet',
|
'qwen-plus-internet',
|
||||||
'qwen-max-internet',
|
'qwen-max-internet',
|
||||||
'qwen-max-longcontext-internet'
|
'qwen-max-longcontext-internet'
|
||||||
]
|
],
|
||||||
|
test_model: 'qwen-turbo'
|
||||||
},
|
},
|
||||||
prompt: {
|
prompt: {
|
||||||
other: '请输入插件参数,即 X-DashScope-Plugin 请求头的取值'
|
other: '请输入插件参数,即 X-DashScope-Plugin 请求头的取值'
|
||||||
@@ -104,7 +112,8 @@ const typeConfig = {
|
|||||||
},
|
},
|
||||||
19: {
|
19: {
|
||||||
input: {
|
input: {
|
||||||
models: ['360GPT_S2_V9', 'embedding-bert-512-v1', 'embedding_s1_v1', 'semantic_similarity_s1_v1']
|
models: ['360GPT_S2_V9', 'embedding-bert-512-v1', 'embedding_s1_v1', 'semantic_similarity_s1_v1'],
|
||||||
|
test_model: '360GPT_S2_V9'
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
22: {
|
22: {
|
||||||
@@ -114,7 +123,8 @@ const typeConfig = {
|
|||||||
},
|
},
|
||||||
23: {
|
23: {
|
||||||
input: {
|
input: {
|
||||||
models: ['hunyuan']
|
models: ['hunyuan'],
|
||||||
|
test_model: 'hunyuan'
|
||||||
},
|
},
|
||||||
prompt: {
|
prompt: {
|
||||||
key: '按照如下格式输入:AppId|SecretId|SecretKey'
|
key: '按照如下格式输入:AppId|SecretId|SecretKey'
|
||||||
@@ -125,11 +135,26 @@ const typeConfig = {
|
|||||||
other: '版本号'
|
other: '版本号'
|
||||||
},
|
},
|
||||||
input: {
|
input: {
|
||||||
models: ['gemini-pro']
|
models: ['gemini-pro', 'gemini-pro-vision'],
|
||||||
|
test_model: 'gemini-pro'
|
||||||
},
|
},
|
||||||
prompt: {
|
prompt: {
|
||||||
other: '请输入版本号,例如:v1'
|
other: '请输入版本号,例如:v1'
|
||||||
}
|
}
|
||||||
|
},
|
||||||
|
26: {
|
||||||
|
input: {
|
||||||
|
models: ['Baichuan2-Turbo', 'Baichuan2-Turbo-192k', 'Baichuan2-53B', 'Baichuan-Text-Embedding'],
|
||||||
|
test_model: 'Baichuan2-Turbo'
|
||||||
|
}
|
||||||
|
},
|
||||||
|
24: {
|
||||||
|
input: {
|
||||||
|
models: ['tts-1', 'tts-1-hd']
|
||||||
|
},
|
||||||
|
prompt: {
|
||||||
|
test_model: ''
|
||||||
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|||||||
@@ -108,7 +108,7 @@ export default function Token() {
|
|||||||
if (success) {
|
if (success) {
|
||||||
showSuccess('操作成功完成!');
|
showSuccess('操作成功完成!');
|
||||||
if (action === 'delete') {
|
if (action === 'delete') {
|
||||||
await loadTokens(0);
|
await handleRefresh();
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
showError(message);
|
showError(message);
|
||||||
@@ -119,9 +119,7 @@ export default function Token() {
|
|||||||
|
|
||||||
// 处理刷新
|
// 处理刷新
|
||||||
const handleRefresh = async () => {
|
const handleRefresh = async () => {
|
||||||
await loadTokens(0);
|
await loadTokens(activePage);
|
||||||
setActivePage(0);
|
|
||||||
setSearchKeyword('');
|
|
||||||
};
|
};
|
||||||
|
|
||||||
const handleOpenModal = (tokenId) => {
|
const handleOpenModal = (tokenId) => {
|
||||||
|
|||||||
@@ -109,9 +109,7 @@ export default function Users() {
|
|||||||
|
|
||||||
// 处理刷新
|
// 处理刷新
|
||||||
const handleRefresh = async () => {
|
const handleRefresh = async () => {
|
||||||
await loadUsers(0);
|
await loadUsers(activePage);
|
||||||
setActivePage(0);
|
|
||||||
setSearchKeyword('');
|
|
||||||
};
|
};
|
||||||
|
|
||||||
const handleOpenModal = (userId) => {
|
const handleOpenModal = (userId) => {
|
||||||
|
|||||||
Reference in New Issue
Block a user