forked from lobehub/lobe-chat
-
Notifications
You must be signed in to change notification settings - Fork 2
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
Merge branch 'main' of https://github.com/lobehub/lobe-chat
- Loading branch information
Showing
53 changed files
with
933 additions
and
538 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,48 @@ | ||
--- | ||
title: Customizing Provider Model List in LobeChat for Deployment | ||
description: >- | ||
Learn how to customize the model list in LobeChat for deployment with the | ||
syntax and extension capabilities | ||
tags: | ||
- LobeChat | ||
- model customization | ||
- deployment | ||
- extension capabilities | ||
--- | ||
|
||
# Model List | ||
|
||
LobeChat supports customizing the model list during deployment. You can use `+` to add a model, `-` to hide a model, and use `model name=display name<extension configuration>` to customize the display name of a model, separated by English commas. The basic syntax is as follows: | ||
|
||
```shell | ||
id=displayName < maxToken:vision:fc:file > ,model2,model3 | ||
``` | ||
|
||
For example: `+qwen-7b-chat,+glm-6b,-gpt-3.5-turbo,gpt-4-0125-preview=gpt-4-turbo` | ||
|
||
In the above example, it adds `qwen-7b-chat` and `glm-6b` to the model list, removes `gpt-3.5-turbo` from the list, and displays the model name of `gpt-4-0125-preview` as `gpt-4-turbo`. If you want to disable all models first and then enable specific models, you can use `-all,+gpt-3.5-turbo`, which means only enabling `gpt-3.5-turbo`. | ||
|
||
## Extension Capabilities | ||
|
||
Considering the diversity of model capabilities, we started to add extension configuration in version `0.147.8`, with the following rules: | ||
|
||
```shell | ||
id=displayName<maxToken:vision:fc:file> | ||
``` | ||
|
||
The first value in angle brackets is designated as the `maxToken` for this model. The second value and beyond are the model's extension capabilities, separated by colons `:`, and the order is not important. | ||
|
||
Examples are as follows: | ||
|
||
- `chatglm-6b=ChatGLM 6B<4096>`: ChatGLM 6B, maximum context of 4k, no advanced capabilities; | ||
- `spark-v3.5=讯飞星火 v3.5<8192:fc>`: Xunfei Spark 3.5 model, maximum context of 8k, supports Function Call; | ||
- `gemini-pro-vision=Gemini Pro Vision<16000:vision>`: Google Vision model, maximum context of 16k, supports image recognition; | ||
- `gpt-4-all=ChatGPT Plus<128000:fc:vision:file>`, hacked version of ChatGPT Plus web, context of 128k, supports image recognition, Function Call, file upload. | ||
|
||
Currently supported extension capabilities are: | ||
|
||
| --- | Description | | ||
| -------- | -------------------------------------------------------- | | ||
| `fc` | Function Calling | | ||
| `vision` | Image Recognition | | ||
| `file` | File Upload (a bit hacky, not recommended for daily use) | |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,46 @@ | ||
--- | ||
title: LobeChat 自定义模型服务商模型列表及扩展能力配置 | ||
description: 了解如何在 LobeChat 中自定义模型列表以及扩展能力配置的基本语法和规则。 | ||
tags: | ||
- LobeChat | ||
- 自定义模型列表 | ||
- 扩展能力配置 | ||
- 模型展示名 | ||
- 模型能力 | ||
--- | ||
# Model List | ||
|
||
LobeChat 支持在部署时自定义模型列表,可以使用 `+` 增加一个模型,使用 `-` 来隐藏一个模型,使用 `模型名=展示名<扩展配置>` 来自定义模型的展示名,用英文逗号隔开。通过 `<>` 来添加扩展配置。基本语法如下: | ||
|
||
```shell | ||
id=displayName < maxToken:vision:fc:file > ,model2,model3 | ||
``` | ||
|
||
例如: `+qwen-7b-chat,+glm-6b,-gpt-3.5-turbo,gpt-4-0125-preview=gpt-4-turbo` | ||
|
||
上面示例表示增加 `qwen-7b-chat` 和 `glm-6b` 到模型列表,而从列表中删除 `gpt-3.5-turbo`,并将 `gpt-4-0125-preview` 模型名字展示为 `gpt-4-turbo`。如果你想先禁用所有模型,再启用指定模型,可以使用 `-all,+gpt-3.5-turbo`,则表示仅启用 `gpt-3.5-turbo`。 | ||
|
||
## 扩展能力 | ||
|
||
考虑到模型的能力多样性,我们在 `0.147.8` 版本开始增加扩展性配置,它的规则如下: | ||
|
||
```shell | ||
id=displayName<maxToken:vision:fc:file> | ||
``` | ||
|
||
尖括号第一个值约定为这个模型的 `maxToken` 。第二个及以后作为模型的扩展能力,能力与能力之间用冒号 `:` 作为分隔符,顺序不重要。 | ||
|
||
举例如下: | ||
|
||
- `chatglm-6b=ChatGLM 6B<4096>`:ChatGLM 6B,最大上下文 4k,没有高阶能力; | ||
- `spark-v3.5=讯飞星火 v3.5<8192:fc>`:讯飞星火 3.5 模型,最大上下文 8k,支持 Function Call; | ||
- `gemini-pro-vision=Gemini Pro Vision<16000:vision>`:Google 视觉模型,最大上下文 16k,支持图像识别; | ||
- `gpt-4-all=ChatGPT Plus<128000:fc:vision:file>`,hack 的 ChatGPT Plus 网页版,上下 128k ,支持图像识别、Function Call、文件上传 | ||
|
||
目前支持的扩展能力有: | ||
|
||
| --- | 描述 | | ||
| -------- | ------------------------------------ | | ||
| `fc` | 函数调用(function calling) | | ||
| `vision` | 视觉识别 | | ||
| `file` | 文件上传(比较hack,不建议日常使用) | |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.