Morler
commited on
Commit
·
d34888e
1
Parent(s):
ce5e150
Updated Model Information for Tongyi-Qianwen and ZHIPU-AI (#2003)
Browse files### What problem does this PR solve?
_Briefly describe what this PR aims to solve. Include background context
that will help reviewers understand the purpose of the PR._
### Type of change
- [X] Bug Fix (non-breaking change which fixes an issue)
- conf/llm_factories.json +38 -2
conf/llm_factories.json
CHANGED
@@ -104,6 +104,12 @@
|
|
104 |
"max_tokens": 2048,
|
105 |
"model_type": "embedding"
|
106 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
107 |
{
|
108 |
"llm_name": "paraformer-realtime-8k-v1",
|
109 |
"tags": "SPEECH2TEXT",
|
@@ -131,13 +137,37 @@
|
|
131 |
"status": "1",
|
132 |
"llm": [
|
133 |
{
|
134 |
-
"llm_name": "glm-
|
135 |
"tags": "LLM,CHAT,",
|
136 |
"max_tokens": 128000,
|
137 |
"model_type": "chat"
|
138 |
},
|
139 |
{
|
140 |
-
"llm_name": "glm-4",
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
141 |
"tags": "LLM,CHAT,",
|
142 |
"max_tokens": 128000,
|
143 |
"model_type": "chat"
|
@@ -153,6 +183,12 @@
|
|
153 |
"tags": "TEXT EMBEDDING",
|
154 |
"max_tokens": 512,
|
155 |
"model_type": "embedding"
|
|
|
|
|
|
|
|
|
|
|
|
|
156 |
}
|
157 |
]
|
158 |
},
|
|
|
104 |
"max_tokens": 2048,
|
105 |
"model_type": "embedding"
|
106 |
},
|
107 |
+
{
|
108 |
+
"llm_name": "text-embedding-v3",
|
109 |
+
"tags": "TEXT EMBEDDING,2K",
|
110 |
+
"max_tokens": 2048,
|
111 |
+
"model_type": "embedding"
|
112 |
+
},
|
113 |
{
|
114 |
"llm_name": "paraformer-realtime-8k-v1",
|
115 |
"tags": "SPEECH2TEXT",
|
|
|
137 |
"status": "1",
|
138 |
"llm": [
|
139 |
{
|
140 |
+
"llm_name": "glm-4",
|
141 |
"tags": "LLM,CHAT,",
|
142 |
"max_tokens": 128000,
|
143 |
"model_type": "chat"
|
144 |
},
|
145 |
{
|
146 |
+
"llm_name": "glm-4-airx",
|
147 |
+
"tags": "LLM,CHAT,",
|
148 |
+
"max_tokens": 8000,
|
149 |
+
"model_type": "chat"
|
150 |
+
},
|
151 |
+
{
|
152 |
+
"llm_name": "glm-4-air",
|
153 |
+
"tags": "LLM,CHAT,",
|
154 |
+
"max_tokens": 128000,
|
155 |
+
"model_type": "chat"
|
156 |
+
},
|
157 |
+
{
|
158 |
+
"llm_name": "glm-4-flash",
|
159 |
+
"tags": "LLM,CHAT,",
|
160 |
+
"max_tokens": 128000,
|
161 |
+
"model_type": "chat"
|
162 |
+
},
|
163 |
+
{
|
164 |
+
"llm_name": "glm-4-long",
|
165 |
+
"tags": "LLM,CHAT,",
|
166 |
+
"max_tokens": 1000000,
|
167 |
+
"model_type": "chat"
|
168 |
+
},
|
169 |
+
{
|
170 |
+
"llm_name": "glm-3-turbo",
|
171 |
"tags": "LLM,CHAT,",
|
172 |
"max_tokens": 128000,
|
173 |
"model_type": "chat"
|
|
|
183 |
"tags": "TEXT EMBEDDING",
|
184 |
"max_tokens": 512,
|
185 |
"model_type": "embedding"
|
186 |
+
},
|
187 |
+
{
|
188 |
+
"llm_name": "embedding-3",
|
189 |
+
"tags": "TEXT EMBEDDING",
|
190 |
+
"max_tokens": 512,
|
191 |
+
"model_type": "embedding"
|
192 |
}
|
193 |
]
|
194 |
},
|