fix: migrate glm-4-flash to glm-4-flash-250414 (model deprecated by Zhipu)
Some checks failed
CI / Lint & TypeCheck (push) Has been cancelled
CI / Unit Tests (push) Has been cancelled
CI / Build Frontend (push) Has been cancelled
CI / Rust Check (push) Has been cancelled
CI / Security Scan (push) Has been cancelled
CI / E2E Tests (push) Has been cancelled
Some checks failed
CI / Lint & TypeCheck (push) Has been cancelled
CI / Unit Tests (push) Has been cancelled
CI / Build Frontend (push) Has been cancelled
CI / Rust Check (push) Has been cancelled
CI / Security Scan (push) Has been cancelled
CI / E2E Tests (push) Has been cancelled
Zhipu AI has deprecated glm-4-flash, causing 404 errors on all chat requests. Updated all references: - config: glm-4-flash → glm-4-flash-250414, added glm-z1-flash - frontend: defaultModel, conversationStore, ChatArea fallback, ModelsAPI
This commit is contained in:
@@ -25,12 +25,19 @@ max_output_tokens = 4096
|
|||||||
supports_streaming = true
|
supports_streaming = true
|
||||||
|
|
||||||
[[llm.providers.models]]
|
[[llm.providers.models]]
|
||||||
id = "glm-4-flash"
|
id = "glm-4-flash-250414"
|
||||||
alias = "GLM-4-Flash"
|
alias = "GLM-4-Flash (免费)"
|
||||||
context_window = 128000
|
context_window = 128000
|
||||||
max_output_tokens = 4096
|
max_output_tokens = 4096
|
||||||
supports_streaming = true
|
supports_streaming = true
|
||||||
|
|
||||||
|
[[llm.providers.models]]
|
||||||
|
id = "glm-z1-flash"
|
||||||
|
alias = "GLM-Z1-Flash (免费推理)"
|
||||||
|
context_window = 128000
|
||||||
|
max_output_tokens = 16384
|
||||||
|
supports_streaming = true
|
||||||
|
|
||||||
[[llm.providers.models]]
|
[[llm.providers.models]]
|
||||||
id = "glm-4v-plus"
|
id = "glm-4v-plus"
|
||||||
alias = "GLM-4V-Plus (视觉)"
|
alias = "GLM-4V-Plus (视觉)"
|
||||||
|
|||||||
@@ -129,7 +129,7 @@ retry_delay = "1s"
|
|||||||
|
|
||||||
[llm.aliases]
|
[llm.aliases]
|
||||||
# 智谱 GLM 模型 (使用正确的 API 模型 ID)
|
# 智谱 GLM 模型 (使用正确的 API 模型 ID)
|
||||||
"glm-4-flash" = "zhipu/glm-4-flash"
|
"glm-4-flash" = "zhipu/glm-4-flash-250414"
|
||||||
"glm-4-plus" = "zhipu/glm-4-plus"
|
"glm-4-plus" = "zhipu/glm-4-plus"
|
||||||
"glm-4.5" = "zhipu/glm-4.5"
|
"glm-4.5" = "zhipu/glm-4.5"
|
||||||
# 其他模型
|
# 其他模型
|
||||||
|
|||||||
@@ -85,7 +85,7 @@ export function ChatArea({ compact, onOpenDetail }: { compact?: boolean; onOpenD
|
|||||||
}
|
}
|
||||||
// Fallback: provide common models when no backend is connected
|
// Fallback: provide common models when no backend is connected
|
||||||
return [
|
return [
|
||||||
{ id: 'glm-4-flash', name: 'GLM-4 Flash', provider: 'zhipu' },
|
{ id: 'glm-4-flash-250414', name: 'GLM-4 Flash (免费)', provider: 'zhipu' },
|
||||||
{ id: 'gpt-4o', name: 'GPT-4o', provider: 'openai' },
|
{ id: 'gpt-4o', name: 'GPT-4o', provider: 'openai' },
|
||||||
{ id: 'gpt-4o-mini', name: 'GPT-4o Mini', provider: 'openai' },
|
{ id: 'gpt-4o-mini', name: 'GPT-4o Mini', provider: 'openai' },
|
||||||
{ id: 'deepseek-chat', name: 'DeepSeek V3', provider: 'deepseek' },
|
{ id: 'deepseek-chat', name: 'DeepSeek V3', provider: 'deepseek' },
|
||||||
@@ -575,7 +575,7 @@ export function ChatArea({ compact, onOpenDetail }: { compact?: boolean; onOpenD
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
function MessageBubble({ message, setInput }: { message: Message; setInput: (text: string) => void }) {
|
function MessageBubble({ message, setInput, onRetry }: { message: Message; setInput: (text: string) => void; onRetry?: () => void }) {
|
||||||
if (message.role === 'tool') {
|
if (message.role === 'tool') {
|
||||||
return null;
|
return null;
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -188,7 +188,7 @@ export function ModelsAPI() {
|
|||||||
// 表单状态
|
// 表单状态
|
||||||
const [formData, setFormData] = useState({
|
const [formData, setFormData] = useState({
|
||||||
provider: 'zhipu',
|
provider: 'zhipu',
|
||||||
modelId: 'glm-4-flash',
|
modelId: 'glm-4-flash-250414',
|
||||||
displayName: '',
|
displayName: '',
|
||||||
apiKey: '',
|
apiKey: '',
|
||||||
apiProtocol: 'openai' as 'openai' | 'anthropic' | 'custom',
|
apiProtocol: 'openai' as 'openai' | 'anthropic' | 'custom',
|
||||||
@@ -678,11 +678,11 @@ export function ModelsAPI() {
|
|||||||
type="text"
|
type="text"
|
||||||
value={formData.modelId}
|
value={formData.modelId}
|
||||||
onChange={(e) => setFormData({ ...formData, modelId: e.target.value })}
|
onChange={(e) => setFormData({ ...formData, modelId: e.target.value })}
|
||||||
placeholder="如:glm-4-flash, glm-4-plus, glm-4.5"
|
placeholder="如:glm-4-flash-250414, glm-4-plus, glm-4.7"
|
||||||
className="w-full px-3 py-2 border border-gray-200 dark:border-gray-600 rounded-lg text-sm bg-white dark:bg-gray-700 text-gray-900 dark:text-white placeholder-gray-400 focus:outline-none focus:ring-2 focus:ring-orange-500"
|
className="w-full px-3 py-2 border border-gray-200 dark:border-gray-600 rounded-lg text-sm bg-white dark:bg-gray-700 text-gray-900 dark:text-white placeholder-gray-400 focus:outline-none focus:ring-2 focus:ring-orange-500"
|
||||||
/>
|
/>
|
||||||
<p className="text-xs text-gray-400 mt-1">
|
<p className="text-xs text-gray-400 mt-1">
|
||||||
智谱: glm-4-flash(免费), glm-4-plus, glm-4.5, glm-4.6
|
智谱: glm-4-flash-250414(免费), glm-4-plus, glm-4.7, glm-z1-flash(免费推理)
|
||||||
</p>
|
</p>
|
||||||
</div>
|
</div>
|
||||||
|
|
||||||
|
|||||||
@@ -51,11 +51,11 @@ export const PROVIDER_DEFAULTS = {
|
|||||||
},
|
},
|
||||||
zhipu: {
|
zhipu: {
|
||||||
baseUrl: 'https://open.bigmodel.cn/api/paas/v4',
|
baseUrl: 'https://open.bigmodel.cn/api/paas/v4',
|
||||||
defaultModel: 'glm-4-flash',
|
defaultModel: 'glm-4-flash-250414',
|
||||||
},
|
},
|
||||||
zhipu_coding: {
|
zhipu_coding: {
|
||||||
baseUrl: 'https://open.bigmodel.cn/api/coding/paas/v4',
|
baseUrl: 'https://open.bigmodel.cn/api/coding/paas/v4',
|
||||||
defaultModel: 'glm-4-flash',
|
defaultModel: 'glm-4-flash-250414',
|
||||||
},
|
},
|
||||||
kimi: {
|
kimi: {
|
||||||
baseUrl: 'https://api.moonshot.cn/v1',
|
baseUrl: 'https://api.moonshot.cn/v1',
|
||||||
|
|||||||
@@ -192,7 +192,7 @@ export const useConversationStore = create<ConversationState>()(
|
|||||||
agents: [DEFAULT_AGENT],
|
agents: [DEFAULT_AGENT],
|
||||||
currentAgent: DEFAULT_AGENT,
|
currentAgent: DEFAULT_AGENT,
|
||||||
sessionKey: null,
|
sessionKey: null,
|
||||||
currentModel: 'glm-4-flash',
|
currentModel: 'glm-4-flash-250414',
|
||||||
|
|
||||||
newConversation: (currentMessages: ChatMessage[]) => {
|
newConversation: (currentMessages: ChatMessage[]) => {
|
||||||
const state = get();
|
const state = get();
|
||||||
|
|||||||
Reference in New Issue
Block a user