diff --git a/changelog/v1.json b/changelog/v1.json index e5afd629264a..12d78e12d100 100644 --- a/changelog/v1.json +++ b/changelog/v1.json @@ -1,12 +1,10 @@ [ { "children": { - "improvements": ["Refactor services code style."], - "features": ["Add changelog modal."], - "fixes": ["Fix o1Models list."] + "improvements": ["Refactor tokens to contextWindowTokens."] }, "date": "2024-12-26", - "version": "1.82.0" + "version": "1.40.2" }, { "children": { @@ -23,13 +21,6 @@ "date": "2024-12-26", "version": "1.40.0" }, - { - "children": { - "improvements": ["Improve loading brand."] - }, - "date": "2024-12-25", - "version": "1.81.2" - }, { "children": { "improvements": ["Improve loading brand."] @@ -44,13 +35,6 @@ "date": "2024-12-25", "version": "1.39.2" }, - { - "children": { - "fixes": ["Fix image input on pglite."] - }, - "date": "2024-12-24", - "version": "1.81.1" - }, { "children": { "fixes": ["Fix image input on pglite."] @@ -58,13 +42,6 @@ "date": "2024-12-24", "version": "1.39.1" }, - { - "children": { - "features": ["Upgrade to next15 and react19."] - }, - "date": "2024-12-23", - "version": "1.81.0" - }, { "children": { "features": ["Upgrade to next15 and react19."] @@ -77,8925 +54,3134 @@ "features": ["Support thread in client pglite."] }, "date": "2024-12-23", - "version": "1.80.0" + "version": "1.38.0" }, { "children": { - "improvements": ["Move pglite to client service, refactor the client service to deprecated."], - "features": ["Support to use pglite as client db."] + "improvements": ["Move pglite to client service."] }, "date": "2024-12-22", - "version": "1.79.0" + "version": "1.37.2" }, { "children": { - "improvements": ["Add o1 model in GitHub models."] + "improvements": ["Refactor the client service to deprecated."] }, - "date": "2024-12-21", - "version": "1.78.18" + "date": "2024-12-22", + "version": "1.37.1" }, { "children": { - "improvements": ["Add Gemini flash thinking model."] + "features": ["Support to use pglite as client db."] }, - "date": "2024-12-21", - "version": "1.78.17" + "date": "2024-12-22", + "version": "1.37.0" }, { "children": { - "improvements": ["Upgrade react scan."], - "fixes": ["Fix HUGGINGFACE endpoint url."] + "improvements": ["Refactor client mode upload to match server mode."] }, "date": "2024-12-21", - "version": "1.78.16" - }, - { - "children": { - "improvements": [ - "Refactor layout props, refactor to use async headers(), seperate user keyVaults encrpyto from user model." - ] - }, - "date": "2024-12-20", - "version": "1.78.15" - }, - { - "children": {}, - "date": "2024-12-19", - "version": "1.78.14" - }, - { - "children": { - "improvements": ["Improve home page loading for better UX."], - "fixes": ["Fix GitHub model fetch, fix pdf preview with capital ext."] - }, - "date": "2024-12-19", - "version": "1.78.13" - }, - { - "children": { - "improvements": ["Improve page loading state, update models of Gitee AI provider."], - "fixes": [ - "Add unique keys to children, fix discover locale with different default lang." - ] - }, - "date": "2024-12-17", - "version": "1.78.12" - }, - { - "children": { - "improvements": ["Add new grok models."] - }, - "date": "2024-12-15", - "version": "1.78.11" - }, - { - "children": { - "fixes": ["One of Gemini functionCall error."], - "improvements": ["Support csv chunking, update locale."] - }, - "date": "2024-12-14", - "version": "1.78.10" - }, - { - "children": { - "improvements": ["Enable googleSearch Tool for gemini-2.0-flash-exp."], - "fixes": ["Fix claude first message can not be assistant."] - }, - "date": "2024-12-12", - "version": "1.78.9" - }, - { - "children": { - "improvements": ["Refactor database file model to remove server env."] - }, - "date": "2024-12-12", - "version": "1.78.8" - }, - { - "children": { - "improvements": ["Add Gemini 2.0 Flash Exp model."] - }, - "date": "2024-12-12", - "version": "1.78.7" + "version": "1.36.46" }, { "children": { - "improvements": ["Refactor data importer to repos."] + "improvements": ["Add o1 model in GitHub models."] }, - "date": "2024-12-11", - "version": "1.78.6" + "date": "2024-12-21", + "version": "1.36.45" }, { "children": { - "improvements": ["Add GLM-4V-Flash from Zhipu."], - "fixes": ["Fix pricing with 0 digit, nullptr errors in NextAuth adapter."] + "improvements": ["Add Gemini flash thinking model."] }, - "date": "2024-12-10", - "version": "1.78.5" + "date": "2024-12-21", + "version": "1.36.44" }, { - "children": { - "improvements": ["Update groq, add llama3.3, Upgrade lobe-ui."] - }, - "date": "2024-12-10", - "version": "1.78.4" + "children": {}, + "date": "2024-12-21", + "version": "1.36.43" }, { "children": { - "fixes": ["Fix wrong email linking in next-auth db adapter."] + "fixes": ["Fix HUGGINGFACE endpoint url."] }, - "date": "2024-12-09", - "version": "1.78.3" + "date": "2024-12-21", + "version": "1.36.42" }, { "children": { - "fixes": ["Support request headers for chat."] + "improvements": ["Upgrade react scan."] }, - "date": "2024-12-08", - "version": "1.78.2" + "date": "2024-12-21", + "version": "1.36.41" }, { "children": { - "improvements": ["Add gemini-exp-1206 model."] + "improvements": ["Seperate user keyVaults encrpyto from user model."] }, - "date": "2024-12-07", - "version": "1.78.1" + "date": "2024-12-20", + "version": "1.36.40" }, { "children": { - "improvements": ["Refactor page params to adapt next15 breaking change."], - "features": ["Add Higress ai model provider."] + "improvements": ["Refactor to use async headers()."] }, - "date": "2024-12-06", - "version": "1.78.0" - }, - { - "children": {}, - "date": "2024-12-06", - "version": "1.77.9" + "date": "2024-12-20", + "version": "1.36.39" }, { "children": { - "fixes": ["Fix typo of prompts."] + "improvements": ["Refactor layout props."] }, - "date": "2024-12-05", - "version": "1.77.8" + "date": "2024-12-20", + "version": "1.36.38" }, { "children": {}, - "date": "2024-12-04", - "version": "1.77.7" - }, - { - "children": { - "improvements": ["Refactor the server db model implement."] - }, - "date": "2024-12-03", - "version": "1.77.6" + "date": "2024-12-19", + "version": "1.36.37" }, { "children": {}, - "date": "2024-12-03", - "version": "1.77.5" - }, - { - "children": { - "improvements": ["Move schema and migration folder."] - }, - "date": "2024-12-03", - "version": "1.77.4" - }, - { - "children": { - "improvements": ["Add QwQ 32B Preview model."] - }, - "date": "2024-12-03", - "version": "1.77.3" - }, - { - "children": { - "improvements": ["Deprecated the current client mode code."] - }, - "date": "2024-12-02", - "version": "1.77.2" - }, - { - "children": { - "improvements": ["Add gpt-4o-2024-11-20 model, improve i18n."] - }, - "date": "2024-12-01", - "version": "1.77.1" - }, - { - "children": { - "features": ["Support ollama tools use."], - "improvements": [ - "Add QWEN_PROXY_URL support for Qwen, update model list, add qwq-32b-preview, update ollama models." - ] - }, - "date": "2024-12-01", - "version": "1.77.0" - }, - { - "children": { - "improvements": ["Add Google LearnLM model."] - }, - "date": "2024-11-28", - "version": "1.76.3" + "date": "2024-12-19", + "version": "1.36.36" }, { "children": { - "fixes": ["Fix fallback behavior of default mode in AgentRuntime."], - "improvements": ["Add switch portal thread."] + "improvements": ["Improve home page loading for better UX."] }, - "date": "2024-11-28", - "version": "1.76.2" + "date": "2024-12-18", + "version": "1.36.35" }, { "children": { - "improvements": ["Improve thread i18n locale."] + "fixes": ["Fix pdf preview with capital ext."] }, - "date": "2024-11-27", - "version": "1.76.1" + "date": "2024-12-18", + "version": "1.36.34" }, { "children": { - "features": ["Forkable Chat Mode."], - "fixes": ["Fix Qwen baseUrl calling."] + "fixes": ["Fix GitHub model fetch."] }, - "date": "2024-11-27", - "version": "1.76.0" + "date": "2024-12-18", + "version": "1.36.33" }, { "children": { - "improvements": ["Update the description translation of Gitee AI."] + "improvements": ["Refactor the drizzle code style."] }, - "date": "2024-11-26", - "version": "1.75.4" + "date": "2024-12-17", + "version": "1.36.32" }, { "children": { - "improvements": ["Refactor getLlmOptionsFromPayload from AgentRuntime."] + "improvements": ["Refactor the data fetch with clientDB init check."] }, - "date": "2024-11-26", - "version": "1.75.3" + "date": "2024-12-17", + "version": "1.36.31" }, { "children": { - "fixes": ["Fix fetchOnClient functional for Moonshot, fix multi-turns tools calling."] + "improvements": ["Improve page loading state."] }, - "date": "2024-11-25", - "version": "1.75.2" + "date": "2024-12-16", + "version": "1.36.30" }, { "children": { - "improvements": ["Add gemini-exp-1121 model."] + "fixes": ["Fix discover locale with different default lang."] }, - "date": "2024-11-25", - "version": "1.75.1" + "date": "2024-12-16", + "version": "1.36.29" }, { - "children": { - "features": ["Add Gitee AI model provider."] - }, - "date": "2024-11-25", - "version": "1.75.0" + "children": {}, + "date": "2024-12-16", + "version": "1.36.28" }, { "children": { - "fixes": ["Fix XAI_PROXY_URL env missing."], - "improvements": ["Support to reset fetched models."] + "fixes": ["Add unique keys to children."] }, - "date": "2024-11-24", - "version": "1.74.5" + "date": "2024-12-16", + "version": "1.36.27" }, { "children": { - "fixes": ["Fix tool message display."] + "improvements": ["Update models of Gitee AI provider."] }, - "date": "2024-11-24", - "version": "1.74.4" + "date": "2024-12-16", + "version": "1.36.26" }, { "children": { - "improvements": ["Refactor the main chat."] + "improvements": ["Add new grok models."] }, - "date": "2024-11-24", - "version": "1.74.3" + "date": "2024-12-14", + "version": "1.36.25" }, { "children": { - "improvements": [ - "Add grok-vision-beta model, update Mistral model list, add pixtral-large-latest." - ] + "improvements": ["Refactor file Url query in message model."] }, - "date": "2024-11-20", - "version": "1.74.2" - }, - { - "children": {}, - "date": "2024-11-19", - "version": "1.74.1" + "date": "2024-12-14", + "version": "1.36.24" }, { "children": { - "features": ["Add support InternLM (书生浦语) provider."], - "fixes": ["Connection check logic, keyword search for chat history & sessions."], - "improvements": ["Support o1 models using streaming."] + "improvements": ["Support csv chunking."] }, - "date": "2024-11-19", - "version": "1.74.0" - }, - { - "children": {}, - "date": "2024-11-16", - "version": "1.73.8" - }, - { - "children": { - "improvements": ["Add gemini-exp-1114 model."] - }, - "date": "2024-11-16", - "version": "1.73.7" - }, - { - "children": { - "improvements": [ - "genServerLLMConfig function, get *_MODEL_LIST from env, move ChatInput to features/ChatInput." - ] - }, - "date": "2024-11-15", - "version": "1.73.6" - }, - { - "children": { - "improvements": ["Refactor the chat conversation implement."] - }, - "date": "2024-11-13", - "version": "1.73.5" - }, - { - "children": { - "improvements": ["Update some provider modellist & fix ai360 baseurl."] - }, - "date": "2024-11-12", - "version": "1.73.4" - }, - { - "children": { - "improvements": ["Fix Cloudflare Workers AI Sort."] - }, - "date": "2024-11-12", - "version": "1.73.3" - }, - { - "children": { - "improvements": ["Refactor languageModel & DEFAULT_LLM_CONFIG generate."] - }, - "date": "2024-11-12", - "version": "1.73.2" - }, - { - "children": { - "improvements": ["Update deepseek model."] - }, - "date": "2024-11-12", - "version": "1.73.1" - }, - { - "children": { - "features": ["Add support xAI provider, Support Cloudflare Workers AI."], - "improvements": ["Fix Windows always showing scrollbar."] - }, - "date": "2024-11-12", - "version": "1.73.0" - }, - { - "children": { - "fixes": ["Fix summary range."], - "improvements": ["Add Sonnet 3.5 v2 inference model to BedRock."] - }, - "date": "2024-11-10", - "version": "1.72.4" - }, - { - "children": { - "fixes": ["Disregard remoteModelCards when showModelFetcher is disabled."] - }, - "date": "2024-11-09", - "version": "1.72.3" - }, - { - "children": { - "fixes": ["Fix the display model of history summary."] - }, - "date": "2024-11-09", - "version": "1.72.2" - }, - { - "children": { - "fixes": ["Fix topic summary field on server db."], - "improvements": ["Allow users to disable SSRF or set a whitelist."] - }, - "date": "2024-11-09", - "version": "1.72.1" - }, - { - "children": { - "features": ["Support compress history messages."], - "fixes": ["Fix env typo of MS Entra ID."] - }, - "date": "2024-11-09", - "version": "1.72.0" - }, - { - "children": {}, - "date": "2024-11-08", - "version": "1.71.3" - }, - { - "children": { - "fixes": ["Disable model fetch for GitHub."] - }, - "date": "2024-11-07", - "version": "1.71.2" - }, - { - "children": { - "improvements": [ - "Move portal code to features folder, refactor and clean some code, update database fields." - ] - }, - "date": "2024-11-06", - "version": "1.71.1" - }, - { - "children": { - "features": ["Support export as markdown and JSON."] - }, - "date": "2024-11-06", - "version": "1.71.0" - }, - { - "children": { - "fixes": ["Remove the 'resetConversation' hot key tip."], - "improvements": ["Add claude 3.5 haiku model, improve group topic styles."] - }, - "date": "2024-11-05", - "version": "1.70.1" - }, - { - "children": { - "features": ["Support group topics by time."], - "fixes": [ - "Fix /webapi/plugin/store server error, if enable login and not signed in, return unauthorized error." - ] - }, - "date": "2024-11-04", - "version": "1.70.0" - }, - { - "children": { - "improvements": ["Disable chunking button for unsupported files."] - }, - "date": "2024-11-04", - "version": "1.69.11" - }, - { - "children": { - "fixes": [ - "Fix duplicate key value violates unique constraint \"slug_user_id_unique\" when create inbox session, Fix MS Entra ID and Azure AD authorization." - ] - }, - "date": "2024-11-03", - "version": "1.69.10" - }, - { - "children": { - "improvements": ["Improve server log on chat api."] - }, - "date": "2024-10-31", - "version": "1.69.9" - }, - { - "children": { - "improvements": ["Refactor embedding as plain vector array."], - "fixes": ["Fix server Network connection lost error."] - }, - "date": "2024-10-31", - "version": "1.69.8" - }, - { - "children": { - "improvements": ["Add fa-ir locale."] - }, - "date": "2024-10-30", - "version": "1.69.7" - }, - { - "children": { - "fixes": ["Fix file image prompts in client mode, fix the artifacts interface not scrolling."] - }, - "date": "2024-10-30", - "version": "1.69.6" - }, - { - "children": { - "improvements": ["Refactor the aiChat slice actions."] - }, - "date": "2024-10-29", - "version": "1.69.5" - }, - { - "children": { - "fixes": ["Update zhipu param process."] - }, - "date": "2024-10-29", - "version": "1.69.4" - }, - { - "children": { - "fixes": ["Remove PWA Install in Firefox and Arc."], - "improvements": [ - "Add Qwen2.5-72B-Instruct model on HF provider, improve user guide when user not login." - ] - }, - "date": "2024-10-29", - "version": "1.69.3" - }, - { - "children": { - "fixes": ["Remove the 'resetConversation' hot key."] - }, - "date": "2024-10-29", - "version": "1.69.2" - }, - { - "children": { - "improvements": ["Remove SenseChat-Vision model, due to model limitation."], - "fixes": [ - "Fix Huggingface API interrupting when the output exceeds 140 tokens, fix page not switch when clicking on the pin assistant." - ] - }, - "date": "2024-10-28", - "version": "1.69.1" - }, - { - "children": { - "features": ["experimentally support to pin assistant to sidebar."], - "fixes": ["Fix the issue of the switch assistant portal not closing."], - "improvements": ["Update stepfun models."] - }, - "date": "2024-10-27", - "version": "1.69.0" - }, - { - "children": { - "fixes": ["Fix modelList merge."] - }, - "date": "2024-10-26", - "version": "1.68.1" - }, - { - "children": { - "features": ["Support ZEN mode."], - "improvements": ["Update Google Model list, add gemini-1.5-flash-8b."] - }, - "date": "2024-10-26", - "version": "1.68.0" - }, - { - "children": { - "improvements": ["Refactor the Google Gen AI."], - "features": ["Add SenseNova (商汤) model provider."] - }, - "date": "2024-10-25", - "version": "1.67.0" - }, - { - "children": { - "improvements": ["Add *_MODEL_LIST env to all models, update Spark model id & display name."] - }, - "date": "2024-10-25", - "version": "1.66.1" - }, - { - "children": { - "features": ["Support system agent config."], - "improvements": ["Add bedrock claude-3.5-sonnect-v2."] - }, - "date": "2024-10-25", - "version": "1.66.0" - }, - { - "children": { - "improvements": ["Fix some custom branding detail, improve error i18n, improve i18n."] - }, - "date": "2024-10-24", - "version": "1.65.12" - }, - { - "children": { - "improvements": ["Refactor cookie/headers to async mode."] - }, - "date": "2024-10-23", - "version": "1.65.11" - }, - { - "children": { - "improvements": ["Add new claude-3.5-sonnet model."] - }, - "date": "2024-10-23", - "version": "1.65.10" - }, - { - "children": { - "improvements": ["Move responsive to server utils folder."] - }, - "date": "2024-10-23", - "version": "1.65.9" - }, - { - "children": { - "improvements": ["Fix dynamic import in rsc layout."] - }, - "date": "2024-10-22", - "version": "1.65.8" - }, - { - "children": { - "improvements": ["Update format utils and shared layout."], - "fixes": ["Fix azure-ad."] - }, - "date": "2024-10-21", - "version": "1.65.7" - }, - { - "children": { - "improvements": [ - "Add Llama 3.1 Nemotron 70B model & reorder some provider model list, add Ministral model, update Together AI model list, add function call & vision, update wenxin 4.0 turbo model to latest." - ] - }, - "date": "2024-10-20", - "version": "1.65.6" - }, - { - "children": { - "improvements": ["Update Fireworks AI model list."] - }, - "date": "2024-10-20", - "version": "1.65.5" - }, - { - "children": { - "improvements": ["Add Yi-Lightning model."] - }, - "date": "2024-10-18", - "version": "1.65.4" - }, - { - "children": { - "improvements": ["Add qwen vision model & update qwen2.5 72b to 128k for siliconcloud."] - }, - "date": "2024-10-17", - "version": "1.65.3" - }, - { - "children": { - "improvements": [ - "Reorder github model list & updata info & add new model, support multi-windows for PWA." - ], - "fixes": ["Fix images not go in to chat context."] - }, - "date": "2024-10-14", - "version": "1.65.2" - }, - { - "children": { - "fixes": ["Allow use email as name in logto."] - }, - "date": "2024-10-13", - "version": "1.65.1" - }, - { - "children": { - "improvements": ["Refactor the chat webapi."], - "features": ["Add HuggingFace Model Provider."], - "fixes": ["Fix function calling issue, disable stream when using tools."] - }, - "date": "2024-10-12", - "version": "1.65.0" - }, - { - "children": { - "improvements": ["Fix artifacts render markdown."] - }, - "date": "2024-10-12", - "version": "1.64.12" - }, - { - "children": { - "improvements": [ - "Refactor agent runtime implement of stream and ZHIPU provider, refactor the jwt code." - ] - }, - "date": "2024-10-11", - "version": "1.64.11" - }, - { - "children": { - "improvements": ["Updata gpt-4o model info."] - }, - "date": "2024-10-11", - "version": "1.64.10" - }, - { - "children": { - "improvements": ["Update qwen vl model to latest."] - }, - "date": "2024-10-10", - "version": "1.64.9" - }, - { - "children": { - "fixes": ["Fix auto rewrite query when user message is too long."], - "improvements": ["Support yml in file chunk."] - }, - "date": "2024-10-08", - "version": "1.64.8" - }, - { - "children": { - "improvements": ["Refactor text-to-image endpoint."] - }, - "date": "2024-10-08", - "version": "1.64.7" - }, - { - "children": { - "improvements": ["Move backend api to (backend) folder group."], - "fixes": ["Fix txt-to-image api."] - }, - "date": "2024-10-05", - "version": "1.64.6" - }, - { - "children": { - "improvements": ["Support shadcn in Artifacts."] - }, - "date": "2024-10-05", - "version": "1.64.5" - }, - { - "children": { - "fixes": ["Fix recharts deps in the Artifacts React Renderer."] - }, - "date": "2024-10-02", - "version": "1.64.4" - }, - { - "children": { - "improvements": ["Move most /api to /webapi."] - }, - "date": "2024-10-01", - "version": "1.64.3" - }, - { - "children": { - "improvements": ["Adjust Wenxin icon size."] - }, - "date": "2024-10-01", - "version": "1.64.2" - }, - { - "children": {}, - "date": "2024-09-30", - "version": "1.64.1" - }, - { - "children": { - "features": ["Add wenxin model provider."] - }, - "date": "2024-09-30", - "version": "1.64.0" - }, - { - "children": {}, - "date": "2024-09-30", - "version": "1.63.5" - }, - { - "children": { - "improvements": ["Update groq model list."] - }, - "date": "2024-09-29", - "version": "1.63.4" - }, - { - "children": {}, - "date": "2024-09-29", - "version": "1.63.3" - }, - { - "children": { - "fixes": ["Improve delete orphan chunks when delete files."] - }, - "date": "2024-09-28", - "version": "1.63.2" - }, - { - "children": { - "improvements": ["Add zhipu glm-4-flashx model."] - }, - "date": "2024-09-27", - "version": "1.63.1" - }, - { - "children": { - "features": ["Add Hunyuan(Tencent) model provider."] - }, - "date": "2024-09-27", - "version": "1.63.0" - }, - { - "children": { - "improvements": [ - "Add llama3.2 model for openrouter provider, add o1-preview and o1-mini model to github model provider." - ] - }, - "date": "2024-09-27", - "version": "1.62.16" - }, - { - "children": {}, - "date": "2024-09-26", - "version": "1.62.15" - }, - { - "children": { - "fixes": ["MiniMax output long content interrupted by non-existent error."], - "improvements": ["Add function call for taichu_llm, update google provider model info."] - }, - "date": "2024-09-25", - "version": "1.62.14" - }, - { - "children": { - "improvements": [ - "Add function call support for Stepfun, add google gemini 1.5 002 series, disable taichu2.0 functioncall & default disable taichu2.0v model, update taichu provider info & add taichu vision model." - ] - }, - "date": "2024-09-24", - "version": "1.62.13" - }, - { - "children": { - "fixes": ["Fix url config import after user state init."], - "improvements": [ - "Add support function call for 360AI, improve images display in chat messages, left sidebar has only assistants." - ] - }, - "date": "2024-09-24", - "version": "1.62.12" - }, - { - "children": { - "fixes": ["Add missing translations."] - }, - "date": "2024-09-24", - "version": "1.62.11" - }, - { - "children": { - "fixes": ["Fix artifacts code language highlight."] - }, - "date": "2024-09-23", - "version": "1.62.10" - }, - { - "children": { - "improvements": ["Add spark max-32k model."] - }, - "date": "2024-09-23", - "version": "1.62.9" - }, - { - "children": { - "fixes": ["Fix ollama model download panel."] - }, - "date": "2024-09-22", - "version": "1.62.8" - }, - { - "children": { - "improvements": ["Refactor to improve branding customization."] - }, - "date": "2024-09-22", - "version": "1.62.7" - }, - { - "children": { - "fixes": ["Casdoor webhooks providerAccountId not found, fix Content-Security-Policy."] - }, - "date": "2024-09-21", - "version": "1.62.6" - }, - { - "children": { - "fixes": ["providerAccountId not exist in provider."], - "improvements": ["Improve i18n for discover and improve version check."] - }, - "date": "2024-09-21", - "version": "1.62.5" - }, - { - "children": { - "fixes": ["Pin shiki@1.17.7 to fix code highlight, try to implement better ssrf-protect."], - "improvements": ["Improve i18n in discover."] - }, - "date": "2024-09-20", - "version": "1.62.4" - }, - { - "children": { - "fixes": ["Custom model initialization not taking effect error."], - "improvements": [ - "Add qwen2.5 math and coder model for siliconcloud provider, support webhooks for casdoor." - ] - }, - "date": "2024-09-20", - "version": "1.62.3" - }, - { - "children": { - "improvements": [ - "Add siliconflow qwen2.5 model, delete siliconflow outdated model & disable stepfun functioncall, enable functioncall for stepfun models, Update qwen models." - ], - "fixes": ["Fix a bug with server agent config when user not exist."] - }, - "date": "2024-09-20", - "version": "1.62.2" - }, - { - "children": { - "improvements": ["Updata qwen model info & add qwen2.5 & reorder provider list."] - }, - "date": "2024-09-19", - "version": "1.62.1" - }, - { - "children": { - "features": [ - "Add Ai21Labs model provider, add Github Models provider, support native Artifacts just like Claude." - ], - "fixes": ["Fix InterceptingRoutes in discover."], - "improvements": ["Add mistral provider new models."] - }, - "date": "2024-09-19", - "version": "1.62.0" - }, - { - "children": { - "features": ["Add Discover Page."] - }, - "date": "2024-09-18", - "version": "1.61.0" - }, - { - "children": { - "fixes": ["Fix a corner case of tools_call with empty object."], - "improvements": ["Delete duplicate models in ollama."] - }, - "date": "2024-09-17", - "version": "1.60.6" - }, - { - "children": { - "improvements": ["Add MiniCPM-V 8B model entries to Ollama model providers."] - }, - "date": "2024-09-16", - "version": "1.60.5" - }, - { - "children": { - "improvements": ["Update fullscreen loading style."] - }, - "date": "2024-09-15", - "version": "1.60.4" - }, - { - "children": { - "improvements": ["Delete \"-\" in deepseek displayname."] - }, - "date": "2024-09-14", - "version": "1.60.3" - }, - { - "children": { - "fixes": ["Fix o1 model list."], - "improvements": ["Update openrouter model list."] - }, - "date": "2024-09-13", - "version": "1.60.2" - }, - { - "children": { - "improvements": ["Update zhipu model info."] - }, - "date": "2024-09-13", - "version": "1.60.1" - }, - { - "children": { - "improvements": [ - "Default disable mistral provider useless models, Remove brackets from model names with dates in OpenAI, Support Google Model List, Update siliconcloud model." - ], - "features": ["Support openai new OpenAI o1-preview/o1-mini models."], - "fixes": ["Qwen model param error, support webhooks for logto."] - }, - "date": "2024-09-13", - "version": "1.60.0" - }, - { - "children": { - "improvements": ["Add model and provider desc and url."] - }, - "date": "2024-09-12", - "version": "1.59.4" - }, - { - "children": { - "improvements": ["Improve models and add more info for providers and models."] - }, - "date": "2024-09-12", - "version": "1.59.3" - }, - { - "children": { - "improvements": ["Optimize model token display method."] - }, - "date": "2024-09-11", - "version": "1.59.2" - }, - { - "children": { - "fixes": ["Pin next@14.2.8 to fix Internal error."] - }, - "date": "2024-09-11", - "version": "1.59.1" - }, - { - "children": { - "features": ["Add Fireworks AI Model Provider, Add Spark model provider."], - "fixes": ["Add LLM_VISION_IMAGE_USE_BASE64 to support local s3 in vision model."], - "improvements": [ - "Reorder the provider list, Update CustomLogo, update spark check model to spark-lite & default disable useless model, update Upstage model list." - ] - }, - "date": "2024-09-11", - "version": "1.59.0" - }, - { - "children": { - "improvements": ["Change empty content stream behavior."] - }, - "date": "2024-09-10", - "version": "1.58.16" - }, - { - "children": { - "improvements": ["Add siliconcloud new model."], - "fixes": [ - "Baichuan should not introduce freequency_penality parameters, fix /etc/resolv.confedit permission in docker image, fix claude 3.5 image with s3 url, fix tools calling in some edge cases, Gemini cannot input images when server database is enabled, Update baichuan param." - ] - }, - "date": "2024-09-10", - "version": "1.58.15" - }, - { - "children": { - "improvements": ["Update perplexity model list."] - }, - "date": "2024-09-09", - "version": "1.58.14" - }, - { - "children": { - "improvements": [ - "Fix title in about settings, Update model display name & Remove Qwen preview model." - ] - }, - "date": "2024-09-09", - "version": "1.58.13" - }, - { - "children": {}, - "date": "2024-09-08", - "version": "1.58.12" - }, - { - "children": {}, - "date": "2024-09-08", - "version": "1.58.11" - }, - { - "children": { - "improvements": ["Support anthropic browser request."] - }, - "date": "2024-09-07", - "version": "1.58.10" - }, - { - "children": { - "fixes": ["Fix auth log."] - }, - "date": "2024-09-06", - "version": "1.58.9" - }, - { - "children": { - "improvements": [ - "Add LLaVA 1.5 7B model in Groq, Update Bedrock model list & add AWS_BEDROCK_MODEL_LIST support." - ] - }, - "date": "2024-09-06", - "version": "1.58.8" - }, - { - "children": {}, - "date": "2024-09-06", - "version": "1.58.7" - }, - { - "children": { - "fixes": ["Fix typo in RAG prompt."] - }, - "date": "2024-09-05", - "version": "1.58.6" - }, - { - "children": { - "fixes": ["Fix .PDF can not be chunked, fix speed and rag prompt."] - }, - "date": "2024-09-04", - "version": "1.58.5" - }, - { - "children": { - "improvements": ["Fix provider disabled title style."] - }, - "date": "2024-09-03", - "version": "1.58.4" - }, - { - "children": { - "improvements": [ - "Stepfun default enabled model, update Groq model list & add GROQ_MODEL_LIST support." - ] - }, - "date": "2024-09-02", - "version": "1.58.3" - }, - { - "children": { - "improvements": [ - "Update Novita AI model info & add NOVITA_MODEL_LIST support, update Together AI model list." - ] - }, - "date": "2024-09-01", - "version": "1.58.2" - }, - { - "children": { - "improvements": [ - "Add *_MODEL_LIST for Qwen and ZeroOne, fix model info, update Claude 3.5 Sonnet maxOutput vaule." - ] - }, - "date": "2024-09-01", - "version": "1.58.1" - }, - { - "children": { - "features": ["Add Upstage model provider support."], - "fixes": ["Fix file relative chunks."], - "improvements": [ - "Fix ms doc file preview, Update Qwen and Gemini models info, Update Stepfun models info, Update the sorting of each provider model, Update the sorting of each provider model." - ] - }, - "date": "2024-08-30", - "version": "1.58.0" - }, - { - "children": {}, - "date": "2024-08-30", - "version": "1.57.4" - }, - { - "children": { - "fixes": ["Fix whisper-1 typo."] - }, - "date": "2024-08-29", - "version": "1.57.3" - }, - { - "children": { - "fixes": ["Disable ChatGPT-4o Tools Calling."], - "improvements": ["Improve chunk and file preview, Update Gemini models."] - }, - "date": "2024-08-29", - "version": "1.57.2" - }, - { - "children": { - "fixes": ["No user name if Cloudflare Zero Trust with onetimepin."], - "improvements": ["Move model and provider icon components to @lobehub/icons."] - }, - "date": "2024-08-28", - "version": "1.57.1" - }, - { - "children": { - "features": ["Supports Cloudflare Zero Trust login."], - "fixes": ["Fix agent setting, improve aysnc error type."], - "improvements": ["Improve zhipu model config."] - }, - "date": "2024-08-27", - "version": "1.57.0" - }, - { - "children": { - "fixes": ["Bypass vercel deployment protection, fix can send message on uploading files."], - "improvements": ["Update Qwen models."] - }, - "date": "2024-08-27", - "version": "1.56.1" - }, - { - "children": { - "features": ["Supports Authelia login."] - }, - "date": "2024-08-27", - "version": "1.56.0" - }, - { - "children": { - "fixes": ["Feature flag knowledge_base doesn't affect ActionBar."] - }, - "date": "2024-08-27", - "version": "1.55.15" - }, - { - "children": { - "fixes": [ - "Fix cannot clone agent when imported from client, fix dayjs error in en-US language, fix multi file upload dupicate." - ] - }, - "date": "2024-08-26", - "version": "1.55.14" - }, - { - "children": { - "fixes": ["Session not found error on mobile."] - }, - "date": "2024-08-24", - "version": "1.55.13" - }, - { - "children": { - "fixes": ["Fix tts file saving in server mode."], - "improvements": ["Update 01.AI models."] - }, - "date": "2024-08-24", - "version": "1.55.12" - }, - { - "children": {}, - "date": "2024-08-24", - "version": "1.55.11" - }, - { - "children": { - "fixes": ["Remove orphan chunks if there is no related file."] - }, - "date": "2024-08-23", - "version": "1.55.10" - }, - { - "children": { - "fixes": ["Refactor and fix dalle."] - }, - "date": "2024-08-23", - "version": "1.55.9" - }, - { - "children": { - "fixes": ["Improve s3 path-style url."] - }, - "date": "2024-08-23", - "version": "1.55.8" - }, - { - "children": { - "fixes": [ - "Fix NEXT_PUBLIC_S3_DOMAIN error on Docker, Logout button not shown on mobile view when using nextauth." - ] - }, - "date": "2024-08-22", - "version": "1.55.7" - }, - { - "children": { - "improvements": ["Refactor s3 env and support path-style for minio."] - }, - "date": "2024-08-22", - "version": "1.55.6" - }, - { - "children": { - "fixes": ["Fix clipboard copy issue and improve upload cors feedback."] - }, - "date": "2024-08-22", - "version": "1.55.5" - }, - { - "children": { - "improvements": ["Fix link style."] - }, - "date": "2024-08-22", - "version": "1.55.4" - }, - { - "children": { - "improvements": ["Hide settings in repo."] - }, - "date": "2024-08-22", - "version": "1.55.3" - }, - { - "children": {}, - "date": "2024-08-22", - "version": "1.55.2" - }, - { - "children": { - "fixes": ["Fix embeddings multi-insert when there is issues with async task."] - }, - "date": "2024-08-21", - "version": "1.55.1" - }, - { - "children": { - "features": ["Files and knowledge base."] - }, - "date": "2024-08-21", - "version": "1.55.0" - }, - { - "children": { - "fixes": ["Fixed bedrock llama model id."] - }, - "date": "2024-08-19", - "version": "1.54.7" - }, - { - "children": { - "improvements": ["Update zhipu models."] - }, - "date": "2024-08-19", - "version": "1.54.6" - }, - { - "children": { - "improvements": ["Refactor the SITE_URL to APP_URL."], - "fixes": ["Fix topic scroll issue."] - }, - "date": "2024-08-19", - "version": "1.54.5" - }, - { - "children": { - "improvements": ["Refactor the fetch method to fix response.undefined."] - }, - "date": "2024-08-18", - "version": "1.54.4" - }, - { - "children": { - "improvements": [ - "Add SILICONCLOUD_MODEL_LIST & SILICONCLOUD_PROXY_URL support for SiliconCloud." - ] - }, - "date": "2024-08-18", - "version": "1.54.3" - }, - { - "children": { - "improvements": ["Refactor PanelTitle and move commit from file uploading."] - }, - "date": "2024-08-17", - "version": "1.54.2" - }, - { - "children": { - "fixes": ["Make S3 upload ACL setting optional."] - }, - "date": "2024-08-15", - "version": "1.54.1" - }, - { - "children": { - "features": ["Add 2 new models to openai provider."] - }, - "date": "2024-08-15", - "version": "1.54.0" - }, - { - "children": { - "features": ["Add SiliconCloud model provider."] - }, - "date": "2024-08-14", - "version": "1.53.0" - }, - { - "children": { - "improvements": ["Resize the image size in chat message."] - }, - "date": "2024-08-13", - "version": "1.52.7" - }, - { - "children": {}, - "date": "2024-08-13", - "version": "1.52.6" - }, - { - "children": {}, - "date": "2024-08-09", - "version": "1.52.5" - }, - { - "children": { - "improvements": ["Updated AWS bedrock model list."] - }, - "date": "2024-08-08", - "version": "1.52.4" - }, - { - "children": { - "fixes": ["Fix import clerk AuthObject from public api."] - }, - "date": "2024-08-07", - "version": "1.52.3" - }, - { - "children": { - "improvements": ["Refactor server db schema for better code organize."] - }, - "date": "2024-08-06", - "version": "1.52.2" - }, - { - "children": { - "fixes": ["Azure modelTag icon display."] - }, - "date": "2024-08-05", - "version": "1.52.1" - }, - { - "children": { - "features": ["Skip login page if only one provider exists."] - }, - "date": "2024-08-05", - "version": "1.52.0" - }, - { - "children": { - "fixes": ["Add PROXY_URL in docker with proxychains-ng."], - "improvements": ["Fix aya, mathstral model tag icon & update ollama model info."] - }, - "date": "2024-08-03", - "version": "1.51.1" - }, - { - "children": { - "features": ["Add NextAuth as authentication service in server database."] - }, - "date": "2024-08-02", - "version": "1.51.0" - }, - { - "children": { - "improvements": ["Add Gemini 1.5 Pro Exp model."] - }, - "date": "2024-08-02", - "version": "1.50.9" - }, - { - "children": { - "fixes": ["Fix Mistral models calling & update model info."], - "improvements": ["Fix stepfun & baichuan model tag icon missing, update Perplexity models."] - }, - "date": "2024-08-02", - "version": "1.50.8" - }, - { - "children": {}, - "date": "2024-07-30", - "version": "1.50.7" - }, - { - "children": { - "improvements": ["Improve tools calling UI."] - }, - "date": "2024-07-30", - "version": "1.50.6" - }, - { - "children": { - "fixes": ["Disable anthropic browser request."] - }, - "date": "2024-07-29", - "version": "1.50.5" - }, - { - "children": { - "fixes": [ - "Fix create_session edit_agent feature flags and add more flags, improve remote model list fetching for Novita AI." - ], - "improvements": ["Update 360GPT model (360GPT2 Pro)."] - }, - "date": "2024-07-29", - "version": "1.50.4" - }, - { - "children": { - "fixes": ["Update minimax models."] - }, - "date": "2024-07-28", - "version": "1.50.3" - }, - { - "children": { - "fixes": ["Avoid baseURL being an empty string, resulting in incorrect client fetch."] - }, - "date": "2024-07-26", - "version": "1.50.2" - }, - { - "children": { - "fixes": ["Fix dalle tools calling prompts to avoid content risk."] - }, - "date": "2024-07-26", - "version": "1.50.1" - }, - { - "children": { - "features": ["Enabled function calling on Deepseek models."], - "improvements": ["Fix file upload height."] - }, - "date": "2024-07-26", - "version": "1.50.0" - }, - { - "children": { - "improvements": ["Improve input file upload."] - }, - "date": "2024-07-26", - "version": "1.49.10" - }, - { - "children": { - "improvements": [ - "Add new models to groq which are llama 3.1, Updated Groq model list to include llama-3.1 and llama3-Groq." - ] - }, - "date": "2024-07-25", - "version": "1.49.9" - }, - { - "children": { - "fixes": ["Fix UNAUTHORIZED issue with clerk auth provider."] - }, - "date": "2024-07-24", - "version": "1.49.8" - }, - { - "children": { - "improvements": ["Fix the scrolling of the return result area of function calling."] - }, - "date": "2024-07-24", - "version": "1.49.7" - }, - { - "children": {}, - "date": "2024-07-23", - "version": "1.49.6" - }, - { - "children": { - "improvements": ["Add new model provider Novita AI."] - }, - "date": "2024-07-23", - "version": "1.49.5" - }, - { - "children": { - "fixes": ["Content lost unexpectedly on Qwen provider when finish_reason is stop."] - }, - "date": "2024-07-22", - "version": "1.49.4" - }, - { - "children": { - "improvements": ["Update Zhipu models (GLM-4-AllTools & CodeGeeX-4)."] - }, - "date": "2024-07-21", - "version": "1.49.3" - }, - { - "children": { - "fixes": ["Fix dayjs render on server."] - }, - "date": "2024-07-21", - "version": "1.49.2" - }, - { - "children": { - "improvements": ["Refactor the DragUpload."] - }, - "date": "2024-07-20", - "version": "1.49.1" - }, - { - "children": { - "features": ["Add gpt-4o-mini in OpenAI Provider and set it as the default model."] - }, - "date": "2024-07-19", - "version": "1.49.0" - }, - { - "children": { - "improvements": ["Added Gemma2 instead of outdated Gemma."] - }, - "date": "2024-07-19", - "version": "1.48.4" - }, - { - "children": { - "fixes": ["Fix delete session group."] - }, - "date": "2024-07-17", - "version": "1.48.3" - }, - { - "children": { - "fixes": ["Fix OpenAI deployment restrictions, fix cant duplicate assistant."] - }, - "date": "2024-07-17", - "version": "1.48.2" - }, - { - "children": { - "fixes": ["Fix session not reorder after send message."] - }, - "date": "2024-07-17", - "version": "1.48.1" - }, - { - "children": { - "features": ["Spport qwen-vl and tool call for qwen."], - "improvements": ["Improve brand url."] - }, - "date": "2024-07-17", - "version": "1.48.0" - }, - { - "children": { - "fixes": ["Save assistant info on blur."] - }, - "date": "2024-07-15", - "version": "1.47.3" - }, - { - "children": { - "improvements": ["Remove code related to gemini-pro-vision."] - }, - "date": "2024-07-13", - "version": "1.47.2" - }, - { - "children": { - "improvements": ["Add cloud promotion banner."] - }, - "date": "2024-07-13", - "version": "1.47.1" - }, - { - "children": { - "features": ["Add 360AI model provider."] - }, - "date": "2024-07-12", - "version": "1.47.0" - }, - { - "children": { - "improvements": ["Improve agent runtime code."] - }, - "date": "2024-07-11", - "version": "1.46.5" - }, - { - "children": { - "fixes": ["Fix assistant meta change race issue."] - }, - "date": "2024-07-10", - "version": "1.46.4" - }, - { - "children": { - "improvements": ["Support disable clientFetch by default."] - }, - "date": "2024-07-09", - "version": "1.46.3" - }, - { - "children": { - "fixes": ["Allow user to use their own WebRTC signaling."] - }, - "date": "2024-07-09", - "version": "1.46.2" - }, - { - "children": { - "fixes": ["Automatic refresh when change language."] - }, - "date": "2024-07-09", - "version": "1.46.1" - }, - { - "children": { - "features": ["Add Taichu model provider."] - }, - "date": "2024-07-09", - "version": "1.46.0" - }, - { - "children": { - "improvements": ["Provider changes with model in model settings."] - }, - "date": "2024-07-08", - "version": "1.45.10" - }, - { - "children": { - "fixes": ["Fix tool message order."] - }, - "date": "2024-07-07", - "version": "1.45.9" - }, - { - "children": { - "fixes": [ - "Fixed mobile web page navigation issue with inbox assistant, support to disable clerk signup with feature flag." - ], - "improvements": ["Update deepseek max token."] - }, - "date": "2024-07-07", - "version": "1.45.8" - }, - { - "children": { - "fixes": ["Improve tools calling error feedback when arguments are not correctly."] - }, - "date": "2024-07-05", - "version": "1.45.7" - }, - { - "children": { - "improvements": ["Fix tool message suspense loading."] - }, - "date": "2024-07-04", - "version": "1.45.6" - }, - { - "children": { - "fixes": ["Allow builtin tools to trigger AI message."], - "improvements": ["Improve delete assistant message with tools."] - }, - "date": "2024-07-03", - "version": "1.45.5" - }, - { - "children": { - "fixes": ["Clerk provider refreshes continously."] - }, - "date": "2024-07-03", - "version": "1.45.4" - }, - { - "children": { - "fixes": ["Fix clerk appearance is not applied correctly."] - }, - "date": "2024-07-02", - "version": "1.45.3" - }, - { - "children": { - "improvements": ["Update ProviderAvatar for Baichuan & Stepfun."] - }, - "date": "2024-07-02", - "version": "1.45.2" - }, - { - "children": { - "improvements": ["Refactor the portal file, refactor the portal implement."], - "fixes": ["Display issue when select default model in System Agent."] - }, - "date": "2024-07-01", - "version": "1.45.1" - }, - { - "children": { - "features": ["Add Baichuan model provider."] - }, - "date": "2024-07-01", - "version": "1.45.0" - }, - { - "children": { - "improvements": ["Add Gemini 1.5 stable version model & 2M context window."] - }, - "date": "2024-07-01", - "version": "1.44.12" - }, - { - "children": { - "fixes": ["Fix clerk UNAUTHORIZED error after long-time hang-up."] - }, - "date": "2024-06-30", - "version": "1.44.11" - }, - { - "children": { - "fixes": ["Pin @azure/core-rest-pipeline@1.16.0 to fix azure openai issue."] - }, - "date": "2024-06-28", - "version": "1.44.10" - }, - { - "children": { - "improvements": ["Refactor plugin state."] - }, - "date": "2024-06-27", - "version": "1.44.9" - }, - { - "children": { - "fixes": ["-check_updates cannot be set by FEATURE_FLAGS."] - }, - "date": "2024-06-27", - "version": "1.44.8" - }, - { - "children": { - "fixes": ["Fix azure tools calling."] - }, - "date": "2024-06-26", - "version": "1.44.7" - }, - { - "children": { - "improvements": ["Refactor format utils."] - }, - "date": "2024-06-25", - "version": "1.44.6" - }, - { - "children": { - "fixes": ["Fix agent tags, fix anthropic parallel tools calling."], - "improvements": ["Always show action on mobile."] - }, - "date": "2024-06-24", - "version": "1.44.5" - }, - { - "children": { - "improvements": ["Add AES-GCM footer and tooltip."] - }, - "date": "2024-06-24", - "version": "1.44.4" - }, - { - "children": { - "improvements": ["Refactor the server db implement, remove deprecated env."], - "fixes": ["Fix incorrect baseURL for Groq in client mode."] - }, - "date": "2024-06-23", - "version": "1.44.3" - }, - { - "children": { - "fixes": ["Create first-time user on server db."] - }, - "date": "2024-06-22", - "version": "1.44.2" - }, - { - "children": { - "fixes": ["Ollama not enabled client fetch by default."] - }, - "date": "2024-06-21", - "version": "1.44.1" - }, - { - "children": { - "features": ["Supoort Claude 3.5 Sonnet."], - "fixes": ["Fix to send image without text."], - "improvements": ["Fixed System Agent missing in mobile layout."] - }, - "date": "2024-06-20", - "version": "1.44.0" - }, - { - "children": { - "fixes": ["Fix and improve tool calling."] - }, - "date": "2024-06-19", - "version": "1.43.7" - }, - { - "children": { - "fixes": ["Fix auto avatar."] - }, - "date": "2024-06-19", - "version": "1.43.6" - }, - { - "children": { - "fixes": ["Typo in reset settings."], - "improvements": ["Fix phi3 icon display under OpenRouter."] - }, - "date": "2024-06-19", - "version": "1.43.5" - }, - { - "children": {}, - "date": "2024-06-19", - "version": "1.43.4" - }, - { - "children": { - "fixes": ["Fix hydration mismatch on macOS."], - "improvements": ["Update brand and improve docs."] - }, - "date": "2024-06-18", - "version": "1.43.3" - }, - { - "children": { - "improvements": ["Remove some i18n text."] - }, - "date": "2024-06-18", - "version": "1.43.2" - }, - { - "children": { - "fixes": [ - "Fix clerk UNAUTHORIZED auth error, fix onboard and auth for community version, openai key and openai proxy are invalid in feature flags." - ], - "improvements": ["Add stepfun as a new provider, fix footer being cut on wide screen."] - }, - "date": "2024-06-17", - "version": "1.43.1" - }, - { - "children": { - "features": [ - "Release LobeChat 1.0, support server db mode with Postgres / Drizzle ORM / tRPC." - ], - "improvements": ["Update LICENSE to Apache 2.0", "update to 1.0.0"] - }, - "date": "2024-06-17", - "version": "1.43.0" - }, - { - "children": { - "fixes": ["Fix issues for client fetch."] - }, - "date": "2024-06-16", - "version": "1.42.19" - }, - { - "children": { - "improvements": ["Update error card style, Update settings footer style and about page."] - }, - "date": "2024-06-14", - "version": "1.42.18" - }, - { - "children": { - "improvements": ["Add Qwen2 models, Add Zhipu new models."] - }, - "date": "2024-06-12", - "version": "1.42.17" - }, - { - "children": { - "improvements": ["Do not show noDescription in new sesstion, Expand Clear tooltip maxWidth."] - }, - "date": "2024-06-09", - "version": "1.42.16" - }, - { - "children": {}, - "date": "2024-06-08", - "version": "1.42.15" - }, - { - "children": { - "fixes": ["Fix OpenAi BaseURL in api form."] - }, - "date": "2024-06-07", - "version": "1.42.14" - }, - { - "children": { - "improvements": ["Refactor model provider implement."] - }, - "date": "2024-06-06", - "version": "1.42.13" - }, - { - "children": { - "fixes": ["Fix response.undefined error with some provider."] - }, - "date": "2024-06-04", - "version": "1.42.12" - }, - { - "children": {}, - "date": "2024-06-04", - "version": "1.42.11" - }, - { - "children": { - "fixes": ["Fix send button loading on only add user message."] - }, - "date": "2024-06-03", - "version": "1.42.10" - }, - { - "children": { - "improvements": ["Improve loading state."] - }, - "date": "2024-06-03", - "version": "1.42.9" - }, - { - "children": { - "improvements": ["Improve config upload modal."] - }, - "date": "2024-06-01", - "version": "1.42.8" - }, - { - "children": { - "improvements": ["Refactor session meta method."] - }, - "date": "2024-05-31", - "version": "1.42.7" - }, - { - "children": { - "improvements": ["Refactor the config import for server import."], - "fixes": ["Fix import config."] - }, - "date": "2024-05-29", - "version": "1.42.6" - }, - { - "children": { - "improvements": ["Refactor the settings to add optimistic updating."] - }, - "date": "2024-05-29", - "version": "1.42.5" - }, - { - "children": { - "improvements": [ - "Add optimistic loading for image uploading, improve display of set limited history messages, randomness and voice input." - ] - }, - "date": "2024-05-28", - "version": "1.42.4" - }, - { - "children": { - "fixes": ["Fix auto focus issues, fix the default agent not work correctly on new device."], - "improvements": ["Add SYSTEM_AGENT env."] - }, - "date": "2024-05-28", - "version": "1.42.3" - }, - { - "children": { - "improvements": ["Refactor agent store data."] - }, - "date": "2024-05-28", - "version": "1.42.2" - }, - { - "children": { - "improvements": ["Improve the display effect of plug-in API name and description."] - }, - "date": "2024-05-27", - "version": "1.42.1" - }, - { - "children": { - "features": ["Support topic agent."], - "fixes": ["Fix trpc/edge path error when setting NEXT_PUBLIC_BASE_PATH."] - }, - "date": "2024-05-27", - "version": "1.42.0" - }, - { - "children": { - "fixes": [ - "Fix the missing user id in chat compeletition and fix remove unstarred topic not working." - ] - }, - "date": "2024-05-27", - "version": "1.41.13" - }, - { - "children": { - "improvements": ["Fix PluginStore layout."], - "fixes": ["Fix connection checker, fix the batch import error."] - }, - "date": "2024-05-27", - "version": "1.41.12" - }, - { - "children": { - "fixes": ["Fix vercel build."] - }, - "date": "2024-05-26", - "version": "1.41.11" - }, - { - "children": { - "fixes": ["Fix aws log."], - "improvements": ["Update token tag popover style."] - }, - "date": "2024-05-25", - "version": "1.41.10" - }, - { - "children": { - "improvements": ["Migrate some agent config to chatConfig."], - "fixes": ["Fix bedrock show by default on vercel."] - }, - "date": "2024-05-25", - "version": "1.41.9" - }, - { - "children": { - "fixes": ["Fix avatar missing on client DB mode."] - }, - "date": "2024-05-24", - "version": "1.41.8" - }, - { - "children": { - "improvements": ["Refactor the global app status and fix PWA installer."] - }, - "date": "2024-05-24", - "version": "1.41.7" - }, - { - "children": { - "improvements": ["Improve PWA install guide."] - }, - "date": "2024-05-23", - "version": "1.41.6" - }, - { - "children": { - "fixes": ["Refactor user store and fix custom model list form."], - "improvements": ["Fix image style and improve drag upload box."] - }, - "date": "2024-05-23", - "version": "1.41.5" - }, - { - "children": { - "improvements": ["Refactor to serverDB ENV."] - }, - "date": "2024-05-22", - "version": "1.41.4" - }, - { - "children": { - "improvements": ["Move feature flags ENV."] - }, - "date": "2024-05-22", - "version": "1.41.3" - }, - { - "children": { - "improvements": ["Refactor the app ENV, refactor the langfuse env, refactor the llm env."], - "fixes": ["Fix market and plugin cache."] - }, - "date": "2024-05-22", - "version": "1.41.2" - }, - { - "children": { - "improvements": ["Fix setting modal style problem."] - }, - "date": "2024-05-22", - "version": "1.41.1" - }, - { - "children": { - "features": ["Add system agent to select another model provider for translation."], - "improvements": [ - "Tooltip should not be selected & Model selector overlaps with reset button." - ] - }, - "date": "2024-05-21", - "version": "1.41.0" - }, - { - "children": { - "improvements": [ - "Add ENABLED_OPENAI env, add fetch error notification, improve openai fetch client switch, improve redirect when login." - ], - "fixes": ["Fix duplicate model panel key."] - }, - "date": "2024-05-21", - "version": "1.40.4" - }, - { - "children": { - "improvements": ["Refactor analytics env."] - }, - "date": "2024-05-20", - "version": "1.40.3" - }, - { - "children": { - "improvements": ["Modify bedrock provided model."] - }, - "date": "2024-05-20", - "version": "1.40.2" - }, - { - "children": { - "fixes": ["Upgrade antd and fix lint type."], - "improvements": ["Add two feature flags: check_updates 、welcome_suggest."] - }, - "date": "2024-05-19", - "version": "1.40.1" - }, - { - "children": { - "features": ["Bump version and add enable ollama env."], - "fixes": ["Fix enable ollama env."] - }, - "date": "2024-05-18", - "version": "1.40.0" - }, - { - "children": { - "improvements": ["Refactor the create message flow to fix some bugs."] - }, - "date": "2024-05-15", - "version": "1.39.5" - }, - { - "children": { - "improvements": ["Add Gemini 1.5 Flash model."] - }, - "date": "2024-05-15", - "version": "1.39.4" - }, - { - "children": { - "improvements": ["Fix setting modal on responsive and some other style problem."] - }, - "date": "2024-05-15", - "version": "1.39.3" - }, - { - "children": { - "fixes": [ - "Fix agent config on page init, fix retry issue when hide page, login button not show on user panel, refresh model config form & mobile footer button lost." - ], - "improvements": [ - "Add GPT-4o model, fix scroll and expand, update perplexity models, updates 01.AI model list." - ] - }, - "date": "2024-05-15", - "version": "1.39.2" - }, - { - "children": { - "fixes": ["Fix DeepSeek using wrong model ID."] - }, - "date": "2024-05-14", - "version": "1.39.1" - }, - { - "children": { - "improvements": ["Move next-auth hooks to user store actions."], - "features": ["Support DeepSeek as new model provider."], - "fixes": ["Dragging text mistakenly as image, pin antd@5.17.0 to fix build error."] - }, - "date": "2024-05-14", - "version": "1.39.0" - }, - { - "children": { - "improvements": [ - "Add PWA install and metadata & ld generate, fix TelemetryNotification zindex." - ] - }, - "date": "2024-05-13", - "version": "1.38.1" - }, - { - "children": { - "features": ["Support thread in client pglite."] - }, - "date": "2024-12-23", - "version": "1.38.0" - }, - { - "children": { - "improvements": ["Move pglite to client service."] - }, - "date": "2024-12-22", - "version": "1.37.2" - }, - { - "children": { - "improvements": ["Refactor the client service to deprecated."] - }, - "date": "2024-12-22", - "version": "1.37.1" - }, - { - "children": { - "features": ["Support to use pglite as client db."] - }, - "date": "2024-12-22", - "version": "1.37.0" - }, - { - "children": { - "improvements": ["Refactor client mode upload to match server mode."] - }, - "date": "2024-12-21", - "version": "1.36.46" - }, - { - "children": { - "improvements": ["Add o1 model in GitHub models."] - }, - "date": "2024-12-21", - "version": "1.36.45" - }, - { - "children": { - "improvements": ["Add Gemini flash thinking model."] - }, - "date": "2024-12-21", - "version": "1.36.44" - }, - { - "children": {}, - "date": "2024-12-21", - "version": "1.36.43" - }, - { - "children": { - "fixes": ["Fix HUGGINGFACE endpoint url."] - }, - "date": "2024-12-21", - "version": "1.36.42" - }, - { - "children": { - "improvements": ["Upgrade react scan."] - }, - "date": "2024-12-21", - "version": "1.36.41" - }, - { - "children": { - "improvements": ["Seperate user keyVaults encrpyto from user model."] - }, - "date": "2024-12-20", - "version": "1.36.40" - }, - { - "children": { - "improvements": ["Refactor to use async headers()."] - }, - "date": "2024-12-20", - "version": "1.36.39" - }, - { - "children": { - "improvements": ["Refactor layout props."] - }, - "date": "2024-12-20", - "version": "1.36.38" - }, - { - "children": {}, - "date": "2024-12-19", - "version": "1.36.37" - }, - { - "children": {}, - "date": "2024-12-19", - "version": "1.36.36" - }, - { - "children": { - "improvements": ["Improve home page loading for better UX."] - }, - "date": "2024-12-18", - "version": "1.36.35" - }, - { - "children": { - "fixes": ["Fix pdf preview with capital ext."] - }, - "date": "2024-12-18", - "version": "1.36.34" - }, - { - "children": { - "fixes": ["Fix GitHub model fetch."] - }, - "date": "2024-12-18", - "version": "1.36.33" - }, - { - "children": { - "improvements": ["Refactor the drizzle code style."] - }, - "date": "2024-12-17", - "version": "1.36.32" - }, - { - "children": { - "improvements": ["Refactor the data fetch with clientDB init check."] - }, - "date": "2024-12-17", - "version": "1.36.31" - }, - { - "children": { - "improvements": ["Improve page loading state."] - }, - "date": "2024-12-16", - "version": "1.36.30" - }, - { - "children": { - "fixes": ["Fix discover locale with different default lang."] - }, - "date": "2024-12-16", - "version": "1.36.29" - }, - { - "children": {}, - "date": "2024-12-16", - "version": "1.36.28" - }, - { - "children": { - "fixes": ["Add unique keys to children."] - }, - "date": "2024-12-16", - "version": "1.36.27" - }, - { - "children": { - "improvements": ["Update models of Gitee AI provider."] - }, - "date": "2024-12-16", - "version": "1.36.26" - }, - { - "children": { - "improvements": ["Add new grok models."] - }, - "date": "2024-12-14", - "version": "1.36.25" - }, - { - "children": { - "improvements": ["Refactor file Url query in message model."] - }, - "date": "2024-12-14", - "version": "1.36.24" - }, - { - "children": { - "improvements": ["Support csv chunking."] - }, - "date": "2024-12-13", - "version": "1.36.23" - }, - { - "children": {}, - "date": "2024-12-13", - "version": "1.36.22" - }, - { - "children": {}, - "date": "2024-12-13", - "version": "1.36.21" - }, - { - "children": { - "improvements": ["Update locale."] - }, - "date": "2024-12-13", - "version": "1.36.20" - }, - { - "children": { - "fixes": ["One of Gemini functionCall error."] - }, - "date": "2024-12-13", - "version": "1.36.19" - }, - { - "children": { - "fixes": ["Fix claude first message can not be assistant."] - }, - "date": "2024-12-12", - "version": "1.36.18" - }, - { - "children": {}, - "date": "2024-12-12", - "version": "1.36.17" - }, - { - "children": { - "improvements": ["Refactor the file model method."] - }, - "date": "2024-12-12", - "version": "1.36.16" - }, - { - "children": { - "improvements": ["Enable googleSearch Tool for gemini-2.0-flash-exp."] - }, - "date": "2024-12-12", - "version": "1.36.15" - }, - { - "children": { - "improvements": ["Refactor database file model to remove server env."] - }, - "date": "2024-12-12", - "version": "1.36.14" - }, - { - "children": { - "improvements": ["Add Gemini 2.0 Flash Exp model."] - }, - "date": "2024-12-11", - "version": "1.36.13" - }, - { - "children": { - "improvements": ["Update sql and types."] - }, - "date": "2024-12-11", - "version": "1.36.12" - }, - { - "children": { - "improvements": ["Refactor data importer to repos."] - }, - "date": "2024-12-11", - "version": "1.36.11" - }, - { - "children": { - "improvements": ["Clean the gpt-4-vision-preview models."] - }, - "date": "2024-12-10", - "version": "1.36.10" - }, - { - "children": { - "improvements": ["Refactor the clerk user service implement."], - "fixes": ["Nullptr errors in NextAuth adapter."] - }, - "date": "2024-12-10", - "version": "1.36.9" - }, - { - "children": { - "improvements": ["Add GLM-4V-Flash from Zhipu."] - }, - "date": "2024-12-10", - "version": "1.36.8" - }, - { - "children": { - "fixes": ["Fix pricing with 0 digit."] - }, - "date": "2024-12-10", - "version": "1.36.7" - }, - { - "children": { - "improvements": ["Update groq, add llama3.3, Upgrade lobe-ui."] - }, - "date": "2024-12-10", - "version": "1.36.6" - }, - { - "children": { - "fixes": ["Fix wrong email linking in next-auth db adapter."] - }, - "date": "2024-12-09", - "version": "1.36.5" - }, - { - "children": {}, - "date": "2024-12-09", - "version": "1.36.4" - }, - { - "children": { - "fixes": ["Support request headers for chat."] - }, - "date": "2024-12-08", - "version": "1.36.3" - }, - { - "children": { - "improvements": ["Refactor async params route to adapt next15 breaking change."] - }, - "date": "2024-12-07", - "version": "1.36.2" - }, - { - "children": { - "improvements": ["Add gemini-exp-1206 model."] - }, - "date": "2024-12-07", - "version": "1.36.1" - }, - { - "children": { - "features": ["Add Higress ai model provider."] - }, - "date": "2024-12-06", - "version": "1.36.0" - }, - { - "children": { - "improvements": ["Refactor page params to adapt next15 breaking change."] - }, - "date": "2024-12-06", - "version": "1.35.14" - }, - { - "children": {}, - "date": "2024-12-06", - "version": "1.35.13" - }, - { - "children": { - "fixes": ["Fix typo of prompts."] - }, - "date": "2024-12-05", - "version": "1.35.12" - }, - { - "children": {}, - "date": "2024-12-04", - "version": "1.35.11" - }, - { - "children": { - "improvements": ["Refactor the server db model implement."] - }, - "date": "2024-12-03", - "version": "1.35.10" - }, - { - "children": {}, - "date": "2024-12-03", - "version": "1.35.9" - }, - { - "children": { - "improvements": ["Move schema and migration folder."] - }, - "date": "2024-12-03", - "version": "1.35.8" - }, - { - "children": {}, - "date": "2024-12-03", - "version": "1.35.7" - }, - { - "children": { - "improvements": ["Add QwQ 32B Preview model."] - }, - "date": "2024-12-02", - "version": "1.35.6" - }, - { - "children": { - "improvements": ["Deprecated the current client mode code."] - }, - "date": "2024-12-02", - "version": "1.35.5" - }, - { - "children": {}, - "date": "2024-12-02", - "version": "1.35.4" - }, - { - "children": { - "improvements": ["Add gpt-4o-2024-11-20 model."] - }, - "date": "2024-12-01", - "version": "1.35.3" - }, - { - "children": { - "improvements": ["Improve i18n."] - }, - "date": "2024-12-01", - "version": "1.35.2" - }, - { - "children": { - "improvements": ["Update ollama models."] - }, - "date": "2024-12-01", - "version": "1.35.1" - }, - { - "children": { - "features": ["Support ollama tools use."] - }, - "date": "2024-12-01", - "version": "1.35.0" - }, - { - "children": { - "improvements": ["Add gemini-1.0-pro-002."] - }, - "date": "2024-05-07", - "version": "1.34.7" - }, - { - "children": { - "improvements": [ - "Add QWEN_PROXY_URL support for Qwen, update model list, add qwq-32b-preview." - ] - }, - "date": "2024-12-01", - "version": "1.34.6" - }, - { - "children": { - "improvements": ["Add Google LearnLM model."] - }, - "date": "2024-11-28", - "version": "1.34.5" - }, - { - "children": { - "improvements": ["Add switch portal thread."] - }, - "date": "2024-11-27", - "version": "1.34.4" - }, - { - "children": { - "fixes": ["Fix fallback behavior of default mode in AgentRuntime."] - }, - "date": "2024-11-27", - "version": "1.34.3" - }, - { - "children": { - "improvements": ["Improve thread i18n locale."] - }, - "date": "2024-11-27", - "version": "1.34.2" - }, - { - "children": { - "fixes": ["Fix Qwen baseUrl calling."] - }, - "date": "2024-11-26", - "version": "1.34.1" - }, - { - "children": { - "features": ["Forkable Chat Mode."] - }, - "date": "2024-11-26", - "version": "1.34.0" - }, - { - "children": { - "improvements": ["Update the description translation of Gitee AI."] - }, - "date": "2024-11-26", - "version": "1.33.5" - }, - { - "children": { - "improvements": ["Refactor getLlmOptionsFromPayload from AgentRuntime."] - }, - "date": "2024-11-26", - "version": "1.33.4" - }, - { - "children": { - "fixes": ["Fix fetchOnClient functional for Moonshot."] - }, - "date": "2024-11-25", - "version": "1.33.3" - }, - { - "children": { - "fixes": ["Fix multi-turns tools calling."] - }, - "date": "2024-11-25", - "version": "1.33.2" - }, - { - "children": { - "improvements": ["Add gemini-exp-1121 model."] - }, - "date": "2024-11-25", - "version": "1.33.1" - }, - { - "children": { - "features": ["Add Gitee AI model provider."] - }, - "date": "2024-11-25", - "version": "1.33.0" - }, - { - "children": { - "improvements": ["Support to reset fetched models."] - }, - "date": "2024-11-24", - "version": "1.32.9" - }, - { - "children": { - "fixes": ["Fix XAI_PROXY_URL env missing."] - }, - "date": "2024-11-24", - "version": "1.32.8" - }, - { - "children": { - "fixes": ["Fix tool message display."] - }, - "date": "2024-11-24", - "version": "1.32.7" - }, - { - "children": {}, - "date": "2024-11-24", - "version": "1.32.6" - }, - { - "children": { - "improvements": ["Refactor the main chat."] - }, - "date": "2024-11-24", - "version": "1.32.5" - }, - { - "children": { - "improvements": ["Refactor the default locale."] - }, - "date": "2024-11-20", - "version": "1.32.4" - }, - { - "children": { - "improvements": [ - "Add grok-vision-beta model, update Mistral model list, add pixtral-large-latest." - ] - }, - "date": "2024-11-20", - "version": "1.32.3" - }, - { - "children": {}, - "date": "2024-11-19", - "version": "1.32.2" - }, - { - "children": { - "fixes": ["Keyword search for chat history & sessions."], - "improvements": ["Support o1 models using streaming."] - }, - "date": "2024-11-19", - "version": "1.32.1" - }, - { - "children": { - "features": ["Add support InternLM (书生浦语) provider."] - }, - "date": "2024-11-19", - "version": "1.32.0" - }, - { - "children": { - "fixes": ["Connection check logic."] - }, - "date": "2024-11-18", - "version": "1.31.11" - }, - { - "children": {}, - "date": "2024-11-16", - "version": "1.31.10" - }, - { - "children": { - "improvements": ["Add gemini-exp-1114 model."] - }, - "date": "2024-11-16", - "version": "1.31.9" - }, - { - "children": { - "improvements": ["Move ChatInput to features/ChatInput."] - }, - "date": "2024-11-15", - "version": "1.31.8" - }, - { - "children": { - "improvements": ["genServerLLMConfig function, get *_MODEL_LIST from env."] - }, - "date": "2024-11-15", - "version": "1.31.7" - }, - { - "children": { - "improvements": ["Refactor the chat conversation implement."] - }, - "date": "2024-11-13", - "version": "1.31.6" - }, - { - "children": { - "improvements": ["Update some provider modellist & fix ai360 baseurl."] - }, - "date": "2024-11-12", - "version": "1.31.5" - }, - { - "children": { - "improvements": ["Fix Cloudflare Workers AI Sort."] - }, - "date": "2024-11-12", - "version": "1.31.4" - }, - { - "children": { - "improvements": ["Refactor languageModel & DEFAULT_LLM_CONFIG generate."] - }, - "date": "2024-11-12", - "version": "1.31.3" - }, - { - "children": { - "improvements": ["Update deepseek model."] - }, - "date": "2024-11-12", - "version": "1.31.2" - }, - { - "children": { - "improvements": ["Fix Windows always showing scrollbar."] - }, - "date": "2024-11-12", - "version": "1.31.1" - }, - { - "children": { - "features": ["Add support xAI provider."] - }, - "date": "2024-11-11", - "version": "1.31.0" - }, - { - "children": { - "improvements": ["Rename globalStore to userStore."] - }, - "date": "2024-04-28", - "version": "1.30.6" - }, - { - "children": { - "fixes": ["Suport to fetch model list on client."] - }, - "date": "2024-04-28", - "version": "1.30.5" - }, - { - "children": { - "fixes": ["Fix /api/proxy internal proxy attack."] - }, - "date": "2024-04-28", - "version": "1.30.4" - }, - { - "children": { - "fixes": ["Fix the plugin string env and search error."] - }, - "date": "2024-04-27", - "version": "1.30.3" - }, - { - "children": { - "fixes": ["Fix display error when using DEFAULT_AGENT_CONFIG env, fix docker build."], - "improvements": ["Hide default model tag and show ollama provider by default."] - }, - "date": "2024-04-27", - "version": "1.30.2" - }, - { - "children": { - "fixes": ["Fix display error when using DEFAULT_AGENT_CONFIG env."] - }, - "date": "2024-04-27", - "version": "1.30.1" - }, - { - "children": { - "features": ["Support Cloudflare Workers AI."] - }, - "date": "2024-11-11", - "version": "1.30.0" - }, - { - "children": { - "improvements": ["Add Sonnet 3.5 v2 inference model to BedRock."] - }, - "date": "2024-11-10", - "version": "1.29.6" - }, - { - "children": { - "fixes": ["Fix summary range."] - }, - "date": "2024-11-10", - "version": "1.29.5" - }, - { - "children": { - "fixes": ["Disregard remoteModelCards when showModelFetcher is disabled."] - }, - "date": "2024-11-09", - "version": "1.29.4" - }, - { - "children": { - "fixes": ["Fix the display model of history summary."] - }, - "date": "2024-11-09", - "version": "1.29.3" - }, - { - "children": { - "improvements": ["Allow users to disable SSRF or set a whitelist."] - }, - "date": "2024-11-09", - "version": "1.29.2" - }, - { - "children": { - "fixes": ["Fix topic summary field on server db."] - }, - "date": "2024-11-09", - "version": "1.29.1" - }, - { - "children": { - "features": ["Support compress history messages."] - }, - "date": "2024-11-09", - "version": "1.29.0" - }, - { - "children": { - "improvements": ["Update Ollama model 240421."] - }, - "date": "2024-04-24", - "version": "1.28.8" - }, - { - "children": { - "improvements": ["Refactor for session server mode."] - }, - "date": "2024-04-23", - "version": "1.28.7" - }, - { - "children": { - "fixes": ["Fix env typo of MS Entra ID."] - }, - "date": "2024-11-09", - "version": "1.28.6" - }, - { - "children": {}, - "date": "2024-11-08", - "version": "1.28.5" - }, - { - "children": { - "fixes": ["Disable model fetch for GitHub."] - }, - "date": "2024-11-07", - "version": "1.28.4" - }, - { - "children": { - "improvements": ["Move portal code to features folder."] - }, - "date": "2024-11-06", - "version": "1.28.3" - }, - { - "children": { - "improvements": ["Refactor and clean some code."] - }, - "date": "2024-11-06", - "version": "1.28.2" - }, - { - "children": { - "improvements": ["Update database fields."] - }, - "date": "2024-11-06", - "version": "1.28.1" - }, - { - "children": { - "features": ["Support export as markdown and JSON."] - }, - "date": "2024-11-05", - "version": "1.28.0" - }, - { - "children": { - "improvements": ["Optimized file upload buttons and prompts."] - }, - "date": "2024-04-19", - "version": "1.27.20" - }, - { - "children": { - "improvements": [ - "Add M and B support max token in ModelInfoTags, improve aync session experience." - ] - }, - "date": "2024-04-18", - "version": "1.27.19" - }, - { - "children": { - "improvements": [ - "Add claude 3 opus to AWS Bedrock, remove custom models from providers, and update Perplexity model names." - ] - }, - "date": "2024-04-17", - "version": "1.27.18" - }, - { - "children": { - "improvements": ["Refactor service to a uniform interface."] - }, - "date": "2024-04-16", - "version": "1.27.17" - }, - { - "children": { - "improvements": ["Refactor the auth."], - "fixes": ["Fix tool call error with gpt-4-turbo."] - }, - "date": "2024-04-14", - "version": "1.27.16" - }, - { - "children": { - "improvements": ["Enable gemini-1.5-pro-latest model by default."] - }, - "date": "2024-04-14", - "version": "1.27.15" - }, - { - "children": { - "improvements": [ - "Move client db to a new folder, refactor the service with browser db invoke." - ] - }, - "date": "2024-04-14", - "version": "1.27.14" - }, - { - "children": { - "fixes": ["Support drag or copy to upload file by model ability."] - }, - "date": "2024-04-14", - "version": "1.27.13" - }, - { - "children": { - "fixes": ["Rename."] - }, - "date": "2024-04-13", - "version": "1.27.12" - }, - { - "children": {}, - "date": "2024-04-13", - "version": "1.27.11" - }, - { - "children": { - "fixes": ["Fix custom model list not display correctly, Google models."] - }, - "date": "2024-04-12", - "version": "1.27.10" - }, - { - "children": { - "fixes": ["Azure models."] - }, - "date": "2024-04-12", - "version": "1.27.9" - }, - { - "children": { - "improvements": ["Update README.md."] - }, - "date": "2024-04-12", - "version": "1.27.8" - }, - { - "children": { - "fixes": ["Pin next to 14.1.4 to fix deployment."] - }, - "date": "2024-04-12", - "version": "1.27.7" - }, - { - "children": { - "improvements": [ - "Add GPT-4-turbo and 2024-04-09 Turbo Vision model and mistral new model name." - ] - }, - "date": "2024-04-11", - "version": "1.27.6" - }, - { - "children": { - "fixes": ["Fix only search topics in current session."] - }, - "date": "2024-04-11", - "version": "1.27.5" - }, - { - "children": { - "fixes": ["Fix Azure deployed model name."] - }, - "date": "2024-04-11", - "version": "1.27.4" - }, - { - "children": { - "improvements": ["Add claude 3.5 haiku model."] - }, - "date": "2024-11-05", - "version": "1.27.3" - }, - { - "children": { - "fixes": ["Remove the 'resetConversation' hot key tip."], - "improvements": ["Improve group topic styles."] - }, - "date": "2024-11-05", - "version": "1.27.2" - }, - { - "children": { - "fixes": ["Fix /webapi/plugin/store server error."] - }, - "date": "2024-11-04", - "version": "1.27.1" - }, - { - "children": { - "features": ["Support group topics by time."] - }, - "date": "2024-11-04", - "version": "1.27.0" - }, - { - "children": { - "fixes": ["If enable login and not signed in, return unauthorized error."] - }, - "date": "2024-11-04", - "version": "1.26.21" - }, - { - "children": { - "improvements": ["Disable chunking button for unsupported files."] - }, - "date": "2024-11-04", - "version": "1.26.20" - }, - { - "children": { - "fixes": [ - "Fix duplicate key value violates unique constraint \"slug_user_id_unique\" when create inbox session." - ] - }, - "date": "2024-11-03", - "version": "1.26.19" - }, - { - "children": { - "fixes": ["Fix MS Entra ID and Azure AD authorization."] - }, - "date": "2024-11-03", - "version": "1.26.18" - }, - { - "children": { - "improvements": ["Improve server log on chat api."] - }, - "date": "2024-10-31", - "version": "1.26.17" - }, - { - "children": { - "fixes": ["Fix server Network connection lost error."] - }, - "date": "2024-10-31", - "version": "1.26.16" - }, - { - "children": { - "improvements": ["Refactor embedding as plain vector array."] - }, - "date": "2024-10-31", - "version": "1.26.15" - }, - { - "children": { - "improvements": ["Add fa-ir locale."] - }, - "date": "2024-10-30", - "version": "1.26.14" - }, - { - "children": { - "fixes": ["Fix the artifacts interface not scrolling."] - }, - "date": "2024-10-30", - "version": "1.26.13" - }, - { - "children": { - "fixes": ["Fix file image prompts in client mode."] - }, - "date": "2024-10-30", - "version": "1.26.12" - }, - { - "children": {}, - "date": "2024-10-29", - "version": "1.26.11" - }, - { - "children": { - "improvements": ["Refactor the aiChat slice actions."] - }, - "date": "2024-10-29", - "version": "1.26.10" - }, - { - "children": {}, - "date": "2024-10-29", - "version": "1.26.9" - }, - { - "children": { - "fixes": ["Update zhipu param process."] - }, - "date": "2024-10-29", - "version": "1.26.8" - }, - { - "children": { - "fixes": ["Remove PWA Install in Firefox and Arc."] - }, - "date": "2024-10-29", - "version": "1.26.7" - }, - { - "children": { - "improvements": ["Add Qwen2.5-72B-Instruct model on HF provider."] - }, - "date": "2024-10-29", - "version": "1.26.6" - }, - { - "children": { - "improvements": ["Improve user guide when user not login."] - }, - "date": "2024-10-29", - "version": "1.26.5" - }, - { - "children": { - "fixes": ["Remove the 'resetConversation' hot key."] - }, - "date": "2024-10-28", - "version": "1.26.4" - }, - { - "children": { - "fixes": ["Fix Huggingface API interrupting when the output exceeds 140 tokens."], - "improvements": ["Remove SenseChat-Vision model, due to model limitation."] - }, - "date": "2024-10-28", - "version": "1.26.3" - }, - { - "children": { - "fixes": ["Fix page not switch when clicking on the pin assistant."] - }, - "date": "2024-10-28", - "version": "1.26.2" - }, - { - "children": { - "improvements": ["Refactor the plugin prompts to xml format."] - }, - "date": "2024-10-27", - "version": "1.26.1" - }, - { - "children": { - "features": ["experimentally support to pin assistant to sidebar."] - }, - "date": "2024-10-27", - "version": "1.26.0" - }, - { - "children": { - "improvements": ["Refactor the model settings for more clean code."], - "fixes": ["Fix normalize russian locale."] - }, - "date": "2024-04-07", - "version": "1.25.11" - }, - { - "children": { - "fixes": ["Fix typo of azure-id sso provider."] - }, - "date": "2024-04-04", - "version": "1.25.10" - }, - { - "children": { - "fixes": [ - "Fix page crash when using browser as stt engine, fix page crash when using browser as the stt engine." - ] - }, - "date": "2024-04-03", - "version": "1.25.9" - }, - { - "children": {}, - "date": "2024-04-03", - "version": "1.25.8" - }, - { - "children": { - "improvements": ["Improve scrollbar style."] - }, - "date": "2024-04-02", - "version": "1.25.7" - }, - { - "children": { - "improvements": ["Refactor SSO providers."], - "fixes": ["Fix plugins dropdown menu overflow."] - }, - "date": "2024-04-02", - "version": "1.25.6" - }, - { - "children": { - "fixes": ["Fix DraggablePanel bar interfere with the operation of the scrollbar."] - }, - "date": "2024-04-02", - "version": "1.25.5" - }, - { - "children": {}, - "date": "2024-04-02", - "version": "1.25.4" - }, - { - "children": { - "fixes": ["Fix the issue of the switch assistant portal not closing."] - }, - "date": "2024-10-27", - "version": "1.25.3" - }, - { - "children": { - "improvements": ["Update stepfun models."] - }, - "date": "2024-10-27", - "version": "1.25.2" - }, - { - "children": { - "fixes": ["Fix modelList merge."] - }, - "date": "2024-10-26", - "version": "1.25.1" - }, - { - "children": { - "features": ["Support ZEN mode."] - }, - "date": "2024-10-26", - "version": "1.25.0" - }, - { - "children": { - "fixes": ["Fix google gemini pro 1.5."] - }, - "date": "2024-03-29", - "version": "1.24.3" - }, - { - "children": { - "improvements": ["Update Google Model list, add gemini-1.5-flash-8b."] - }, - "date": "2024-10-26", - "version": "1.24.2" - }, - { - "children": { - "improvements": ["Refactor the Google Gen AI."] - }, - "date": "2024-10-25", - "version": "1.24.1" - }, - { - "children": { - "features": ["Add SenseNova (商汤) model provider."] - }, - "date": "2024-10-25", - "version": "1.24.0" - }, - { - "children": { - "improvements": ["Add *_MODEL_LIST env to all models, update Spark model id & display name."] - }, - "date": "2024-10-25", - "version": "1.23.1" - }, - { - "children": { - "features": ["Support system agent config."] - }, - "date": "2024-10-25", - "version": "1.23.0" - }, - { - "children": { - "improvements": ["Add bedrock claude-3.5-sonnect-v2."] - }, - "date": "2024-10-25", - "version": "1.22.27" - }, - { - "children": { - "improvements": ["Fix some custom branding detail."] - }, - "date": "2024-10-23", - "version": "1.22.26" - }, - { - "children": { - "improvements": ["Remove unused user tables."] - }, - "date": "2024-10-23", - "version": "1.22.25" - }, - { - "children": { - "improvements": ["Support plugin flag."] - }, - "date": "2024-10-23", - "version": "1.22.24" - }, - { - "children": { - "improvements": ["Improve error i18n."] - }, - "date": "2024-10-23", - "version": "1.22.23" - }, - { - "children": { - "improvements": ["Improve i18n."] - }, - "date": "2024-10-23", - "version": "1.22.22" - }, - { - "children": { - "improvements": ["Refactor cookie/headers to async mode."] - }, - "date": "2024-10-23", - "version": "1.22.21" - }, - { - "children": { - "improvements": ["Add new claude-3.5-sonnet model."] - }, - "date": "2024-10-23", - "version": "1.22.20" - }, - { - "children": { - "improvements": ["Move responsive to server utils folder."] - }, - "date": "2024-10-22", - "version": "1.22.19" - }, - { - "children": {}, - "date": "2024-10-22", - "version": "1.22.18" - }, - { - "children": { - "improvements": ["Fix dynamic import in rsc layout."] - }, - "date": "2024-10-22", - "version": "1.22.17" - }, - { - "children": { - "fixes": ["Fix azure-ad."] - }, - "date": "2024-10-21", - "version": "1.22.16" - }, - { - "children": { - "improvements": ["Update format utils and shared layout."] - }, - "date": "2024-10-21", - "version": "1.22.15" - }, - { - "children": { - "improvements": ["Update wenxin 4.0 turbo model to latest."] - }, - "date": "2024-10-20", - "version": "1.22.14" - }, - { - "children": { - "improvements": [ - "Add Ministral model, update Together AI model list, add function call & vision." - ] - }, - "date": "2024-10-20", - "version": "1.22.13" - }, - { - "children": { - "improvements": ["Add Llama 3.1 Nemotron 70B model & reorder some provider model list."] - }, - "date": "2024-10-20", - "version": "1.22.12" - }, - { - "children": { - "improvements": ["Refactor azure ad to ms entra id."] - }, - "date": "2024-10-20", - "version": "1.22.11" - }, - { - "children": {}, - "date": "2024-10-20", - "version": "1.22.10" - }, - { - "children": { - "improvements": ["Update Fireworks AI model list."] - }, - "date": "2024-10-18", - "version": "1.22.9" - }, - { - "children": { - "improvements": ["Add Yi-Lightning model."] - }, - "date": "2024-10-17", - "version": "1.22.8" - }, - { - "children": { - "improvements": ["Add qwen vision model & update qwen2.5 72b to 128k for siliconcloud."] - }, - "date": "2024-10-17", - "version": "1.22.7" - }, - { - "children": { - "fixes": ["Fix images not go in to chat context."] - }, - "date": "2024-10-13", - "version": "1.22.6" - }, - { - "children": { - "improvements": ["Reorder github model list & updata info & add new model."] - }, - "date": "2024-10-13", - "version": "1.22.5" - }, - { - "children": { - "improvements": ["Separate message slice and aiChat slice."] - }, - "date": "2024-10-13", - "version": "1.22.4" - }, - { - "children": { - "improvements": ["Support multi-windows for PWA."] - }, - "date": "2024-10-13", - "version": "1.22.3" - }, - { - "children": { - "fixes": ["Allow use email as name in logto."] - }, - "date": "2024-10-13", - "version": "1.22.2" - }, - { - "children": { - "fixes": ["Fix function calling issue, disable stream when using tools."] - }, - "date": "2024-10-12", - "version": "1.22.1" - }, - { - "children": { - "improvements": ["Refactor the chat webapi."], - "features": ["Add HuggingFace Model Provider."] - }, - "date": "2024-10-12", - "version": "1.22.0" - }, - { - "children": {}, - "date": "2024-10-12", - "version": "1.21.16" - }, - { - "children": {}, - "date": "2024-10-12", - "version": "1.21.15" - }, - { - "children": { - "improvements": ["Fix artifacts render markdown."] - }, - "date": "2024-10-12", - "version": "1.21.14" - }, - { - "children": { - "improvements": ["Refactor agent runtime implement of stream and ZHIPU provider."] - }, - "date": "2024-10-11", - "version": "1.21.13" - }, - { - "children": { - "improvements": ["Refactor the jwt code."] - }, - "date": "2024-10-11", - "version": "1.21.12" - }, - { - "children": { - "improvements": ["Refactor the backend code for better organization."] - }, - "date": "2024-10-11", - "version": "1.21.11" - }, - { - "children": { - "improvements": ["Updata gpt-4o model info."] - }, - "date": "2024-10-11", - "version": "1.21.10" - }, - { - "children": { - "improvements": ["Update qwen vl model to latest."] - }, - "date": "2024-10-10", - "version": "1.21.9" - }, - { - "children": { - "fixes": ["Fix auto rewrite query when user message is too long."], - "improvements": ["Support yml in file chunk."] - }, - "date": "2024-10-08", - "version": "1.21.8" - }, - { - "children": { - "improvements": ["Refactor text-to-image endpoint."] - }, - "date": "2024-10-08", - "version": "1.21.7" - }, - { - "children": { - "improvements": ["Move backend api to (backend) folder group."], - "fixes": ["Fix txt-to-image api."] - }, - "date": "2024-10-05", - "version": "1.21.6" - }, - { - "children": { - "improvements": ["Support shadcn in Artifacts."] - }, - "date": "2024-10-05", - "version": "1.21.5" - }, - { - "children": { - "fixes": ["Fix recharts deps in the Artifacts React Renderer."] - }, - "date": "2024-10-02", - "version": "1.21.4" - }, - { - "children": { - "improvements": ["Move most /api to /webapi."] - }, - "date": "2024-10-01", - "version": "1.21.3" - }, - { - "children": { - "improvements": ["Adjust Wenxin icon size."] - }, - "date": "2024-10-01", - "version": "1.21.2" - }, - { - "children": {}, - "date": "2024-09-30", - "version": "1.21.1" - }, - { - "children": { - "features": ["Add wenxin model provider."] - }, - "date": "2024-09-30", - "version": "1.21.0" - }, - { - "children": {}, - "date": "2024-09-30", - "version": "1.20.8" - }, - { - "children": { - "improvements": ["Update groq model list."] - }, - "date": "2024-09-29", - "version": "1.20.7" - }, - { - "children": {}, - "date": "2024-09-29", - "version": "1.20.6" - }, - { - "children": {}, - "date": "2024-09-29", - "version": "1.20.5" - }, - { - "children": {}, - "date": "2024-09-28", - "version": "1.20.4" - }, - { - "children": { - "fixes": ["Improve delete orphan chunks when delete files."] - }, - "date": "2024-09-28", - "version": "1.20.3" - }, - { - "children": { - "improvements": ["Add zhipu glm-4-flashx model."] - }, - "date": "2024-09-27", - "version": "1.20.2" - }, - { - "children": {}, - "date": "2024-09-27", - "version": "1.20.1" - }, - { - "children": { - "features": ["Add Hunyuan(Tencent) model provider."] - }, - "date": "2024-09-27", - "version": "1.20.0" - }, - { - "children": { - "improvements": ["Add llama3.2 model for openrouter provider."] - }, - "date": "2024-09-27", - "version": "1.19.36" - }, - { - "children": { - "improvements": ["Add o1-preview and o1-mini model to github model provider."] - }, - "date": "2024-09-27", - "version": "1.19.35" - }, - { - "children": {}, - "date": "2024-09-26", - "version": "1.19.34" - }, - { - "children": { - "fixes": ["MiniMax output long content interrupted by non-existent error."], - "improvements": ["Update google provider model info."] - }, - "date": "2024-09-25", - "version": "1.19.33" - }, - { - "children": { - "improvements": ["Add function call for taichu_llm."] - }, - "date": "2024-09-25", - "version": "1.19.32" - }, - { - "children": { - "improvements": ["Add google gemini 1.5 002 series."] - }, - "date": "2024-09-24", - "version": "1.19.31" - }, - { - "children": { - "improvements": ["Disable taichu2.0 functioncall & default disable taichu2.0v model."] - }, - "date": "2024-09-24", - "version": "1.19.30" - }, - { - "children": { - "improvements": ["Update taichu provider info & add taichu vision model."] - }, - "date": "2024-09-24", - "version": "1.19.29" - }, - { - "children": { - "improvements": ["Add function call support for Stepfun."] - }, - "date": "2024-09-24", - "version": "1.19.28" - }, - { - "children": { - "improvements": ["Improve images display in chat messages."] - }, - "date": "2024-09-24", - "version": "1.19.27" - }, - { - "children": { - "fixes": ["Fix url config import after user state init."], - "improvements": ["Add support function call for 360AI, left sidebar has only assistants."] - }, - "date": "2024-09-24", - "version": "1.19.26" - }, - { - "children": { - "fixes": ["Add missing translations."] - }, - "date": "2024-09-24", - "version": "1.19.25" - }, - { - "children": { - "fixes": ["Fix artifacts code language highlight."] - }, - "date": "2024-09-23", - "version": "1.19.24" - }, - { - "children": { - "improvements": ["Add spark max-32k model."] - }, - "date": "2024-09-23", - "version": "1.19.23" - }, - { - "children": { - "fixes": ["Fix ollama model download panel."] - }, - "date": "2024-09-22", - "version": "1.19.22" - }, - { - "children": { - "improvements": ["Refactor to improve branding customization."] - }, - "date": "2024-09-21", - "version": "1.19.21" - }, - { - "children": { - "fixes": ["Fix Content-Security-Policy."] - }, - "date": "2024-09-21", - "version": "1.19.20" - }, - { - "children": { - "fixes": ["Casdoor webhooks providerAccountId not found."] - }, - "date": "2024-09-21", - "version": "1.19.19" - }, - { - "children": {}, - "date": "2024-09-21", - "version": "1.19.18" - }, - { - "children": { - "fixes": ["providerAccountId not exist in provider."] - }, - "date": "2024-09-21", - "version": "1.19.17" - }, - { - "children": { - "improvements": ["Improve i18n for discover and improve version check."] - }, - "date": "2024-09-21", - "version": "1.19.16" - }, - { - "children": { - "improvements": ["Improve i18n in discover."] - }, - "date": "2024-09-20", - "version": "1.19.15" - }, - { - "children": { - "fixes": ["Pin shiki@1.17.7 to fix code highlight."] - }, - "date": "2024-09-20", - "version": "1.19.14" - }, - { - "children": { - "fixes": ["Try to implement better ssrf-protect."] - }, - "date": "2024-09-20", - "version": "1.19.13" - }, - { - "children": { - "improvements": ["Support webhooks for casdoor."] - }, - "date": "2024-09-20", - "version": "1.19.12" - }, - { - "children": { - "fixes": ["Custom model initialization not taking effect error."] - }, - "date": "2024-09-20", - "version": "1.19.11" - }, - { - "children": { - "improvements": ["Add qwen2.5 math and coder model for siliconcloud provider."] - }, - "date": "2024-09-20", - "version": "1.19.10" - }, - { - "children": { - "fixes": ["Fix a bug with server agent config when user not exist."] - }, - "date": "2024-09-20", - "version": "1.19.9" - }, - { - "children": { - "improvements": ["Delete siliconflow outdated model & disable stepfun functioncall."] - }, - "date": "2024-09-19", - "version": "1.19.8" - }, - { - "children": { - "improvements": ["Add siliconflow qwen2.5 model."] - }, - "date": "2024-09-19", - "version": "1.19.7" - }, - { - "children": { - "improvements": ["Refactor the tts route url."] - }, - "date": "2024-09-19", - "version": "1.19.6" - }, - { - "children": { - "improvements": ["Enable functioncall for stepfun models, Update qwen models."] - }, - "date": "2024-09-19", - "version": "1.19.5" - }, - { - "children": { - "improvements": ["Refactor the sitemap implement."] - }, - "date": "2024-09-19", - "version": "1.19.4" - }, - { - "children": {}, - "date": "2024-09-19", - "version": "1.19.3" - }, - { - "children": { - "improvements": ["Updata qwen model info & add qwen2.5 & reorder provider list."] - }, - "date": "2024-09-19", - "version": "1.19.2" - }, - { - "children": { - "improvements": ["Add mistral provider new models."] - }, - "date": "2024-09-19", - "version": "1.19.1" - }, - { - "children": { - "features": [ - "Add Ai21Labs model provider, add Github Models provider, support native Artifacts just like Claude." - ] - }, - "date": "2024-09-18", - "version": "1.19.0" - }, - { - "children": { - "fixes": ["Fix InterceptingRoutes in discover."] - }, - "date": "2024-09-18", - "version": "1.18.2" - }, - { - "children": {}, - "date": "2024-09-18", - "version": "1.18.1" - }, - { - "children": { - "features": ["Add Discover Page."] - }, - "date": "2024-09-18", - "version": "1.18.0" - }, - { - "children": { - "fixes": ["Fix a corner case of tools_call with empty object."], - "improvements": ["Delete duplicate models in ollama."] - }, - "date": "2024-09-16", - "version": "1.17.7" - }, - { - "children": { - "improvements": ["Rename artifacts to plugins in portal."] - }, - "date": "2024-09-15", - "version": "1.17.6" - }, - { - "children": { - "improvements": ["Add MiniCPM-V 8B model entries to Ollama model providers."] - }, - "date": "2024-09-15", - "version": "1.17.5" - }, - { - "children": { - "improvements": ["Update fullscreen loading style."] - }, - "date": "2024-09-15", - "version": "1.17.4" - }, - { - "children": { - "improvements": ["Delete \"-\" in deepseek displayname."] - }, - "date": "2024-09-14", - "version": "1.17.3" - }, - { - "children": { - "fixes": ["Fix o1 model list."], - "improvements": ["Update openrouter model list."] - }, - "date": "2024-09-13", - "version": "1.17.2" - }, - { - "children": { - "improvements": ["Update zhipu model info."] - }, - "date": "2024-09-13", - "version": "1.17.1" - }, - { - "children": { - "features": ["Support openai new OpenAI o1-preview/o1-mini models."], - "improvements": ["Support Google Model List."] - }, - "date": "2024-09-13", - "version": "1.17.0" - }, - { - "children": {}, - "date": "2024-09-13", - "version": "1.16.14" - }, - { - "children": { - "improvements": ["Update siliconcloud model."] - }, - "date": "2024-09-13", - "version": "1.16.13" - }, - { - "children": { - "improvements": ["Remove brackets from model names with dates in OpenAI."] - }, - "date": "2024-09-12", - "version": "1.16.12" - }, - { - "children": { - "fixes": ["Support webhooks for logto."], - "improvements": ["Default disable mistral provider useless models."] - }, - "date": "2024-09-12", - "version": "1.16.11" - }, - { - "children": { - "improvements": ["Support Environment Variable Inference For NextAuth."], - "fixes": ["Qwen model param error."] - }, - "date": "2024-09-12", - "version": "1.16.10" - }, - { - "children": { - "improvements": ["Add model and provider desc and url."] - }, - "date": "2024-09-12", - "version": "1.16.9" - }, - { - "children": { - "improvements": ["Improve models and add more info for providers and models."] - }, - "date": "2024-09-12", - "version": "1.16.8" - }, - { - "children": { - "improvements": ["Optimize model token display method."] - }, - "date": "2024-09-11", - "version": "1.16.7" - }, - { - "children": { - "fixes": ["Pin next@14.2.8 to fix Internal error."] - }, - "date": "2024-09-11", - "version": "1.16.6" - }, - { - "children": {}, - "date": "2024-09-11", - "version": "1.16.5" - }, - { - "children": {}, - "date": "2024-09-11", - "version": "1.16.4" - }, - { - "children": { - "fixes": ["Add LLM_VISION_IMAGE_USE_BASE64 to support local s3 in vision model."] - }, - "date": "2024-09-11", - "version": "1.16.3" - }, - { - "children": { - "improvements": ["Update Upstage model list."] - }, - "date": "2024-09-11", - "version": "1.16.2" - }, - { - "children": { - "improvements": [ - "Reorder the provider list, update spark check model to spark-lite & default disable useless model." - ] - }, - "date": "2024-09-10", - "version": "1.16.1" - }, - { - "children": { - "features": ["Add Fireworks AI Model Provider, Add Spark model provider."] - }, - "date": "2024-09-10", - "version": "1.16.0" - }, - { - "children": { - "improvements": ["Update CustomLogo."] - }, - "date": "2024-09-10", - "version": "1.15.35" - }, - { - "children": { - "improvements": ["Change empty content stream behavior."] - }, - "date": "2024-09-10", - "version": "1.15.34" - }, - { - "children": { - "fixes": ["Fix /etc/resolv.confedit permission in docker image."] - }, - "date": "2024-09-10", - "version": "1.15.33" - }, - { - "children": { - "fixes": ["Fix tools calling in some edge cases."] - }, - "date": "2024-09-10", - "version": "1.15.32" - }, - { - "children": { - "fixes": ["Baichuan should not introduce freequency_penality parameters."] - }, - "date": "2024-09-10", - "version": "1.15.31" - }, - { - "children": { - "fixes": ["Fix claude 3.5 image with s3 url."] - }, - "date": "2024-09-09", - "version": "1.15.30" - }, - { - "children": { - "fixes": ["Gemini cannot input images when server database is enabled."] - }, - "date": "2024-09-09", - "version": "1.15.29" - }, - { - "children": { - "fixes": ["Update baichuan param."] - }, - "date": "2024-09-09", - "version": "1.15.28" - }, - { - "children": { - "improvements": ["Add siliconcloud new model."] - }, - "date": "2024-09-09", - "version": "1.15.27" - }, - { - "children": { - "improvements": ["Update perplexity model list."] - }, - "date": "2024-09-09", - "version": "1.15.26" - }, - { - "children": {}, - "date": "2024-09-09", - "version": "1.15.25" - }, - { - "children": { - "improvements": ["Fix title in about settings."] - }, - "date": "2024-09-09", - "version": "1.15.24" - }, - { - "children": { - "improvements": ["Improve branding implement."] - }, - "date": "2024-09-08", - "version": "1.15.23" - }, - { - "children": { - "improvements": ["Update model display name & Remove Qwen preview model."] - }, - "date": "2024-09-08", - "version": "1.15.22" - }, - { - "children": { - "improvements": ["Temperature range from 0 to 2."] - }, - "date": "2024-09-08", - "version": "1.15.21" - }, - { - "children": {}, - "date": "2024-09-08", - "version": "1.15.20" - }, - { - "children": {}, - "date": "2024-09-08", - "version": "1.15.19" - }, - { - "children": { - "improvements": ["Support anthropic browser request."] - }, - "date": "2024-09-06", - "version": "1.15.18" - }, - { - "children": { - "fixes": ["Fix auth log."] - }, - "date": "2024-09-06", - "version": "1.15.17" - }, - { - "children": { - "improvements": ["Update Bedrock model list & add AWS_BEDROCK_MODEL_LIST support."] - }, - "date": "2024-09-06", - "version": "1.15.16" - }, - { - "children": { - "improvements": ["Add LLaVA 1.5 7B model in Groq."] - }, - "date": "2024-09-06", - "version": "1.15.15" - }, - { - "children": {}, - "date": "2024-09-06", - "version": "1.15.14" - }, - { - "children": {}, - "date": "2024-09-06", - "version": "1.15.13" - }, - { - "children": { - "fixes": ["Fix typo in RAG prompt."] - }, - "date": "2024-09-04", - "version": "1.15.12" - }, - { - "children": {}, - "date": "2024-09-04", - "version": "1.15.11" - }, - { - "children": {}, - "date": "2024-09-03", - "version": "1.15.10" - }, - { - "children": { - "fixes": ["Fix speed and rag prompt."] - }, - "date": "2024-09-03", - "version": "1.15.9" - }, - { - "children": { - "fixes": ["Fix .PDF can not be chunked."] - }, - "date": "2024-09-03", - "version": "1.15.8" - }, - { - "children": { - "improvements": ["Fix provider disabled title style."] - }, - "date": "2024-09-03", - "version": "1.15.7" - }, - { - "children": { - "improvements": [ - "Stepfun default enabled model, update Groq model list & add GROQ_MODEL_LIST support." - ] - }, - "date": "2024-09-01", - "version": "1.15.6" - }, - { - "children": { - "improvements": ["Update Together AI model list."] - }, - "date": "2024-09-01", - "version": "1.15.5" - }, - { - "children": { - "improvements": ["Update Novita AI model info & add NOVITA_MODEL_LIST support."] - }, - "date": "2024-09-01", - "version": "1.15.4" - }, - { - "children": { - "improvements": [ - "Add *_MODEL_LIST for Qwen and ZeroOne, fix model info, update Claude 3.5 Sonnet maxOutput vaule." - ] - }, - "date": "2024-09-01", - "version": "1.15.3" - }, - { - "children": { - "improvements": ["Update Qwen and Gemini models info."] - }, - "date": "2024-08-30", - "version": "1.15.2" - }, - { - "children": { - "improvements": ["Update the sorting of each provider model."] - }, - "date": "2024-08-30", - "version": "1.15.1" - }, - { - "children": { - "features": ["Add Upstage model provider support."] - }, - "date": "2024-08-30", - "version": "1.15.0" - }, - { - "children": { - "improvements": ["Fix ms doc file preview, Update the sorting of each provider model."] - }, - "date": "2024-08-30", - "version": "1.14.12" - }, - { - "children": { - "improvements": ["Update Stepfun models info."] - }, - "date": "2024-08-30", - "version": "1.14.11" - }, - { - "children": { - "fixes": ["Fix file relative chunks."] - }, - "date": "2024-08-30", - "version": "1.14.10" - }, - { - "children": {}, - "date": "2024-08-29", - "version": "1.14.9" - }, - { - "children": { - "fixes": ["Fix whisper-1 typo."] - }, - "date": "2024-08-29", - "version": "1.14.8" - }, - { - "children": { - "fixes": ["Disable ChatGPT-4o Tools Calling."], - "improvements": ["Improve chunk and file preview."] - }, - "date": "2024-08-28", - "version": "1.14.7" - }, - { - "children": { - "improvements": ["Update Gemini models."] - }, - "date": "2024-08-28", - "version": "1.14.6" - }, - { - "children": { - "fixes": ["No user name if Cloudflare Zero Trust with onetimepin."] - }, - "date": "2024-08-28", - "version": "1.14.5" - }, - { - "children": { - "improvements": ["Move model and provider icon components to @lobehub/icons."] - }, - "date": "2024-08-28", - "version": "1.14.4" - }, - { - "children": { - "fixes": ["Improve aysnc error type."] - }, - "date": "2024-08-27", - "version": "1.14.3" - }, - { - "children": { - "fixes": ["Fix agent setting."] - }, - "date": "2024-08-27", - "version": "1.14.2" - }, - { - "children": { - "improvements": ["Improve zhipu model config."] - }, - "date": "2024-08-27", - "version": "1.14.1" - }, - { - "children": { - "features": ["Supports Cloudflare Zero Trust login."] - }, - "date": "2024-08-27", - "version": "1.14.0" - }, - { - "children": { - "fixes": ["Bypass vercel deployment protection, fix can send message on uploading files."] - }, - "date": "2024-08-27", - "version": "1.13.2" - }, - { - "children": { - "improvements": ["Update Qwen models."] - }, - "date": "2024-08-27", - "version": "1.13.1" - }, - { - "children": { - "features": ["Supports Authelia login."] - }, - "date": "2024-08-27", - "version": "1.13.0" - }, - { - "children": { - "fixes": ["Feature flag knowledge_base doesn't affect ActionBar."] - }, - "date": "2024-08-26", - "version": "1.12.20" - }, - { - "children": { - "fixes": ["Fix cannot clone agent when imported from client."] - }, - "date": "2024-08-25", - "version": "1.12.19" - }, - { - "children": { - "fixes": ["Fix dayjs error in en-US language."] - }, - "date": "2024-08-25", - "version": "1.12.18" - }, - { - "children": { - "fixes": ["Fix multi file upload dupicate."] - }, - "date": "2024-08-25", - "version": "1.12.17" - }, - { - "children": { - "fixes": ["Session not found error on mobile."] - }, - "date": "2024-08-24", - "version": "1.12.16" - }, - { - "children": {}, - "date": "2024-08-24", - "version": "1.12.15" - }, - { - "children": { - "fixes": ["Fix tts file saving in server mode."] - }, - "date": "2024-08-24", - "version": "1.12.14" - }, - { - "children": { - "improvements": ["Update 01.AI models."] - }, - "date": "2024-08-24", - "version": "1.12.13" - }, - { - "children": {}, - "date": "2024-08-24", - "version": "1.12.12" - }, - { - "children": { - "fixes": ["Remove orphan chunks if there is no related file."] - }, - "date": "2024-08-23", - "version": "1.12.11" - }, - { - "children": { - "fixes": ["Refactor and fix dalle."] - }, - "date": "2024-08-23", - "version": "1.12.10" - }, - { - "children": { - "fixes": ["Improve s3 path-style url."] - }, - "date": "2024-08-23", - "version": "1.12.9" - }, - { - "children": { - "fixes": ["Fix NEXT_PUBLIC_S3_DOMAIN error on Docker."] - }, - "date": "2024-08-22", - "version": "1.12.8" - }, - { - "children": { - "fixes": ["Logout button not shown on mobile view when using nextauth."] - }, - "date": "2024-08-22", - "version": "1.12.7" - }, - { - "children": { - "improvements": ["Refactor s3 env and support path-style for minio."] - }, - "date": "2024-08-22", - "version": "1.12.6" - }, - { - "children": { - "fixes": ["Fix clipboard copy issue and improve upload cors feedback."] - }, - "date": "2024-08-22", - "version": "1.12.5" - }, - { - "children": { - "improvements": ["Fix link style."] - }, - "date": "2024-08-22", - "version": "1.12.4" - }, - { - "children": { - "improvements": ["Hide settings in repo."] - }, - "date": "2024-08-22", - "version": "1.12.3" - }, - { - "children": {}, - "date": "2024-08-22", - "version": "1.12.2" - }, - { - "children": { - "fixes": ["Fix embeddings multi-insert when there is issues with async task."] - }, - "date": "2024-08-21", - "version": "1.12.1" - }, - { - "children": { - "features": ["Files and knowledge base."] - }, - "date": "2024-08-21", - "version": "1.12.0" - }, - { - "children": { - "fixes": ["Fixed bedrock llama model id."] - }, - "date": "2024-08-19", - "version": "1.11.9" - }, - { - "children": { - "improvements": ["Update zhipu models."] - }, - "date": "2024-08-19", - "version": "1.11.8" - }, - { - "children": { - "fixes": ["Fix topic scroll issue."] - }, - "date": "2024-08-18", - "version": "1.11.7" - }, - { - "children": { - "improvements": ["Refactor the SITE_URL to APP_URL."] - }, - "date": "2024-08-18", - "version": "1.11.6" - }, - { - "children": { - "improvements": ["Refactor the fetch method to fix response.undefined."] - }, - "date": "2024-08-18", - "version": "1.11.5" - }, - { - "children": { - "improvements": [ - "Add SILICONCLOUD_MODEL_LIST & SILICONCLOUD_PROXY_URL support for SiliconCloud." - ] - }, - "date": "2024-08-18", - "version": "1.11.4" - }, - { - "children": { - "improvements": ["Refactor PanelTitle and move commit from file uploading."] - }, - "date": "2024-08-17", - "version": "1.11.3" - }, - { - "children": {}, - "date": "2024-08-17", - "version": "1.11.2" - }, - { - "children": { - "fixes": ["Make S3 upload ACL setting optional."] - }, - "date": "2024-08-15", - "version": "1.11.1" - }, - { - "children": { - "features": ["Add 2 new models to openai provider."] - }, - "date": "2024-08-14", - "version": "1.11.0" - }, - { - "children": {}, - "date": "2024-08-14", - "version": "1.10.1" - }, - { - "children": { - "features": ["Add SiliconCloud model provider."] - }, - "date": "2024-08-14", - "version": "1.10.0" - }, - { - "children": { - "improvements": ["Resize the image size in chat message."] - }, - "date": "2024-08-13", - "version": "1.9.8" - }, - { - "children": {}, - "date": "2024-08-13", - "version": "1.9.7" - }, - { - "children": {}, - "date": "2024-08-09", - "version": "1.9.6" - }, - { - "children": { - "improvements": ["Updated AWS bedrock model list."] - }, - "date": "2024-08-08", - "version": "1.9.5" - }, - { - "children": { - "fixes": ["Fix import clerk AuthObject from public api."] - }, - "date": "2024-08-06", - "version": "1.9.4" - }, - { - "children": { - "improvements": ["Refactor server db schema for better code organize."] - }, - "date": "2024-08-06", - "version": "1.9.3" - }, - { - "children": {}, - "date": "2024-08-05", - "version": "1.9.2" - }, - { - "children": { - "fixes": ["Azure modelTag icon display."] - }, - "date": "2024-08-05", - "version": "1.9.1" - }, - { - "children": { - "features": ["Skip login page if only one provider exists."] - }, - "date": "2024-08-05", - "version": "1.9.0" - }, - { - "children": { - "fixes": ["Add PROXY_URL in docker with proxychains-ng."] - }, - "date": "2024-08-03", - "version": "1.8.2" - }, - { - "children": { - "improvements": ["Fix aya, mathstral model tag icon & update ollama model info."] - }, - "date": "2024-08-03", - "version": "1.8.1" - }, - { - "children": { - "features": ["Add NextAuth as authentication service in server database."] - }, - "date": "2024-08-02", - "version": "1.8.0" - }, - { - "children": { - "improvements": ["Add Gemini 1.5 Pro Exp model."] - }, - "date": "2024-08-02", - "version": "1.7.10" - }, - { - "children": { - "fixes": ["Fix Mistral models calling & update model info."], - "improvements": ["Fix stepfun & baichuan model tag icon missing, update Perplexity models."] - }, - "date": "2024-08-01", - "version": "1.7.9" - }, - { - "children": {}, - "date": "2024-07-30", - "version": "1.7.8" - }, - { - "children": { - "improvements": ["Improve tools calling UI."] - }, - "date": "2024-07-30", - "version": "1.7.7" - }, - { - "children": { - "fixes": ["Disable anthropic browser request."] - }, - "date": "2024-07-29", - "version": "1.7.6" - }, - { - "children": { - "fixes": ["Fix create_session edit_agent feature flags and add more flags."], - "improvements": ["Update 360GPT model (360GPT2 Pro)."] - }, - "date": "2024-07-29", - "version": "1.7.5" - }, - { - "children": { - "fixes": ["Improve remote model list fetching for Novita AI."] - }, - "date": "2024-07-29", - "version": "1.7.4" - }, - { - "children": { - "fixes": ["Update minimax models."] - }, - "date": "2024-07-28", - "version": "1.7.3" - }, - { - "children": { - "fixes": ["Avoid baseURL being an empty string, resulting in incorrect client fetch."] - }, - "date": "2024-07-26", - "version": "1.7.2" - }, - { - "children": { - "fixes": ["Fix dalle tools calling prompts to avoid content risk."] - }, - "date": "2024-07-26", - "version": "1.7.1" - }, - { - "children": { - "features": ["Enabled function calling on Deepseek models."] - }, - "date": "2024-07-26", - "version": "1.7.0" - }, - { - "children": { - "improvements": ["Fix file upload height."] - }, - "date": "2024-07-26", - "version": "1.6.15" - }, - { - "children": { - "improvements": ["Improve input file upload."] - }, - "date": "2024-07-26", - "version": "1.6.14" - }, - { - "children": { - "improvements": ["Updated Groq model list to include llama-3.1 and llama3-Groq."] - }, - "date": "2024-07-25", - "version": "1.6.13" - }, - { - "children": { - "improvements": ["Add new models to groq which are llama 3.1."] - }, - "date": "2024-07-25", - "version": "1.6.12" - }, - { - "children": { - "fixes": ["Fix UNAUTHORIZED issue with clerk auth provider."] - }, - "date": "2024-07-24", - "version": "1.6.11" - }, - { - "children": { - "improvements": ["Fix the scrolling of the return result area of function calling."] - }, - "date": "2024-07-23", - "version": "1.6.10" - }, - { - "children": {}, - "date": "2024-07-23", - "version": "1.6.9" - }, - { - "children": { - "improvements": ["Move server modules."] - }, - "date": "2024-07-23", - "version": "1.6.8" - }, - { - "children": { - "improvements": ["Add new model provider Novita AI."] - }, - "date": "2024-07-23", - "version": "1.6.7" - }, - { - "children": {}, - "date": "2024-07-22", - "version": "1.6.6" - }, - { - "children": { - "fixes": ["Content lost unexpectedly on Qwen provider when finish_reason is stop."] - }, - "date": "2024-07-22", - "version": "1.6.5" - }, - { - "children": { - "improvements": ["Add trpc query client with react-query."] - }, - "date": "2024-07-21", - "version": "1.6.4" - }, - { - "children": { - "improvements": ["Update Zhipu models (GLM-4-AllTools & CodeGeeX-4)."] - }, - "date": "2024-07-21", - "version": "1.6.3" - }, - { - "children": { - "fixes": ["Fix dayjs render on server."] - }, - "date": "2024-07-21", - "version": "1.6.2" - }, - { - "children": { - "improvements": ["Refactor the DragUpload."] - }, - "date": "2024-07-19", - "version": "1.6.1" - }, - { - "children": { - "features": ["Add gpt-4o-mini in OpenAI Provider and set it as the default model."] - }, - "date": "2024-07-19", - "version": "1.6.0" - }, - { - "children": { - "improvements": ["Added Gemma2 instead of outdated Gemma."] - }, - "date": "2024-07-19", - "version": "1.5.5" - }, - { - "children": { - "fixes": ["Fix delete session group."] - }, - "date": "2024-07-17", - "version": "1.5.4" - }, - { - "children": { - "fixes": ["Fix OpenAI deployment restrictions, fix cant duplicate assistant."] - }, - "date": "2024-07-17", - "version": "1.5.3" - }, - { - "children": { - "fixes": ["Fix session not reorder after send message."] - }, - "date": "2024-07-17", - "version": "1.5.2" - }, - { - "children": { - "improvements": ["Improve brand url."] - }, - "date": "2024-07-17", - "version": "1.5.1" - }, - { - "children": { - "features": ["Spport qwen-vl and tool call for qwen."] - }, - "date": "2024-07-17", - "version": "1.5.0" - }, - { - "children": { - "fixes": ["Save assistant info on blur."] - }, - "date": "2024-07-15", - "version": "1.4.3" - }, - { - "children": { - "improvements": ["Remove code related to gemini-pro-vision."] - }, - "date": "2024-07-13", - "version": "1.4.2" - }, - { - "children": { - "improvements": ["Add cloud promotion banner."] - }, - "date": "2024-07-13", - "version": "1.4.1" - }, - { - "children": { - "features": ["Add 360AI model provider."] - }, - "date": "2024-07-12", - "version": "1.4.0" - }, - { - "children": { - "improvements": ["Improve agent runtime code."] - }, - "date": "2024-07-11", - "version": "1.3.6" - }, - { - "children": { - "fixes": ["Fix assistant meta change race issue."] - }, - "date": "2024-07-10", - "version": "1.3.5" - }, - { - "children": { - "improvements": ["Support disable clientFetch by default."] - }, - "date": "2024-07-09", - "version": "1.3.4" - }, - { - "children": { - "fixes": ["Allow user to use their own WebRTC signaling."] - }, - "date": "2024-07-09", - "version": "1.3.3" - }, - { - "children": { - "fixes": ["Automatic refresh when change language."] - }, - "date": "2024-07-09", - "version": "1.3.2" - }, - { - "children": {}, - "date": "2024-07-09", - "version": "1.3.1" - }, - { - "children": { - "features": ["Add Taichu model provider."] - }, - "date": "2024-07-09", - "version": "1.3.0" - }, - { - "children": { - "improvements": ["Provider changes with model in model settings."] - }, - "date": "2024-07-08", - "version": "1.2.14" - }, - { - "children": { - "fixes": ["Fix tool message order."] - }, - "date": "2024-07-07", - "version": "1.2.13" - }, - { - "children": { - "fixes": [ - "Fixed mobile web page navigation issue with inbox assistant, support to disable clerk signup with feature flag." - ] - }, - "date": "2024-07-07", - "version": "1.2.12" - }, - { - "children": { - "improvements": ["Update deepseek max token."] - }, - "date": "2024-07-07", - "version": "1.2.11" - }, - { - "children": { - "fixes": ["Improve tools calling error feedback when arguments are not correctly."] - }, - "date": "2024-07-05", - "version": "1.2.10" - }, - { - "children": { - "improvements": ["Fix tool message suspense loading."] - }, - "date": "2024-07-04", - "version": "1.2.9" - }, - { - "children": { - "fixes": ["Allow builtin tools to trigger AI message."] - }, - "date": "2024-07-03", - "version": "1.2.8" - }, - { - "children": { - "improvements": ["Improve delete assistant message with tools."] - }, - "date": "2024-07-03", - "version": "1.2.7" - }, - { - "children": { - "fixes": ["Clerk provider refreshes continously."] - }, - "date": "2024-07-03", - "version": "1.2.6" - }, - { - "children": { - "fixes": ["Fix clerk appearance is not applied correctly."] - }, - "date": "2024-07-02", - "version": "1.2.5" - }, - { - "children": { - "improvements": ["Update ProviderAvatar for Baichuan & Stepfun."] - }, - "date": "2024-07-02", - "version": "1.2.4" - }, - { - "children": { - "improvements": ["Refactor the portal implement."] - }, - "date": "2024-07-01", - "version": "1.2.3" - }, - { - "children": { - "fixes": ["Display issue when select default model in System Agent."] - }, - "date": "2024-07-01", - "version": "1.2.2" - }, - { - "children": { - "improvements": ["Refactor the portal file."] - }, - "date": "2024-07-01", - "version": "1.2.1" - }, - { - "children": { - "features": ["Add Baichuan model provider."] - }, - "date": "2024-07-01", - "version": "1.2.0" - }, - { - "children": { - "improvements": ["Add Gemini 1.5 stable version model & 2M context window."] - }, - "date": "2024-07-01", - "version": "1.1.18" - }, - { - "children": { - "improvements": ["Refactor to prepare for Chat Portal."] - }, - "date": "2024-07-01", - "version": "1.1.17" - }, - { - "children": { - "fixes": ["Fix clerk UNAUTHORIZED error after long-time hang-up."] - }, - "date": "2024-06-29", - "version": "1.1.16" - }, - { - "children": { - "fixes": ["Pin @azure/core-rest-pipeline@1.16.0 to fix azure openai issue."] - }, - "date": "2024-06-28", - "version": "1.1.15" - }, - { - "children": { - "improvements": ["Refactor plugin state."] - }, - "date": "2024-06-27", - "version": "1.1.14" - }, - { - "children": { - "fixes": ["-check_updates cannot be set by FEATURE_FLAGS."] - }, - "date": "2024-06-27", - "version": "1.1.13" - }, - { - "children": { - "fixes": ["Fix azure tools calling."] - }, - "date": "2024-06-26", - "version": "1.1.12" - }, - { - "children": { - "improvements": ["Refactor format utils."] - }, - "date": "2024-06-25", - "version": "1.1.11" - }, - { - "children": {}, - "date": "2024-06-24", - "version": "1.1.10" - }, - { - "children": { - "fixes": ["Fix agent tags."], - "improvements": ["Always show action on mobile."] - }, - "date": "2024-06-24", - "version": "1.1.9" - }, - { - "children": { - "fixes": ["Fix anthropic parallel tools calling."] - }, - "date": "2024-06-24", - "version": "1.1.8" - }, - { - "children": { - "improvements": ["Add AES-GCM footer and tooltip."] - }, - "date": "2024-06-24", - "version": "1.1.7" - }, - { - "children": { - "improvements": ["Refactor the server db implement."], - "fixes": ["Fix incorrect baseURL for Groq in client mode."] - }, - "date": "2024-06-23", - "version": "1.1.6" - }, - { - "children": { - "improvements": ["Remove deprecated env."] - }, - "date": "2024-06-23", - "version": "1.1.5" - }, - { - "children": { - "fixes": ["Create first-time user on server db."] - }, - "date": "2024-06-22", - "version": "1.1.4" - }, - { - "children": { - "fixes": ["Ollama not enabled client fetch by default."] - }, - "date": "2024-06-21", - "version": "1.1.3" - }, - { - "children": {}, - "date": "2024-06-20", - "version": "1.1.2" - }, - { - "children": { - "improvements": ["Fixed System Agent missing in mobile layout."] - }, - "date": "2024-06-20", - "version": "1.1.1" - }, - { - "children": { - "features": ["Supoort Claude 3.5 Sonnet."] - }, - "date": "2024-06-20", - "version": "1.1.0" - }, - { - "children": { - "fixes": ["Fix to send image without text."] - }, - "date": "2024-06-20", - "version": "1.0.14" - }, - { - "children": { - "fixes": ["Fix and improve tool calling."] - }, - "date": "2024-06-19", - "version": "1.0.13" - }, - { - "children": { - "fixes": ["Fix auto avatar."] - }, - "date": "2024-06-19", - "version": "1.0.12" - }, - { - "children": { - "improvements": ["Fix phi3 icon display under OpenRouter."] - }, - "date": "2024-06-19", - "version": "1.0.11" - }, - { - "children": { - "fixes": ["Typo in reset settings."] - }, - "date": "2024-06-19", - "version": "1.0.10" - }, - { - "children": {}, - "date": "2024-06-19", - "version": "1.0.9" - }, - { - "children": { - "fixes": ["Fix hydration mismatch on macOS."], - "improvements": ["Update brand and improve docs."] - }, - "date": "2024-06-18", - "version": "1.0.8" - }, - { - "children": { - "improvements": ["Remove some i18n text."] - }, - "date": "2024-06-18", - "version": "1.0.7" - }, - { - "children": { - "improvements": ["Fix footer being cut on wide screen."] - }, - "date": "2024-06-17", - "version": "1.0.6" - }, - { - "children": { - "fixes": ["Fix onboard and auth for community version."] - }, - "date": "2024-06-17", - "version": "1.0.5" - }, - { - "children": { - "improvements": ["Add stepfun as a new provider."] - }, - "date": "2024-06-17", - "version": "1.0.4" - }, - { - "children": { - "fixes": ["Fix clerk UNAUTHORIZED auth error."] - }, - "date": "2024-06-17", - "version": "1.0.3" - }, - { - "children": { - "fixes": ["Openai key and openai proxy are invalid in feature flags."] - }, - "date": "2024-06-17", - "version": "1.0.2" - }, - { - "children": { - "improvements": ["Remove r2 cdn url."] - }, - "date": "2024-06-17", - "version": "1.0.1" - }, - { - "children": { - "features": ["Release LobeChat 1.0."], - "improvements": [ - "Bump version.", - "Release LobeChat 1.0, closes #2897 (97bb377)", - "Bump version, closes #2896 (6e2d03b)", - "Update LICENSE to Apache 2.0", - "update to 1.0.0" - ] - }, - "date": "2024-06-17", - "version": "1.0.0" - }, - { - "children": { - "fixes": ["Fix qwen model id and improve anthropic logo text color."] - }, - "date": "2024-03-10", - "version": "0.133.2" - }, - { - "children": { - "fixes": ["Fix sitemap config."] - }, - "date": "2024-03-08", - "version": "0.133.1" - }, - { - "children": { - "features": ["Support Mistral model provider."] - }, - "date": "2024-03-07", - "version": "0.133.0" + "date": "2024-12-13", + "version": "1.36.23" }, { - "children": { - "fixes": ["Fix anthropic streaming on Vercel/Cloudflare."] - }, - "date": "2024-03-07", - "version": "0.132.2" + "children": {}, + "date": "2024-12-13", + "version": "1.36.22" }, { - "children": { - "fixes": ["Fix hydration error while OAuth callback."] - }, - "date": "2024-03-06", - "version": "0.132.1" + "children": {}, + "date": "2024-12-13", + "version": "1.36.21" }, { "children": { - "features": ["Support anthropic as model provider."] + "improvements": ["Update locale."] }, - "date": "2024-03-05", - "version": "0.132.0" + "date": "2024-12-13", + "version": "1.36.20" }, { "children": { - "features": ["Support langfuse integration."] + "fixes": ["One of Gemini functionCall error."] }, - "date": "2024-03-05", - "version": "0.131.0" + "date": "2024-12-13", + "version": "1.36.19" }, { "children": { - "improvements": ["Update gpt-3.5-turbo model card."] + "fixes": ["Fix claude first message can not be assistant."] }, - "date": "2024-03-03", - "version": "0.130.7" + "date": "2024-12-12", + "version": "1.36.18" }, { - "children": { - "improvements": ["Refactor the plugin and tool slice."] - }, - "date": "2024-03-01", - "version": "0.130.6" + "children": {}, + "date": "2024-12-12", + "version": "1.36.17" }, { "children": { - "improvements": ["Support switch model with tag."] + "improvements": ["Refactor the file model method."] }, - "date": "2024-03-01", - "version": "0.130.5" + "date": "2024-12-12", + "version": "1.36.16" }, { "children": { - "improvements": ["Refactor the core chatStream and plugin gateway auth."] + "improvements": ["Enable googleSearch Tool for gemini-2.0-flash-exp."] }, - "date": "2024-02-29", - "version": "0.130.4" + "date": "2024-12-12", + "version": "1.36.15" }, { "children": { - "improvements": ["Refactor the google api route and add more tests for chat route."] + "improvements": ["Refactor database file model to remove server env."] }, - "date": "2024-02-29", - "version": "0.130.3" + "date": "2024-12-12", + "version": "1.36.14" }, { "children": { - "fixes": ["Update azure OpenAI api version options to latest."] + "improvements": ["Add Gemini 2.0 Flash Exp model."] }, - "date": "2024-02-29", - "version": "0.130.2" - }, - { - "children": {}, - "date": "2024-02-28", - "version": "0.130.1" + "date": "2024-12-11", + "version": "1.36.13" }, { "children": { - "features": ["Support multiple API Keys."] + "improvements": ["Update sql and types."] }, - "date": "2024-02-27", - "version": "0.130.0" + "date": "2024-12-11", + "version": "1.36.12" }, { "children": { - "fixes": ["Fix github url."] + "improvements": ["Refactor data importer to repos."] }, - "date": "2024-02-25", - "version": "0.129.6" + "date": "2024-12-11", + "version": "1.36.11" }, { "children": { - "fixes": [ - "Fix eliminate UI jitter on navigation, improving experience for users sensitive to motion." - ] + "improvements": ["Clean the gpt-4-vision-preview models."] }, - "date": "2024-02-25", - "version": "0.129.5" - }, - { - "children": {}, - "date": "2024-02-24", - "version": "0.129.4" + "date": "2024-12-10", + "version": "1.36.10" }, { "children": { - "improvements": ["Add gemma model logo for ollama."] + "improvements": ["Refactor the clerk user service implement."], + "fixes": ["Nullptr errors in NextAuth adapter."] }, - "date": "2024-02-23", - "version": "0.129.3" + "date": "2024-12-10", + "version": "1.36.9" }, { "children": { - "fixes": ["Fix OAuth don't get user id from session."] + "improvements": ["Add GLM-4V-Flash from Zhipu."] }, - "date": "2024-02-23", - "version": "0.129.2" - }, - { - "children": {}, - "date": "2024-02-22", - "version": "0.129.1" + "date": "2024-12-10", + "version": "1.36.8" }, { "children": { - "features": ["Support perplexity AI provider."] + "fixes": ["Fix pricing with 0 digit."] }, - "date": "2024-02-22", - "version": "0.129.0" + "date": "2024-12-10", + "version": "1.36.7" }, { "children": { - "fixes": ["Fix the robots.txt config."] + "improvements": ["Update groq, add llama3.3, Upgrade lobe-ui."] }, - "date": "2024-02-21", - "version": "0.128.10" + "date": "2024-12-10", + "version": "1.36.6" }, { "children": { - "fixes": ["Fix the robots.txt config."] + "fixes": ["Fix wrong email linking in next-auth db adapter."] }, - "date": "2024-02-20", - "version": "0.128.9" + "date": "2024-12-09", + "version": "1.36.5" }, { "children": {}, - "date": "2024-02-20", - "version": "0.128.8" - }, - { - "children": { - "improvements": ["Improve docs url and add more docs."] - }, - "date": "2024-02-20", - "version": "0.128.7" + "date": "2024-12-09", + "version": "1.36.4" }, { "children": { - "fixes": ["Fix OAuth errors on Docker deployment."] + "fixes": ["Support request headers for chat."] }, - "date": "2024-02-20", - "version": "0.128.6" + "date": "2024-12-08", + "version": "1.36.3" }, { "children": { - "fixes": ["Fix the document url."] + "improvements": ["Refactor async params route to adapt next15 breaking change."] }, - "date": "2024-02-18", - "version": "0.128.5" + "date": "2024-12-07", + "version": "1.36.2" }, { "children": { - "fixes": ["Fix documents i18n."] + "improvements": ["Add gemini-exp-1206 model."] }, - "date": "2024-02-18", - "version": "0.128.4" + "date": "2024-12-07", + "version": "1.36.1" }, { "children": { - "improvements": ["Refactor with chat docs site."] + "features": ["Add Higress ai model provider."] }, - "date": "2024-02-18", - "version": "0.128.3" + "date": "2024-12-06", + "version": "1.36.0" }, { "children": { - "fixes": ["Fix agent avatar click wrong navigation."] + "improvements": ["Refactor page params to adapt next15 breaking change."] }, - "date": "2024-02-15", - "version": "0.128.2" + "date": "2024-12-06", + "version": "1.35.14" }, { - "children": { - "fixes": ["Fix auto lang switch."] - }, - "date": "2024-02-15", - "version": "0.128.1" + "children": {}, + "date": "2024-12-06", + "version": "1.35.13" }, { "children": { - "features": ["Support define default agent config with DEFAULT_AGENT_CONFIG ENV."] + "fixes": ["Fix typo of prompts."] }, - "date": "2024-02-14", - "version": "0.128.0" + "date": "2024-12-05", + "version": "1.35.12" }, { - "children": { - "improvements": ["Refactor the sidebar to fix first render state."] - }, - "date": "2024-02-14", - "version": "0.127.2" + "children": {}, + "date": "2024-12-04", + "version": "1.35.11" }, { "children": { - "improvements": ["Improve settings tabs style and refactor the LLM setting page."] + "improvements": ["Refactor the server db model implement."] }, - "date": "2024-02-14", - "version": "0.127.1" + "date": "2024-12-03", + "version": "1.35.10" }, { "children": {}, - "date": "2024-02-13", - "version": "0.127.0" + "date": "2024-12-03", + "version": "1.35.9" }, { "children": { - "improvements": ["Refactor with the auth code."], - "fixes": ["Fix middleware auth console error."] + "improvements": ["Move schema and migration folder."] }, - "date": "2024-02-12", - "version": "0.126.5" + "date": "2024-12-03", + "version": "1.35.8" }, { - "children": { - "improvements": ["Update Model provider request url."], - "fixes": ["Fix auth error in console, fix token tag usage display."] - }, - "date": "2024-02-11", - "version": "0.126.4" + "children": {}, + "date": "2024-12-03", + "version": "1.35.7" }, { "children": { - "fixes": ["Fix auth layout error."] + "improvements": ["Add QwQ 32B Preview model."] }, - "date": "2024-02-09", - "version": "0.126.3" + "date": "2024-12-02", + "version": "1.35.6" }, { "children": { - "fixes": ["Fix OAuth throws an error on Vercel deploy."] + "improvements": ["Deprecated the current client mode code."] }, - "date": "2024-02-09", - "version": "0.126.2" + "date": "2024-12-02", + "version": "1.35.5" }, { - "children": { - "fixes": ["Add basePath to support subdirectory."] - }, - "date": "2024-02-09", - "version": "0.126.1" + "children": {}, + "date": "2024-12-02", + "version": "1.35.4" }, { "children": { - "features": ["Support umami analytics."], - "fixes": [ - "The back button on the chat setting page can correctly return to the configured Agent chat page." - ] + "improvements": ["Add gpt-4o-2024-11-20 model."] }, - "date": "2024-02-09", - "version": "0.126.0" + "date": "2024-12-01", + "version": "1.35.3" }, { "children": { - "features": ["Support login & session authentication via OAuth 2.0 (Auth0)."] + "improvements": ["Improve i18n."] }, - "date": "2024-02-08", - "version": "0.125.0" + "date": "2024-12-01", + "version": "1.35.2" }, { "children": { - "fixes": ["Fix use azure api key error."] + "improvements": ["Update ollama models."] }, - "date": "2024-02-07", - "version": "0.124.3" + "date": "2024-12-01", + "version": "1.35.1" }, { "children": { - "improvements": ["Add moonshot i18n."] + "features": ["Support ollama tools use."] }, - "date": "2024-02-06", - "version": "0.124.2" + "date": "2024-12-01", + "version": "1.35.0" }, { "children": { - "improvements": ["Improve direction UX."] + "improvements": [ + "Add QWEN_PROXY_URL support for Qwen, update model list, add qwq-32b-preview." + ] }, - "date": "2024-02-06", - "version": "0.124.1" + "date": "2024-12-01", + "version": "1.34.6" }, { "children": { - "features": ["Support Moonshot AI Provider."] + "improvements": ["Add Google LearnLM model."] }, - "date": "2024-02-06", - "version": "0.124.0" + "date": "2024-11-28", + "version": "1.34.5" }, { "children": { - "improvements": ["Improve clear topic tips."] + "improvements": ["Add switch portal thread."] }, - "date": "2024-02-06", - "version": "0.123.4" + "date": "2024-11-27", + "version": "1.34.4" }, { "children": { - "fixes": ["Fix non-https crypto.subtile missing error."] + "fixes": ["Fix fallback behavior of default mode in AgentRuntime."] }, - "date": "2024-02-06", - "version": "0.123.3" + "date": "2024-11-27", + "version": "1.34.3" }, { "children": { - "fixes": ["Fix docker build."] + "improvements": ["Improve thread i18n locale."] }, - "date": "2024-02-06", - "version": "0.123.2" + "date": "2024-11-27", + "version": "1.34.2" }, { "children": { - "fixes": ["Improve auth control of plugin gateways, update dockerfile."], - "improvements": ["Add gpt-4-all feature flag."] + "fixes": ["Fix Qwen baseUrl calling."] }, - "date": "2024-02-05", - "version": "0.123.1" + "date": "2024-11-26", + "version": "1.34.1" }, { "children": { - "features": ["Support Google / Zhipu / AWS Bedrock model providers."] + "features": ["Forkable Chat Mode."] }, - "date": "2024-02-05", - "version": "0.123.0" - }, - { - "children": {}, - "date": "2024-02-05", - "version": "0.122.9" + "date": "2024-11-26", + "version": "1.34.0" }, { "children": { - "improvements": ["Allow user to add agent without redirection."] + "improvements": ["Update the description translation of Gitee AI."] }, - "date": "2024-02-03", - "version": "0.122.8" + "date": "2024-11-26", + "version": "1.33.5" }, { "children": { - "improvements": ["Update the gpt-4-1106-preview model to gpt-4-0125-preview."] + "improvements": ["Refactor getLlmOptionsFromPayload from AgentRuntime."] }, - "date": "2024-02-02", - "version": "0.122.7" - }, - { - "children": {}, - "date": "2024-01-31", - "version": "0.122.6" + "date": "2024-11-26", + "version": "1.33.4" }, { "children": { - "fixes": ["The plugin has a hallucination and gets stuck."] + "fixes": ["Fix fetchOnClient functional for Moonshot."] }, - "date": "2024-01-31", - "version": "0.122.5" + "date": "2024-11-25", + "version": "1.33.3" }, { "children": { - "fixes": ["Fix plugin gateway auth."] + "fixes": ["Fix multi-turns tools calling."] }, - "date": "2024-01-30", - "version": "0.122.4" + "date": "2024-11-25", + "version": "1.33.2" }, { "children": { - "improvements": ["Refactor the setting storage from localStorage to indexedDB."] + "improvements": ["Add gemini-exp-1121 model."] }, - "date": "2024-01-30", - "version": "0.122.3" + "date": "2024-11-25", + "version": "1.33.1" }, { "children": { - "fixes": ["Fix unexpected topic switch when favoriting topic."] + "features": ["Add Gitee AI model provider."] }, - "date": "2024-01-30", - "version": "0.122.2" + "date": "2024-11-25", + "version": "1.33.0" }, { "children": { - "improvements": ["Fix antd tab width flicker when show function debug."] + "improvements": ["Support to reset fetched models."] }, - "date": "2024-01-29", - "version": "0.122.1" + "date": "2024-11-24", + "version": "1.32.9" }, { "children": { - "features": ["Add create agent action in group menu."] + "fixes": ["Fix XAI_PROXY_URL env missing."] }, - "date": "2024-01-29", - "version": "0.122.0" + "date": "2024-11-24", + "version": "1.32.8" }, { "children": { - "fixes": ["Pin ahooks to fix test ci and settings crash."] + "fixes": ["Fix tool message display."] }, - "date": "2024-01-29", - "version": "0.121.4" + "date": "2024-11-24", + "version": "1.32.7" }, { - "children": { - "improvements": ["Improve stop loading icon."] - }, - "date": "2024-01-26", - "version": "0.121.3" + "children": {}, + "date": "2024-11-24", + "version": "1.32.6" }, { "children": { - "improvements": ["Remove centered prop from CreateGroupModal."] + "improvements": ["Refactor the main chat."] }, - "date": "2024-01-25", - "version": "0.121.2" + "date": "2024-11-24", + "version": "1.32.5" }, { "children": { - "fixes": ["Automatically fill in the wrong password."], - "improvements": ["Fix default plugins height unstabled when scrolling."] + "improvements": ["Refactor the default locale."] }, - "date": "2024-01-24", - "version": "0.121.1" + "date": "2024-11-20", + "version": "1.32.4" }, { "children": { - "features": ["Add session group manager."] + "improvements": [ + "Add grok-vision-beta model, update Mistral model list, add pixtral-large-latest." + ] }, - "date": "2024-01-24", - "version": "0.121.0" + "date": "2024-11-20", + "version": "1.32.3" }, { - "children": { - "improvements": ["Fix share image tags not align."] - }, - "date": "2024-01-22", - "version": "0.120.6" + "children": {}, + "date": "2024-11-19", + "version": "1.32.2" }, { "children": { - "improvements": ["Update locale and add test for globalStore."] + "fixes": ["Keyword search for chat history & sessions."], + "improvements": ["Support o1 models using streaming."] }, - "date": "2024-01-21", - "version": "0.120.5" + "date": "2024-11-19", + "version": "1.32.1" }, { "children": { - "fixes": ["Refactor url state management and fix some detail experience."] + "features": ["Add support InternLM (书生浦语) provider."] }, - "date": "2024-01-21", - "version": "0.120.4" + "date": "2024-11-19", + "version": "1.32.0" }, { "children": { - "improvements": ["Refactor antd i18n and improve locale order."] + "fixes": ["Connection check logic."] }, - "date": "2024-01-19", - "version": "0.120.3" + "date": "2024-11-18", + "version": "1.31.11" + }, + { + "children": {}, + "date": "2024-11-16", + "version": "1.31.10" }, { "children": { - "fixes": ["Fix setPluginMessage can not stop create ai message."] + "improvements": ["Add gemini-exp-1114 model."] }, - "date": "2024-01-17", - "version": "0.120.2" + "date": "2024-11-16", + "version": "1.31.9" }, { "children": { - "fixes": ["Fix list scrolling white screen on mobile."] + "improvements": ["Move ChatInput to features/ChatInput."] }, - "date": "2024-01-16", - "version": "0.120.1" + "date": "2024-11-15", + "version": "1.31.8" }, { "children": { - "features": ["Adding Arabic Language Support."] + "improvements": ["genServerLLMConfig function, get *_MODEL_LIST from env."] }, - "date": "2024-01-15", - "version": "0.120.0" + "date": "2024-11-15", + "version": "1.31.7" }, { "children": { - "improvements": ["Add delete and regenerate for function message."] + "improvements": ["Refactor the chat conversation implement."] }, - "date": "2024-01-10", - "version": "0.119.13" + "date": "2024-11-13", + "version": "1.31.6" }, { "children": { - "fixes": ["Fix new line after sending messages with enter key."] + "improvements": ["Update some provider modellist & fix ai360 baseurl."] }, - "date": "2024-01-09", - "version": "0.119.12" + "date": "2024-11-12", + "version": "1.31.5" }, { "children": { - "improvements": ["Refactor ChatInput to support cmd+enter."] + "improvements": ["Fix Cloudflare Workers AI Sort."] }, - "date": "2024-01-09", - "version": "0.119.11" + "date": "2024-11-12", + "version": "1.31.4" }, { "children": { - "fixes": ["Debug information cannot be selected."] + "improvements": ["Refactor languageModel & DEFAULT_LLM_CONFIG generate."] }, - "date": "2024-01-08", - "version": "0.119.10" + "date": "2024-11-12", + "version": "1.31.3" }, { "children": { - "improvements": ["Fix ChatInput fullscreen display not correct."] + "improvements": ["Update deepseek model."] }, - "date": "2024-01-08", - "version": "0.119.9" + "date": "2024-11-12", + "version": "1.31.2" }, { "children": { - "fixes": ["Fix spotting tool call correctly."] + "improvements": ["Fix Windows always showing scrollbar."] }, - "date": "2024-01-07", - "version": "0.119.8" + "date": "2024-11-12", + "version": "1.31.1" }, { "children": { - "improvements": ["Improve share modal style."] + "features": ["Add support xAI provider."] }, - "date": "2024-01-07", - "version": "0.119.7" + "date": "2024-11-11", + "version": "1.31.0" }, { "children": { - "improvements": ["Improve conversation style."] + "features": ["Support Cloudflare Workers AI."] }, - "date": "2024-01-06", - "version": "0.119.6" + "date": "2024-11-11", + "version": "1.30.0" }, { "children": { - "improvements": ["Fix topic i18n."] + "improvements": ["Add Sonnet 3.5 v2 inference model to BedRock."] }, - "date": "2024-01-06", - "version": "0.119.5" + "date": "2024-11-10", + "version": "1.29.6" }, { "children": { - "improvements": [ - "Fix BackBottom zIndex, improve chat list on mobile, improve chat list scrolling to bottom at initial render, improve custom model input, improve topic scroll." - ], - "fixes": ["Fix auto scroll error and BackBottom error."] + "fixes": ["Fix summary range."] }, - "date": "2024-01-06", - "version": "0.119.4" + "date": "2024-11-10", + "version": "1.29.5" }, { "children": { - "fixes": ["Fix deploy error. Changed SquareAsterisk to AsteriskSquare."] + "fixes": ["Disregard remoteModelCards when showModelFetcher is disabled."] }, - "date": "2024-01-06", - "version": "0.119.3" + "date": "2024-11-09", + "version": "1.29.4" }, { "children": { - "fixes": ["Fix function call error with smooth animation."] + "fixes": ["Fix the display model of history summary."] }, - "date": "2024-01-05", - "version": "0.119.2" - }, - { - "children": {}, - "date": "2024-01-05", - "version": "0.119.1" + "date": "2024-11-09", + "version": "1.29.3" }, { "children": { - "improvements": ["Refactor the ChatList."], - "features": [ - "Support auto rename topic, support delete and regenerate message, support duplicate session, support duplicate topic." - ], - "fixes": ["Fix can't uninstall custom plugin in custom plugin modal."] + "improvements": ["Allow users to disable SSRF or set a whitelist."] }, - "date": "2024-01-04", - "version": "0.119.0" + "date": "2024-11-09", + "version": "1.29.2" }, { "children": { - "fixes": ["Add chat defaultNS."] + "fixes": ["Fix topic summary field on server db."] }, - "date": "2024-01-03", - "version": "0.118.10" + "date": "2024-11-09", + "version": "1.29.1" }, { "children": { - "improvements": ["Add leaving protect alert."] + "features": ["Support compress history messages."] }, - "date": "2024-01-03", - "version": "0.118.9" + "date": "2024-11-09", + "version": "1.29.0" }, { "children": { - "improvements": ["Add Vietnamese files and add the vi-VN option in the General Settings."] + "fixes": ["Fix env typo of MS Entra ID."] }, - "date": "2024-01-03", - "version": "0.118.8" + "date": "2024-11-09", + "version": "1.28.6" }, { - "children": { - "fixes": ["Desensitize openai base url in the error response."] - }, - "date": "2024-01-03", - "version": "0.118.7" + "children": {}, + "date": "2024-11-08", + "version": "1.28.5" }, { "children": { - "improvements": ["Migration the ChatList into Conversation."] + "fixes": ["Disable model fetch for GitHub."] }, - "date": "2024-01-03", - "version": "0.118.6" + "date": "2024-11-07", + "version": "1.28.4" }, { "children": { - "fixes": ["Mobile device return to the previous page error."] + "improvements": ["Move portal code to features folder."] }, - "date": "2024-01-02", - "version": "0.118.5" + "date": "2024-11-06", + "version": "1.28.3" }, { "children": { - "fixes": ["Update dalle identifier to fix unstable dalle function call."] + "improvements": ["Refactor and clean some code."] }, - "date": "2024-01-02", - "version": "0.118.4" + "date": "2024-11-06", + "version": "1.28.2" }, { "children": { - "fixes": ["Fix parse error of tool calls at end."] + "improvements": ["Update database fields."] }, - "date": "2024-01-01", - "version": "0.118.3" + "date": "2024-11-06", + "version": "1.28.1" }, { "children": { - "fixes": ["Pin antd version to fix chat page crash."] + "features": ["Support export as markdown and JSON."] }, - "date": "2023-12-31", - "version": "0.118.2" + "date": "2024-11-05", + "version": "1.28.0" }, { "children": { - "fixes": ["Fix dalle image download error."] + "improvements": ["Add claude 3.5 haiku model."] }, - "date": "2023-12-30", - "version": "0.118.1" + "date": "2024-11-05", + "version": "1.27.3" }, { "children": { - "features": ["Support markdown type plugin."] + "fixes": ["Remove the 'resetConversation' hot key tip."], + "improvements": ["Improve group topic styles."] }, - "date": "2023-12-29", - "version": "0.118.0" + "date": "2024-11-05", + "version": "1.27.2" }, { "children": { - "fixes": ["The input box is prone to losing focus."] + "fixes": ["Fix /webapi/plugin/store server error."] }, - "date": "2023-12-29", - "version": "0.117.5" + "date": "2024-11-04", + "version": "1.27.1" }, { "children": { - "fixes": ["Fix messages not refresh when creating a new topic."] + "features": ["Support group topics by time."] }, - "date": "2023-12-28", - "version": "0.117.4" + "date": "2024-11-04", + "version": "1.27.0" }, { "children": { - "fixes": ["Fix tool calls at end, fix vision model max tokens, improve vision model checker."] + "fixes": ["If enable login and not signed in, return unauthorized error."] }, - "date": "2023-12-28", - "version": "0.117.3" + "date": "2024-11-04", + "version": "1.26.21" }, { "children": { - "fixes": ["Fix market locale missing."] + "improvements": ["Disable chunking button for unsupported files."] }, - "date": "2023-12-28", - "version": "0.117.2" + "date": "2024-11-04", + "version": "1.26.20" }, { "children": { - "improvements": ["Add image download functionality to DALL·E render component."] + "fixes": [ + "Fix duplicate key value violates unique constraint \"slug_user_id_unique\" when create inbox session." + ] }, - "date": "2023-12-27", - "version": "0.117.1" + "date": "2024-11-03", + "version": "1.26.19" }, { "children": { - "features": ["Support plugin settings env."], - "fixes": ["Improve topic search experience."] + "fixes": ["Fix MS Entra ID and Azure AD authorization."] }, - "date": "2023-12-27", - "version": "0.117.0" + "date": "2024-11-03", + "version": "1.26.18" }, { "children": { - "fixes": ["Fix input box losing focus after sending a message on the desktop."] + "improvements": ["Improve server log on chat api."] }, - "date": "2023-12-27", - "version": "0.116.5" + "date": "2024-10-31", + "version": "1.26.17" }, { "children": { - "fixes": ["Fix ShareModal."] + "fixes": ["Fix server Network connection lost error."] }, - "date": "2023-12-26", - "version": "0.116.4" + "date": "2024-10-31", + "version": "1.26.16" }, { "children": { - "improvements": ["Fix typo."] + "improvements": ["Refactor embedding as plain vector array."] }, - "date": "2023-12-26", - "version": "0.116.3" + "date": "2024-10-31", + "version": "1.26.15" }, { "children": { - "improvements": ["Update Modal style."] + "improvements": ["Add fa-ir locale."] }, - "date": "2023-12-26", - "version": "0.116.2" + "date": "2024-10-30", + "version": "1.26.14" }, { "children": { - "improvements": ["Support slider and select plugin setting render."] + "fixes": ["Fix the artifacts interface not scrolling."] }, - "date": "2023-12-26", - "version": "0.116.1" + "date": "2024-10-30", + "version": "1.26.13" }, { "children": { - "features": ["Support OpenAI tool calls."] + "fixes": ["Fix file image prompts in client mode."] }, - "date": "2023-12-26", - "version": "0.116.0" + "date": "2024-10-30", + "version": "1.26.12" }, { - "children": { - "fixes": ["Fix remove tts and translate not working."] - }, - "date": "2023-12-26", - "version": "0.115.13" + "children": {}, + "date": "2024-10-29", + "version": "1.26.11" }, { "children": { - "fixes": ["Fix active setting tab after click agent setting button."] + "improvements": ["Refactor the aiChat slice actions."] }, - "date": "2023-12-25", - "version": "0.115.12" + "date": "2024-10-29", + "version": "1.26.10" }, { - "children": { - "fixes": ["Fix agent system role modal scrolling when content is too long."] - }, - "date": "2023-12-25", - "version": "0.115.11" + "children": {}, + "date": "2024-10-29", + "version": "1.26.9" }, { "children": { - "improvements": ["Fix some style problem."] + "fixes": ["Update zhipu param process."] }, - "date": "2023-12-25", - "version": "0.115.10" + "date": "2024-10-29", + "version": "1.26.8" }, { "children": { - "fixes": [ - "Fix PLUGINS_INDEX_URL not working, fix a translation error in Traditional Chinese." - ] + "fixes": ["Remove PWA Install in Firefox and Arc."] }, - "date": "2023-12-24", - "version": "0.115.9" + "date": "2024-10-29", + "version": "1.26.7" }, { "children": { - "fixes": ["Fix CUSTOM_MODEL - operator not working."] + "improvements": ["Add Qwen2.5-72B-Instruct model on HF provider."] }, - "date": "2023-12-24", - "version": "0.115.8" + "date": "2024-10-29", + "version": "1.26.6" }, { "children": { - "fixes": ["Fix auto scrolling when generating message."] + "improvements": ["Improve user guide when user not login."] }, - "date": "2023-12-23", - "version": "0.115.7" + "date": "2024-10-29", + "version": "1.26.5" }, { "children": { - "fixes": ["Fix maxTokens params still work when disable enableMaxTokens."] + "fixes": ["Remove the 'resetConversation' hot key."] }, - "date": "2023-12-23", - "version": "0.115.6" + "date": "2024-10-28", + "version": "1.26.4" }, { "children": { - "fixes": ["Fix image display error."] + "fixes": ["Fix Huggingface API interrupting when the output exceeds 140 tokens."], + "improvements": ["Remove SenseChat-Vision model, due to model limitation."] }, - "date": "2023-12-23", - "version": "0.115.5" + "date": "2024-10-28", + "version": "1.26.3" }, { "children": { - "improvements": ["Refactor the ChatMessage type."] + "fixes": ["Fix page not switch when clicking on the pin assistant."] }, - "date": "2023-12-23", - "version": "0.115.4" + "date": "2024-10-28", + "version": "1.26.2" }, { "children": { - "improvements": ["Refactor and clean global store and chat store."] + "improvements": ["Refactor the plugin prompts to xml format."] }, - "date": "2023-12-23", - "version": "0.115.3" + "date": "2024-10-27", + "version": "1.26.1" }, { "children": { - "fixes": ["Fix envs like CUSTOM_MODELS don't work with docker deployment."] + "features": ["experimentally support to pin assistant to sidebar."] }, - "date": "2023-12-23", - "version": "0.115.2" + "date": "2024-10-27", + "version": "1.26.0" }, { "children": { - "improvements": ["Lock ui version to fix setting form style."] + "fixes": ["Fix the issue of the switch assistant portal not closing."] }, - "date": "2023-12-22", - "version": "0.115.1" + "date": "2024-10-27", + "version": "1.25.3" }, { "children": { - "features": ["Support Dall·E 3."] + "improvements": ["Update stepfun models."] }, - "date": "2023-12-22", - "version": "0.115.0" + "date": "2024-10-27", + "version": "1.25.2" }, { "children": { - "improvements": ["Support it-IT nl-NL and pl-PL locales."] + "fixes": ["Fix modelList merge."] }, - "date": "2023-12-22", - "version": "0.114.9" - }, - { - "children": {}, - "date": "2023-12-22", - "version": "0.114.8" + "date": "2024-10-26", + "version": "1.25.1" }, { "children": { - "improvements": ["Move the conversation and chatInput to features folder."] + "features": ["Support ZEN mode."] }, - "date": "2023-12-22", - "version": "0.114.7" - }, - { - "children": {}, - "date": "2023-12-22", - "version": "0.114.6" + "date": "2024-10-26", + "version": "1.25.0" }, { "children": { - "improvements": ["Fix plugin iframe width."] + "improvements": ["Update Google Model list, add gemini-1.5-flash-8b."] }, - "date": "2023-12-19", - "version": "0.114.5" + "date": "2024-10-26", + "version": "1.24.2" }, { "children": { - "fixes": ["Fix agent system role modal scrolling when content is too long."] + "improvements": ["Refactor the Google Gen AI."] }, - "date": "2023-12-19", - "version": "0.114.4" - }, - { - "children": {}, - "date": "2023-12-18", - "version": "0.114.3" + "date": "2024-10-25", + "version": "1.24.1" }, { "children": { - "fixes": ["Fix chat error when message has image with non-vision model."] + "features": ["Add SenseNova (商汤) model provider."] }, - "date": "2023-12-17", - "version": "0.114.2" + "date": "2024-10-25", + "version": "1.24.0" }, { "children": { - "fixes": ["Inject tool description into agent system role."] + "improvements": ["Add *_MODEL_LIST env to all models, update Spark model id & display name."] }, - "date": "2023-12-16", - "version": "0.114.1" + "date": "2024-10-25", + "version": "1.23.1" }, { "children": { - "features": ["Supports setting multiple access code."] + "features": ["Support system agent config."] }, - "date": "2023-12-16", - "version": "0.114.0" + "date": "2024-10-25", + "version": "1.23.0" }, { "children": { - "fixes": ["Fix fontsize setting and audio download style."] + "improvements": ["Add bedrock claude-3.5-sonnect-v2."] }, - "date": "2023-12-16", - "version": "0.113.1" - }, - { - "children": {}, - "date": "2023-12-16", - "version": "0.113.0" + "date": "2024-10-25", + "version": "1.22.27" }, { "children": { - "fixes": ["Fix locales."] + "improvements": ["Fix some custom branding detail."] }, - "date": "2023-12-16", - "version": "0.112.1" + "date": "2024-10-23", + "version": "1.22.26" }, { "children": { - "improvements": [ - "Fix function message style, fix mobile padding of plugin dev modal, improve settings display, Update tool style." - ], - "features": [ - "Introduce plugin detail modal, support OpenAI plugin manifest, support OpenAPI Authentication, support OpenAPI schema in lobe plugin, support parse openapi schema." - ], - "fixes": ["Fix function apiName length, try with node mode plugins."] + "improvements": ["Remove unused user tables."] }, - "date": "2023-12-16", - "version": "0.112.0" + "date": "2024-10-23", + "version": "1.22.25" }, { "children": { - "fixes": ["Fix deployment build failure."] + "improvements": ["Support plugin flag."] }, - "date": "2023-12-15", - "version": "0.111.6" + "date": "2024-10-23", + "version": "1.22.24" }, { "children": { - "fixes": ["Wrong locale language in en_US."] + "improvements": ["Improve error i18n."] }, - "date": "2023-12-14", - "version": "0.111.5" + "date": "2024-10-23", + "version": "1.22.23" }, { "children": { - "fixes": ["Revert \"🐛 fix: clean up gpt-3.5 model\"."] + "improvements": ["Improve i18n."] }, - "date": "2023-12-14", - "version": "0.111.4" + "date": "2024-10-23", + "version": "1.22.22" }, { "children": { - "fixes": ["Fix the history-count text."] + "improvements": ["Refactor cookie/headers to async mode."] }, - "date": "2023-12-14", - "version": "0.111.3" + "date": "2024-10-23", + "version": "1.22.21" }, { "children": { - "fixes": ["Change topic-deletion hotkey."], - "improvements": ["Fix image display in safari (fix."] + "improvements": ["Add new claude-3.5-sonnet model."] }, - "date": "2023-12-13", - "version": "0.111.2" + "date": "2024-10-23", + "version": "1.22.20" }, { "children": { - "fixes": ["Fix locale typo."] + "improvements": ["Move responsive to server utils folder."] }, - "date": "2023-12-13", - "version": "0.111.1" + "date": "2024-10-22", + "version": "1.22.19" }, { "children": {}, - "date": "2023-12-13", - "version": "0.111.0" + "date": "2024-10-22", + "version": "1.22.18" }, { "children": { - "fixes": ["Add cancel button text i18n for delete assistant modal."] + "improvements": ["Fix dynamic import in rsc layout."] }, - "date": "2023-12-13", - "version": "0.110.10" + "date": "2024-10-22", + "version": "1.22.17" }, { "children": { - "fixes": ["ChatInput should have maxHeight."] + "fixes": ["Fix azure-ad."] }, - "date": "2023-12-13", - "version": "0.110.9" + "date": "2024-10-21", + "version": "1.22.16" }, { "children": { - "fixes": ["Clean up gpt-3.5 model."] + "improvements": ["Update format utils and shared layout."] }, - "date": "2023-12-12", - "version": "0.110.8" + "date": "2024-10-21", + "version": "1.22.15" }, { "children": { - "fixes": ["Fix language settings may not take effect."] + "improvements": ["Update wenxin 4.0 turbo model to latest."] }, - "date": "2023-12-11", - "version": "0.110.7" + "date": "2024-10-20", + "version": "1.22.14" }, { "children": { - "fixes": ["Sharp missing in docker production."] + "improvements": [ + "Add Ministral model, update Together AI model list, add function call & vision." + ] }, - "date": "2023-12-11", - "version": "0.110.6" + "date": "2024-10-20", + "version": "1.22.13" }, { "children": { - "fixes": ["Fix setting plugin i18n."] + "improvements": ["Add Llama 3.1 Nemotron 70B model & reorder some provider model list."] }, - "date": "2023-12-10", - "version": "0.110.5" - }, - { - "children": {}, - "date": "2023-12-08", - "version": "0.110.4" + "date": "2024-10-20", + "version": "1.22.12" }, { "children": { - "improvements": ["Refactor with new plugin implement with dexie db."] + "improvements": ["Refactor azure ad to ms entra id."] }, - "date": "2023-12-08", - "version": "0.110.3" + "date": "2024-10-20", + "version": "1.22.11" }, { - "children": { - "improvements": ["Fix ChatInputArea style and typo (resolve."] - }, - "date": "2023-12-08", - "version": "0.110.2" + "children": {}, + "date": "2024-10-20", + "version": "1.22.10" }, { "children": { - "fixes": ["Sharp missing in production."] + "improvements": ["Update Fireworks AI model list."] }, - "date": "2023-12-08", - "version": "0.110.1" + "date": "2024-10-18", + "version": "1.22.9" }, { "children": { - "features": ["Local TTS Player."] + "improvements": ["Add Yi-Lightning model."] }, - "date": "2023-12-07", - "version": "0.110.0" + "date": "2024-10-17", + "version": "1.22.8" }, { "children": { - "fixes": ["Fix agent settings crash with old pluginManifest."] + "improvements": ["Add qwen vision model & update qwen2.5 72b to 128k for siliconcloud."] }, - "date": "2023-12-07", - "version": "0.109.1" + "date": "2024-10-17", + "version": "1.22.7" }, { "children": { - "features": ["Introducing plugin store and refactor with tool concept."] + "fixes": ["Fix images not go in to chat context."] }, - "date": "2023-12-06", - "version": "0.109.0" + "date": "2024-10-13", + "version": "1.22.6" }, { "children": { - "features": ["Hide the password form item in the settings when there is no ACCESS_CODE env."] + "improvements": ["Reorder github model list & updata info & add new model."] }, - "date": "2023-12-03", - "version": "0.108.0" + "date": "2024-10-13", + "version": "1.22.5" }, { "children": { - "fixes": ["Fix custom agent meta issue."] + "improvements": ["Separate message slice and aiChat slice."] }, - "date": "2023-12-03", - "version": "0.107.16" + "date": "2024-10-13", + "version": "1.22.4" }, { "children": { - "fixes": ["Fix messages flickering when creating topic."] + "improvements": ["Support multi-windows for PWA."] }, - "date": "2023-12-03", - "version": "0.107.15" + "date": "2024-10-13", + "version": "1.22.3" }, { "children": { - "fixes": ["Fix opt+delete fail in inputing (resolve."] + "fixes": ["Allow use email as name in logto."] }, - "date": "2023-12-03", - "version": "0.107.14" + "date": "2024-10-13", + "version": "1.22.2" }, { "children": { - "improvements": [ - "Change image fit to cover mode, Fix empty files style, Move file inside chat input in mobile mode, Update editable image style, Update image default background color, Update image editable style, Update image grid, Update Image grid, Update image remove button hover style." - ], - "fixes": [ - "Fix a bug that can't send only images with empty content, Fix image gallery sort index, Fix image gallery sort index, Fix image sort index, Fix image upload error, Fix import." - ] + "fixes": ["Fix function calling issue, disable stream when using tools."] }, - "date": "2023-12-03", - "version": "0.107.13" + "date": "2024-10-12", + "version": "1.22.1" }, { "children": { - "fixes": ["Fix topic not refresh when switching sessions quickly."] + "improvements": ["Refactor the chat webapi."], + "features": ["Add HuggingFace Model Provider."] }, - "date": "2023-12-02", - "version": "0.107.12" + "date": "2024-10-12", + "version": "1.22.0" }, { - "children": { - "fixes": ["Fix switch model don't work on mobile."] - }, - "date": "2023-12-01", - "version": "0.107.11" + "children": {}, + "date": "2024-10-12", + "version": "1.21.16" }, { "children": {}, - "date": "2023-11-30", - "version": "0.107.10" + "date": "2024-10-12", + "version": "1.21.15" }, { "children": { - "fixes": ["Switch session causing problem."] + "improvements": ["Fix artifacts render markdown."] }, - "date": "2023-11-30", - "version": "0.107.9" + "date": "2024-10-12", + "version": "1.21.14" }, { "children": { - "improvements": ["Fix chatitem gap."] + "improvements": ["Refactor agent runtime implement of stream and ZHIPU provider."] }, - "date": "2023-11-30", - "version": "0.107.8" + "date": "2024-10-11", + "version": "1.21.13" }, { "children": { - "fixes": ["Improve plugin message display."] + "improvements": ["Refactor the jwt code."] }, - "date": "2023-11-30", - "version": "0.107.7" + "date": "2024-10-11", + "version": "1.21.12" }, { "children": { - "fixes": ["修正调用插件查询的显示问题."] + "improvements": ["Refactor the backend code for better organization."] }, - "date": "2023-11-30", - "version": "0.107.6" + "date": "2024-10-11", + "version": "1.21.11" }, { "children": { - "fixes": ["修正调用插件查询的显示问题."] + "improvements": ["Updata gpt-4o model info."] }, - "date": "2023-11-30", - "version": "0.107.5" + "date": "2024-10-11", + "version": "1.21.10" }, { "children": { - "fixes": ["Fix a bug that remove all topics when clear message."] + "improvements": ["Update qwen vl model to latest."] }, - "date": "2023-11-30", - "version": "0.107.4" + "date": "2024-10-10", + "version": "1.21.9" }, { "children": { - "fixes": ["Fix a bug that trigger plugin's message type error."] + "fixes": ["Fix auto rewrite query when user message is too long."], + "improvements": ["Support yml in file chunk."] }, - "date": "2023-11-30", - "version": "0.107.3" + "date": "2024-10-08", + "version": "1.21.8" }, { "children": { - "fixes": ["Fix a bug that export a session without messages."] + "improvements": ["Refactor text-to-image endpoint."] }, - "date": "2023-11-30", - "version": "0.107.2" + "date": "2024-10-08", + "version": "1.21.7" }, { "children": { - "improvements": ["优化文案."] + "improvements": ["Move backend api to (backend) folder group."], + "fixes": ["Fix txt-to-image api."] }, - "date": "2023-11-30", - "version": "0.107.1" + "date": "2024-10-05", + "version": "1.21.6" }, { "children": { - "features": ["Refactor the persist layer from zustand's persist to dexie ORM."] + "improvements": ["Support shadcn in Artifacts."] }, - "date": "2023-11-30", - "version": "0.107.0" + "date": "2024-10-05", + "version": "1.21.5" }, { "children": { - "features": ["Support custom deletion, addition, and renaming of models."] + "fixes": ["Fix recharts deps in the Artifacts React Renderer."] }, - "date": "2023-11-29", - "version": "0.106.0" + "date": "2024-10-02", + "version": "1.21.4" }, { "children": { - "fixes": ["Add some env to Dockerfile."] + "improvements": ["Move most /api to /webapi."] }, - "date": "2023-11-27", - "version": "0.105.2" + "date": "2024-10-01", + "version": "1.21.3" }, { "children": { - "fixes": ["Fix agent market detail scroll error."] + "improvements": ["Adjust Wenxin icon size."] }, - "date": "2023-11-27", - "version": "0.105.1" + "date": "2024-10-01", + "version": "1.21.2" }, { - "children": { - "features": ["Standalone pluginn can get more arguments on init."] - }, - "date": "2023-11-22", - "version": "0.105.0" + "children": {}, + "date": "2024-09-30", + "version": "1.21.1" }, { "children": { - "features": ["Support using env variable to set regions for OpenAI Edge Functions.."] + "features": ["Add wenxin model provider."] }, - "date": "2023-11-21", - "version": "0.104.0" + "date": "2024-09-30", + "version": "1.21.0" }, { - "children": { - "fixes": ["Image optimization in docker standalone build."] - }, - "date": "2023-11-21", - "version": "0.103.1" + "children": {}, + "date": "2024-09-30", + "version": "1.20.8" }, { "children": { - "features": ["Support the auto create topic configuration."] + "improvements": ["Update groq model list."] }, - "date": "2023-11-20", - "version": "0.103.0" + "date": "2024-09-29", + "version": "1.20.7" }, { "children": {}, - "date": "2023-11-20", - "version": "0.102.4" + "date": "2024-09-29", + "version": "1.20.6" }, { - "children": { - "fixes": ["Fix plugin not work correct when adding agent from market."] - }, - "date": "2023-11-20", - "version": "0.102.3" + "children": {}, + "date": "2024-09-29", + "version": "1.20.5" }, { - "children": { - "fixes": ["Fix model tag missing."] - }, - "date": "2023-11-20", - "version": "0.102.2" + "children": {}, + "date": "2024-09-28", + "version": "1.20.4" }, { "children": { - "fixes": ["Fix image upload list missing."] + "fixes": ["Improve delete orphan chunks when delete files."] }, - "date": "2023-11-19", - "version": "0.102.1" + "date": "2024-09-28", + "version": "1.20.3" }, { "children": { - "features": ["Support TTS & STT."] + "improvements": ["Add zhipu glm-4-flashx model."] }, - "date": "2023-11-19", - "version": "0.102.0" + "date": "2024-09-27", + "version": "1.20.2" }, { - "children": { - "fixes": ["Agent details sidebar and market page height overflow."] - }, - "date": "2023-11-18", - "version": "0.101.7" + "children": {}, + "date": "2024-09-27", + "version": "1.20.1" }, { "children": { - "improvements": ["Add config to renderErrorMessages, Use new Alert ui."] + "features": ["Add Hunyuan(Tencent) model provider."] }, - "date": "2023-11-17", - "version": "0.101.6" + "date": "2024-09-27", + "version": "1.20.0" }, { "children": { - "fixes": ["Improve openai error info."] + "improvements": ["Add llama3.2 model for openrouter provider."] }, - "date": "2023-11-17", - "version": "0.101.5" + "date": "2024-09-27", + "version": "1.19.36" }, { "children": { - "fixes": ["Fix the plugin array merge error when fork agent from market."] + "improvements": ["Add o1-preview and o1-mini model to github model provider."] }, - "date": "2023-11-14", - "version": "0.101.4" + "date": "2024-09-27", + "version": "1.19.35" }, { - "children": { - "improvements": ["Improve password ui to make it more clear."] - }, - "date": "2023-11-14", - "version": "0.101.3" + "children": {}, + "date": "2024-09-26", + "version": "1.19.34" }, { "children": { - "improvements": ["upload image to vision model adapting to mobile device."] + "fixes": ["MiniMax output long content interrupted by non-existent error."], + "improvements": ["Update google provider model info."] }, - "date": "2023-11-14", - "version": "0.101.2" + "date": "2024-09-25", + "version": "1.19.33" }, { "children": { - "fixes": ["Fix market search (fix."] + "improvements": ["Add function call for taichu_llm."] }, - "date": "2023-11-14", - "version": "0.101.1" + "date": "2024-09-25", + "version": "1.19.32" }, { "children": { - "features": ["Support upload images to chat with gpt4-vision model."] + "improvements": ["Add google gemini 1.5 002 series."] }, - "date": "2023-11-14", - "version": "0.101.0" + "date": "2024-09-24", + "version": "1.19.31" }, { "children": { - "improvements": ["Refactor the input area to suit the files upload feature."] + "improvements": ["Disable taichu2.0 functioncall & default disable taichu2.0v model."] }, - "date": "2023-11-11", - "version": "0.100.5" + "date": "2024-09-24", + "version": "1.19.30" }, { "children": { - "fixes": ["Hotkey disabled in form tags."] + "improvements": ["Update taichu provider info & add taichu vision model."] }, - "date": "2023-11-11", - "version": "0.100.4" + "date": "2024-09-24", + "version": "1.19.29" }, { "children": { - "fixes": ["Fix market error."] + "improvements": ["Add function call support for Stepfun."] }, - "date": "2023-11-09", - "version": "0.100.3" + "date": "2024-09-24", + "version": "1.19.28" }, { "children": { - "fixes": ["Upgrade viewport for nextjs 14."] + "improvements": ["Improve images display in chat messages."] }, - "date": "2023-11-09", - "version": "0.100.2" - }, - { - "children": {}, - "date": "2023-11-09", - "version": "0.100.1" + "date": "2024-09-24", + "version": "1.19.27" }, { "children": { - "features": ["Platform check utils."] + "fixes": ["Fix url config import after user state init."], + "improvements": ["Add support function call for 360AI, left sidebar has only assistants."] }, - "date": "2023-11-09", - "version": "0.100.0" + "date": "2024-09-24", + "version": "1.19.26" }, { "children": { - "improvements": ["Add max height to model menu in chat input area."] + "fixes": ["Add missing translations."] }, - "date": "2023-11-08", - "version": "0.99.1" + "date": "2024-09-24", + "version": "1.19.25" }, { "children": { - "features": ["Add Environment Variable for custom model name when deploying."] + "fixes": ["Fix artifacts code language highlight."] }, - "date": "2023-11-08", - "version": "0.99.0" + "date": "2024-09-23", + "version": "1.19.24" }, { "children": { - "fixes": ["Fix redirect to welcome problem when there are topics in inbox."] + "improvements": ["Add spark max-32k model."] }, - "date": "2023-11-07", - "version": "0.98.3" + "date": "2024-09-23", + "version": "1.19.23" }, { "children": { - "improvements": ["Refactor antd locale file to useSWR."] + "fixes": ["Fix ollama model download panel."] }, - "date": "2023-11-07", - "version": "0.98.2" + "date": "2024-09-22", + "version": "1.19.22" }, { "children": { - "improvements": ["Update welcome assetes."] + "improvements": ["Refactor to improve branding customization."] }, - "date": "2023-11-07", - "version": "0.98.1" + "date": "2024-09-21", + "version": "1.19.21" }, { "children": { - "features": ["Support latest openai model."] + "fixes": ["Fix Content-Security-Policy."] }, - "date": "2023-11-07", - "version": "0.98.0" + "date": "2024-09-21", + "version": "1.19.20" }, { "children": { - "fixes": ["Use pnpm to fix docker release."] + "fixes": ["Casdoor webhooks providerAccountId not found."] }, - "date": "2023-11-06", - "version": "0.97.1" + "date": "2024-09-21", + "version": "1.19.19" }, { - "children": { - "features": ["Add open new topic when open a topic."], - "fixes": ["Fix toggle back to default topic when clearing topic."] - }, - "date": "2023-11-05", - "version": "0.97.0" + "children": {}, + "date": "2024-09-21", + "version": "1.19.18" }, { "children": { - "improvements": ["Update topic list header."] + "fixes": ["providerAccountId not exist in provider."] }, - "date": "2023-11-04", - "version": "0.96.9" + "date": "2024-09-21", + "version": "1.19.17" }, { "children": { - "fixes": ["Fix a bug that weather plugin is not work correctly, template remove sharp deps."] + "improvements": ["Improve i18n for discover and improve version check."] }, - "date": "2023-10-31", - "version": "0.96.8" + "date": "2024-09-21", + "version": "1.19.16" }, { "children": { - "fixes": ["Fix a bug when click inbox not switch back to chat page."] + "improvements": ["Improve i18n in discover."] }, - "date": "2023-10-31", - "version": "0.96.7" + "date": "2024-09-20", + "version": "1.19.15" }, { "children": { - "fixes": ["Improve plausible analytics ENV."] + "fixes": ["Pin shiki@1.17.7 to fix code highlight."] }, - "date": "2023-10-30", - "version": "0.96.6" + "date": "2024-09-20", + "version": "1.19.14" }, { "children": { - "fixes": ["Fix docker image optimization error log."] + "fixes": ["Try to implement better ssrf-protect."] }, - "date": "2023-10-29", - "version": "0.96.5" + "date": "2024-09-20", + "version": "1.19.13" }, { "children": { - "fixes": ["Fix agents market locale fallback to english."] + "improvements": ["Support webhooks for casdoor."] }, - "date": "2023-10-29", - "version": "0.96.4" + "date": "2024-09-20", + "version": "1.19.12" }, { "children": { - "improvements": ["Fix SessionList on mobile."] + "fixes": ["Custom model initialization not taking effect error."] }, - "date": "2023-10-28", - "version": "0.96.3" + "date": "2024-09-20", + "version": "1.19.11" }, { "children": { - "improvements": ["Fix some styles and make updates to various files."] + "improvements": ["Add qwen2.5 math and coder model for siliconcloud provider."] }, - "date": "2023-10-28", - "version": "0.96.2" + "date": "2024-09-20", + "version": "1.19.10" }, { "children": { - "improvements": ["Add guide to market page."] + "fixes": ["Fix a bug with server agent config when user not exist."] }, - "date": "2023-10-28", - "version": "0.96.1" + "date": "2024-09-20", + "version": "1.19.9" }, { "children": { - "features": ["Improve pin mode about session group."] + "improvements": ["Delete siliconflow outdated model & disable stepfun functioncall."] }, - "date": "2023-10-27", - "version": "0.96.0" + "date": "2024-09-19", + "version": "1.19.8" }, { "children": { - "improvements": ["Improve plugin message ui."] + "improvements": ["Add siliconflow qwen2.5 model."] }, - "date": "2023-10-25", - "version": "0.95.1" + "date": "2024-09-19", + "version": "1.19.7" }, { "children": { - "improvements": ["优化 plugin 文件夹命名以支持 standalone 类型的插件."], - "features": [ - "Support function call at message end, support plugin settings modal, support plugin state and settings." - ] + "improvements": ["Refactor the tts route url."] }, - "date": "2023-10-24", - "version": "0.95.0" + "date": "2024-09-19", + "version": "1.19.6" }, { "children": { - "fixes": ["Fallback agent market index to en when not find correct locale."] + "improvements": ["Enable functioncall for stepfun models, Update qwen models."] }, - "date": "2023-10-22", - "version": "0.94.5" + "date": "2024-09-19", + "version": "1.19.5" }, { "children": { - "fixes": ["Fix break cn chars in stream mode."] + "improvements": ["Refactor the sitemap implement."] }, - "date": "2023-10-21", - "version": "0.94.4" + "date": "2024-09-19", + "version": "1.19.4" }, { - "children": { - "fixes": ["Fix agent share format."] - }, - "date": "2023-10-19", - "version": "0.94.3" + "children": {}, + "date": "2024-09-19", + "version": "1.19.3" }, { "children": { - "fixes": ["Fix agent market with other locales."] + "improvements": ["Updata qwen model info & add qwen2.5 & reorder provider list."] }, - "date": "2023-10-19", - "version": "0.94.2" + "date": "2024-09-19", + "version": "1.19.2" }, { "children": { - "improvements": ["Update ShareAgentButton."] + "improvements": ["Add mistral provider new models."] }, - "date": "2023-10-19", - "version": "0.94.1" + "date": "2024-09-19", + "version": "1.19.1" }, { "children": { - "features": ["Add agent share."] + "features": [ + "Add Ai21Labs model provider, add Github Models provider, support native Artifacts just like Claude." + ] }, - "date": "2023-10-18", - "version": "0.94.0" + "date": "2024-09-18", + "version": "1.19.0" }, { "children": { - "improvements": ["Refactor chain."], - "features": ["Support multi-language translate."] + "fixes": ["Fix InterceptingRoutes in discover."] }, - "date": "2023-10-18", - "version": "0.93.0" + "date": "2024-09-18", + "version": "1.18.2" }, { - "children": { - "features": ["Support translate message to current language."] - }, - "date": "2023-10-18", - "version": "0.92.0" + "children": {}, + "date": "2024-09-18", + "version": "1.18.1" }, { "children": { - "features": ["Add hotkeys."] + "features": ["Add Discover Page."] }, - "date": "2023-10-17", - "version": "0.91.0" + "date": "2024-09-18", + "version": "1.18.0" }, { "children": { - "fixes": ["Fix ActionBar props and regenerate btn with error message."] + "fixes": ["Fix a corner case of tools_call with empty object."], + "improvements": ["Delete duplicate models in ollama."] }, - "date": "2023-10-17", - "version": "0.90.3" + "date": "2024-09-16", + "version": "1.17.7" }, { "children": { - "improvements": ["Refactor OpenAIStreamPayload with chat name."] + "improvements": ["Rename artifacts to plugins in portal."] }, - "date": "2023-10-17", - "version": "0.90.2" + "date": "2024-09-15", + "version": "1.17.6" }, { "children": { - "improvements": ["Fix lazyload height."] + "improvements": ["Add MiniCPM-V 8B model entries to Ollama model providers."] }, - "date": "2023-10-17", - "version": "0.90.1" + "date": "2024-09-15", + "version": "1.17.5" }, { "children": { - "features": ["Add Lazyload."] + "improvements": ["Update fullscreen loading style."] }, - "date": "2023-10-17", - "version": "0.90.0" + "date": "2024-09-15", + "version": "1.17.4" }, { "children": { - "improvements": ["Refactor ChatList onActionsClick."] + "improvements": ["Delete \"-\" in deepseek displayname."] }, - "date": "2023-10-17", - "version": "0.89.10" + "date": "2024-09-14", + "version": "1.17.3" }, { "children": { - "fixes": ["Fix ChatList FC Render."] + "fixes": ["Fix o1 model list."], + "improvements": ["Update openrouter model list."] }, - "date": "2023-10-17", - "version": "0.89.9" + "date": "2024-09-13", + "version": "1.17.2" }, { "children": { - "improvements": ["Refactor ChatList."], - "fixes": ["Fix type."] + "improvements": ["Update zhipu model info."] }, - "date": "2023-10-16", - "version": "0.89.8" + "date": "2024-09-13", + "version": "1.17.1" }, { "children": { - "fixes": ["Fix setting tab highlight (fix."] + "features": ["Support openai new OpenAI o1-preview/o1-mini models."], + "improvements": ["Support Google Model List."] }, - "date": "2023-10-16", - "version": "0.89.7" + "date": "2024-09-13", + "version": "1.17.0" }, { "children": {}, - "date": "2023-10-15", - "version": "0.89.6" - }, - { - "children": { - "fixes": [ - "Fix fallback to en when the locale is zh, fix reset button not clear plugin settings." - ] - }, - "date": "2023-10-15", - "version": "0.89.5" + "date": "2024-09-13", + "version": "1.16.14" }, { "children": { - "fixes": ["Fix qwen, chatglm request failed."] + "improvements": ["Update siliconcloud model."] }, - "date": "2023-10-15", - "version": "0.89.4" + "date": "2024-09-13", + "version": "1.16.13" }, { "children": { - "fixes": ["Fix plugin error with nginx reverse proxy."] + "improvements": ["Remove brackets from model names with dates in OpenAI."] }, - "date": "2023-10-12", - "version": "0.89.3" + "date": "2024-09-12", + "version": "1.16.12" }, { "children": { - "improvements": ["Modify onClick event in SessionHeader, change title in Loading component,."] + "fixes": ["Support webhooks for logto."], + "improvements": ["Default disable mistral provider useless models."] }, - "date": "2023-10-12", - "version": "0.89.2" + "date": "2024-09-12", + "version": "1.16.11" }, { "children": { - "fixes": ["Remove useless dynamic import."] + "improvements": ["Support Environment Variable Inference For NextAuth."], + "fixes": ["Qwen model param error."] }, - "date": "2023-10-12", - "version": "0.89.1" - }, - { - "children": {}, - "date": "2023-10-12", - "version": "0.89.0" + "date": "2024-09-12", + "version": "1.16.10" }, { "children": { - "features": [ - "Add mobile responsiveness, create new component, modify properties, make API calls, Dynamically import components using \"dynamic\" function." - ] + "improvements": ["Add model and provider desc and url."] }, - "date": "2023-10-11", - "version": "0.88.0" + "date": "2024-09-12", + "version": "1.16.9" }, { "children": { - "features": ["Support custom model name."] + "improvements": ["Improve models and add more info for providers and models."] }, - "date": "2023-10-11", - "version": "0.87.0" + "date": "2024-09-12", + "version": "1.16.8" }, { "children": { - "fixes": ["Fix clear session error."] + "improvements": ["Optimize model token display method."] }, - "date": "2023-10-11", - "version": "0.86.5" + "date": "2024-09-11", + "version": "1.16.7" }, { "children": { - "improvements": ["Improve api key form."] + "fixes": ["Pin next@14.2.8 to fix Internal error."] }, - "date": "2023-10-11", - "version": "0.86.4" + "date": "2024-09-11", + "version": "1.16.6" }, { - "children": { - "fixes": ["Fix docker image."] - }, - "date": "2023-10-11", - "version": "0.86.3" + "children": {}, + "date": "2024-09-11", + "version": "1.16.5" }, { "children": {}, - "date": "2023-10-11", - "version": "0.86.2" + "date": "2024-09-11", + "version": "1.16.4" }, { "children": { - "fixes": ["Fix docker reverse proxy don't work."] + "fixes": ["Add LLM_VISION_IMAGE_USE_BASE64 to support local s3 in vision model."] }, - "date": "2023-10-11", - "version": "0.86.1" + "date": "2024-09-11", + "version": "1.16.3" }, { "children": { - "features": ["Support docker deploy."] + "improvements": ["Update Upstage model list."] }, - "date": "2023-10-10", - "version": "0.86.0" + "date": "2024-09-11", + "version": "1.16.2" }, { "children": { "improvements": [ - "Add new components, modify display properties, and update settings feature, Replace 100vh with 100% to fix mobile scroll problem." + "Reorder the provider list, update spark check model to spark-lite & default disable useless model." ] }, - "date": "2023-10-10", - "version": "0.85.3" + "date": "2024-09-10", + "version": "1.16.1" }, { "children": { - "fixes": ["Add apikey form when there is no default api key in env."] + "features": ["Add Fireworks AI Model Provider, Add Spark model provider."] }, - "date": "2023-10-10", - "version": "0.85.2" + "date": "2024-09-10", + "version": "1.16.0" }, { "children": { - "fixes": ["Fix mobile safe area (fix."] + "improvements": ["Update CustomLogo."] }, - "date": "2023-10-10", - "version": "0.85.1" + "date": "2024-09-10", + "version": "1.15.35" }, { "children": { - "features": ["Add ja_JP, ko_KR and update workflow."] + "improvements": ["Change empty content stream behavior."] }, - "date": "2023-10-10", - "version": "0.85.0" + "date": "2024-09-10", + "version": "1.15.34" }, { "children": { - "features": ["Support detect new version and upgrade action."] + "fixes": ["Fix /etc/resolv.confedit permission in docker image."] }, - "date": "2023-10-10", - "version": "0.84.0" + "date": "2024-09-10", + "version": "1.15.33" }, { "children": { - "fixes": ["Fix rsc layout."] + "fixes": ["Fix tools calling in some edge cases."] }, - "date": "2023-10-09", - "version": "0.83.10" + "date": "2024-09-10", + "version": "1.15.32" }, { - "children": {}, - "date": "2023-10-08", - "version": "0.83.9" + "children": { + "fixes": ["Baichuan should not introduce freequency_penality parameters."] + }, + "date": "2024-09-10", + "version": "1.15.31" }, { - "children": {}, - "date": "2023-10-07", - "version": "0.83.8" + "children": { + "fixes": ["Fix claude 3.5 image with s3 url."] + }, + "date": "2024-09-09", + "version": "1.15.30" }, { "children": { - "fixes": ["Fix shuffle, use search url with agent item."], - "improvements": ["Better tag style, improve loading state."] + "fixes": ["Gemini cannot input images when server database is enabled."] }, - "date": "2023-10-07", - "version": "0.83.7" + "date": "2024-09-09", + "version": "1.15.29" }, { "children": { - "improvements": ["Update modal style."] + "fixes": ["Update baichuan param."] }, - "date": "2023-10-06", - "version": "0.83.6" + "date": "2024-09-09", + "version": "1.15.28" }, { "children": { - "fixes": ["Fix agent market list."] + "improvements": ["Add siliconcloud new model."] }, - "date": "2023-10-06", - "version": "0.83.5" + "date": "2024-09-09", + "version": "1.15.27" }, { "children": { - "fixes": ["Fix agent settings."] + "improvements": ["Update perplexity model list."] }, - "date": "2023-10-06", - "version": "0.83.4" + "date": "2024-09-09", + "version": "1.15.26" + }, + { + "children": {}, + "date": "2024-09-09", + "version": "1.15.25" }, { "children": { - "improvements": ["Refactor the settings layout to rsc."] + "improvements": ["Fix title in about settings."] }, - "date": "2023-10-06", - "version": "0.83.3" + "date": "2024-09-09", + "version": "1.15.24" }, { "children": { - "fixes": ["Fix setCookie method that set cookie with sub-path."] + "improvements": ["Improve branding implement."] }, - "date": "2023-10-06", - "version": "0.83.2" + "date": "2024-09-08", + "version": "1.15.23" }, { "children": { - "improvements": ["Refactor settings page entry."] + "improvements": ["Update model display name & Remove Qwen preview model."] }, - "date": "2023-10-06", - "version": "0.83.1" + "date": "2024-09-08", + "version": "1.15.22" }, { "children": { - "features": ["Upgrade locale with SSR."] + "improvements": ["Temperature range from 0 to 2."] }, - "date": "2023-10-06", - "version": "0.83.0" + "date": "2024-09-08", + "version": "1.15.21" }, { "children": {}, - "date": "2023-10-05", - "version": "0.82.9" + "date": "2024-09-08", + "version": "1.15.20" }, { - "children": { - "improvements": ["Refactor / route to reduce page js size."] - }, - "date": "2023-09-30", - "version": "0.82.8" + "children": {}, + "date": "2024-09-08", + "version": "1.15.19" }, { "children": { - "improvements": ["Refactor the api router to app route handlers."] + "improvements": ["Support anthropic browser request."] }, - "date": "2023-09-30", - "version": "0.82.7" + "date": "2024-09-06", + "version": "1.15.18" }, { "children": { - "fixes": ["Fix share default config, pin openai to fix type error."] + "fixes": ["Fix auth log."] }, - "date": "2023-09-29", - "version": "0.82.6" + "date": "2024-09-06", + "version": "1.15.17" }, { "children": { - "improvements": ["Update theme color and styling of mobile settings page."] + "improvements": ["Update Bedrock model list & add AWS_BEDROCK_MODEL_LIST support."] }, - "date": "2023-09-29", - "version": "0.82.5" + "date": "2024-09-06", + "version": "1.15.16" }, { "children": { - "fixes": ["修正 localStorage 不存在造成设置页刷新 500 保存的问题."] + "improvements": ["Add LLaVA 1.5 7B model in Groq."] }, - "date": "2023-09-29", - "version": "0.82.4" + "date": "2024-09-06", + "version": "1.15.15" }, { - "children": { - "fixes": [ - "修正 access code 校验逻辑,修正 api key 无法正常显示在秘钥输入框,并增加显示关闭按钮,修正移动端输入 access code 默认打开数据键盘的问题." - ] - }, - "date": "2023-09-29", - "version": "0.82.3" + "children": {}, + "date": "2024-09-06", + "version": "1.15.14" }, { - "children": { - "improvements": ["Refactor settings page and mobile ux."] - }, - "date": "2023-09-28", - "version": "0.82.2" + "children": {}, + "date": "2024-09-06", + "version": "1.15.13" }, { "children": { - "fixes": ["Fix share screenshot scrollbar."] + "fixes": ["Fix typo in RAG prompt."] }, - "date": "2023-09-27", - "version": "0.82.1" + "date": "2024-09-04", + "version": "1.15.12" }, { "children": {}, - "date": "2023-09-27", - "version": "0.82.0" + "date": "2024-09-04", + "version": "1.15.11" }, { - "children": { - "features": ["Add several analytics sdk."] - }, - "date": "2023-09-27", - "version": "0.81.0" + "children": {}, + "date": "2024-09-03", + "version": "1.15.10" }, { "children": { - "improvements": ["Switch Modal components to @bentwnghk/ui."] + "fixes": ["Fix speed and rag prompt."] }, - "date": "2023-09-27", - "version": "0.80.2" + "date": "2024-09-03", + "version": "1.15.9" }, { "children": { - "improvements": ["Fix conversation mobile view area."] + "fixes": ["Fix .PDF can not be chunked."] }, - "date": "2023-09-27", - "version": "0.80.1" + "date": "2024-09-03", + "version": "1.15.8" }, { "children": { - "features": ["Improve user experience and ensure consistency."] + "improvements": ["Fix provider disabled title style."] }, - "date": "2023-09-27", - "version": "0.80.0" + "date": "2024-09-03", + "version": "1.15.7" }, { "children": { - "improvements": ["Fix safearea in mobile."] + "improvements": [ + "Stepfun default enabled model, update Groq model list & add GROQ_MODEL_LIST support." + ] }, - "date": "2023-09-27", - "version": "0.79.8" + "date": "2024-09-01", + "version": "1.15.6" }, { "children": { - "improvements": ["Use hook to check PWA env."] + "improvements": ["Update Together AI model list."] }, - "date": "2023-09-27", - "version": "0.79.7" + "date": "2024-09-01", + "version": "1.15.5" }, { "children": { - "improvements": ["Optimize PWA style and scroll effect."] + "improvements": ["Update Novita AI model info & add NOVITA_MODEL_LIST support."] }, - "date": "2023-09-27", - "version": "0.79.6" + "date": "2024-09-01", + "version": "1.15.4" }, { "children": { - "fixes": ["Fix URI error."] + "improvements": [ + "Add *_MODEL_LIST for Qwen and ZeroOne, fix model info, update Claude 3.5 Sonnet maxOutput vaule." + ] }, - "date": "2023-09-26", - "version": "0.79.5" + "date": "2024-09-01", + "version": "1.15.3" }, { "children": { - "improvements": ["Move dir from page to app and remove .page suffix."] + "improvements": ["Update Qwen and Gemini models info."] }, - "date": "2023-09-26", - "version": "0.79.4" - }, - { - "children": {}, - "date": "2023-09-25", - "version": "0.79.3" - }, - { - "children": {}, - "date": "2023-09-25", - "version": "0.79.2" - }, - { - "children": {}, - "date": "2023-09-25", - "version": "0.79.1" - }, - { - "children": {}, - "date": "2023-09-25", - "version": "0.79.0" + "date": "2024-08-30", + "version": "1.15.2" }, { "children": { - "improvements": ["Show topic tooltip on left side."] + "improvements": ["Update the sorting of each provider model."] }, - "date": "2023-09-21", - "version": "0.78.1" + "date": "2024-08-30", + "version": "1.15.1" }, { "children": { - "features": ["Auto create topic when chatting."] + "features": ["Add Upstage model provider support."] }, - "date": "2023-09-17", - "version": "0.78.0" - }, - { - "children": {}, - "date": "2023-09-15", - "version": "0.77.2" + "date": "2024-08-30", + "version": "1.15.0" }, { "children": { - "fixes": ["Fix lint."] + "improvements": ["Fix ms doc file preview, Update the sorting of each provider model."] }, - "date": "2023-09-14", - "version": "0.77.1" + "date": "2024-08-30", + "version": "1.14.12" }, { "children": { - "features": ["Update localization files and add translations for different languages."] + "improvements": ["Update Stepfun models info."] }, - "date": "2023-09-14", - "version": "0.77.0" + "date": "2024-08-30", + "version": "1.14.11" }, { "children": { - "fixes": ["Fix client config."] + "fixes": ["Fix file relative chunks."] }, - "date": "2023-09-11", - "version": "0.76.2" + "date": "2024-08-30", + "version": "1.14.10" }, { - "children": { - "fixes": ["Fix save topic button."] - }, - "date": "2023-09-11", - "version": "0.76.1" + "children": {}, + "date": "2024-08-29", + "version": "1.14.9" }, { "children": { - "features": ["Support Azure OpenAI Deploy env."] + "fixes": ["Fix whisper-1 typo."] }, - "date": "2023-09-11", - "version": "0.76.0" + "date": "2024-08-29", + "version": "1.14.8" }, { "children": { - "improvements": ["Update loading style and compatible with unknown agent identifier."], - "features": [ - "Add agents market and improve UI components, Add and refactor components for chat input feature, Add functions for generating and analyzing JSON files, generating resource files and table of contents, and formatting console output, Add new settings for Azure OpenAI and OpenAI in locales files, Add new string, create AgentModal component, implement GridCardItem and Loading components, import AgentModal, Add SideBar component, new actions, and update market store state and selectors, Add translations and new setting to \"setting.json\", Improve functionality and user interface of market page, Modify market features components and update CSS styles, support add agent to chat." - ], - "fixes": ["Fix fetcher, Fix market sidebar scroll and add i18n."] + "fixes": ["Disable ChatGPT-4o Tools Calling."], + "improvements": ["Improve chunk and file preview."] }, - "date": "2023-09-11", - "version": "0.75.0" + "date": "2024-08-28", + "version": "1.14.7" }, { "children": { - "features": ["Add russian locally, Update Russian and English locally (LLM tab)."] + "improvements": ["Update Gemini models."] }, - "date": "2023-09-11", - "version": "0.74.0" + "date": "2024-08-28", + "version": "1.14.6" }, { "children": { - "features": ["Support Azure OpenAI."] + "fixes": ["No user name if Cloudflare Zero Trust with onetimepin."] }, - "date": "2023-09-10", - "version": "0.73.0" + "date": "2024-08-28", + "version": "1.14.5" }, { "children": { - "fixes": ["Use en-US when no suit lang with plugin index."] + "improvements": ["Move model and provider icon components to @lobehub/icons."] }, - "date": "2023-09-10", - "version": "0.72.4" + "date": "2024-08-28", + "version": "1.14.4" }, { "children": { - "fixes": ["Fix sessionList double click on mobile."] + "fixes": ["Improve aysnc error type."] }, - "date": "2023-09-09", - "version": "0.72.3" + "date": "2024-08-27", + "version": "1.14.3" }, { "children": { - "fixes": ["Fix mobile switch when session selected."] + "fixes": ["Fix agent setting."] }, - "date": "2023-09-09", - "version": "0.72.2" + "date": "2024-08-27", + "version": "1.14.2" }, { "children": { - "fixes": ["修正异步水合造成的初始状态不稳定的问题."] + "improvements": ["Improve zhipu model config."] }, - "date": "2023-09-09", - "version": "0.72.1" + "date": "2024-08-27", + "version": "1.14.1" }, { "children": { - "features": [ - "Add plugin market Setting Modal, 支持快速刷新与预览 manifest, 适配插件 i18n 方案." - ], - "fixes": ["修正删除插件时错误开启的问题."], - "improvements": ["优化 manifest 预览的尺寸."] + "features": ["Supports Cloudflare Zero Trust login."] }, - "date": "2023-09-09", - "version": "0.72.0" + "date": "2024-08-27", + "version": "1.14.0" }, { "children": { - "fixes": ["Fix mobile route."] + "fixes": ["Bypass vercel deployment protection, fix can send message on uploading files."] }, - "date": "2023-09-09", - "version": "0.71.1" + "date": "2024-08-27", + "version": "1.13.2" }, { "children": { - "features": ["Migrate localStorage to indexedDB."] + "improvements": ["Update Qwen models."] }, - "date": "2023-09-09", - "version": "0.71.0" + "date": "2024-08-27", + "version": "1.13.1" }, { "children": { - "fixes": ["Fix route."] + "features": ["Supports Authelia login."] }, - "date": "2023-09-09", - "version": "0.70.4" + "date": "2024-08-27", + "version": "1.13.0" }, { "children": { - "improvements": ["Better mobile style."] + "fixes": ["Feature flag knowledge_base doesn't affect ActionBar."] }, - "date": "2023-09-09", - "version": "0.70.3" + "date": "2024-08-26", + "version": "1.12.20" }, { "children": { - "fixes": ["修正移动端路由问题."] + "fixes": ["Fix cannot clone agent when imported from client."] }, - "date": "2023-09-08", - "version": "0.70.2" + "date": "2024-08-25", + "version": "1.12.19" }, { "children": { - "improvements": ["Refactor settingsSelectors to globalSelectors."] + "fixes": ["Fix dayjs error in en-US language."] }, - "date": "2023-09-08", - "version": "0.70.1" + "date": "2024-08-25", + "version": "1.12.18" }, { "children": { - "features": ["Refactor to url state."] + "fixes": ["Fix multi file upload dupicate."] }, - "date": "2023-09-08", - "version": "0.70.0" + "date": "2024-08-25", + "version": "1.12.17" }, { "children": { - "improvements": ["Migrate openai-edge to openai."] + "fixes": ["Session not found error on mobile."] }, - "date": "2023-09-06", - "version": "0.69.1" + "date": "2024-08-24", + "version": "1.12.16" }, { - "children": { - "features": ["Add new import statement for \"Flexbox\" component in \"Empty\" component."] - }, - "date": "2023-09-06", - "version": "0.69.0" + "children": {}, + "date": "2024-08-24", + "version": "1.12.15" }, { "children": { - "fixes": ["修正数组合并逻辑,修正被移除插件无法看到的问题."] + "fixes": ["Fix tts file saving in server mode."] }, - "date": "2023-09-03", - "version": "0.68.1" + "date": "2024-08-24", + "version": "1.12.14" }, { "children": { - "features": ["Plugin default use iframe render."] + "improvements": ["Update 01.AI models."] }, - "date": "2023-09-03", - "version": "0.68.0" + "date": "2024-08-24", + "version": "1.12.13" }, { "children": {}, - "date": "2023-09-02", - "version": "0.67.0" + "date": "2024-08-24", + "version": "1.12.12" }, { "children": { - "features": ["Add russian locally."] + "fixes": ["Remove orphan chunks if there is no related file."] }, - "date": "2023-09-02", - "version": "0.66.0" + "date": "2024-08-23", + "version": "1.12.11" }, { "children": { - "fixes": ["修正 defaultAgent 无法正常变更的问题."] + "fixes": ["Refactor and fix dalle."] }, - "date": "2023-09-01", - "version": "0.65.1" + "date": "2024-08-23", + "version": "1.12.10" }, { "children": { - "features": ["支持本地插件自定义 gateway."] + "fixes": ["Improve s3 path-style url."] }, - "date": "2023-08-29", - "version": "0.65.0" + "date": "2024-08-23", + "version": "1.12.9" }, { "children": { - "improvements": ["Update i18n."] + "fixes": ["Fix NEXT_PUBLIC_S3_DOMAIN error on Docker."] }, - "date": "2023-08-29", - "version": "0.64.1" + "date": "2024-08-22", + "version": "1.12.8" }, { "children": { - "improvements": ["Remove no need i18n."], - "features": [ - "增加自定义插件的增删改配置功能,完善自定义插件表单的校验逻辑,支持本地插件侧的请求与错误呈现,新增插件配置 Dev 弹窗,绑定本地插件的增删改逻辑." - ] + "fixes": ["Logout button not shown on mobile view when using nextauth."] }, - "date": "2023-08-29", - "version": "0.64.0" + "date": "2024-08-22", + "version": "1.12.7" }, { "children": { - "improvements": ["Refactor with new market url."] + "improvements": ["Refactor s3 env and support path-style for minio."] }, - "date": "2023-08-28", - "version": "0.63.3" + "date": "2024-08-22", + "version": "1.12.6" }, { "children": { - "improvements": ["Refactor AgentSettings."] + "fixes": ["Fix clipboard copy issue and improve upload cors feedback."] }, - "date": "2023-08-27", - "version": "0.63.2" + "date": "2024-08-22", + "version": "1.12.5" }, { "children": { - "improvements": ["Refactor the selectors import."] + "improvements": ["Fix link style."] }, - "date": "2023-08-27", - "version": "0.63.1" + "date": "2024-08-22", + "version": "1.12.4" }, { "children": { - "features": ["support sharing to shareGPT."] + "improvements": ["Hide settings in repo."] }, - "date": "2023-08-27", - "version": "0.63.0" + "date": "2024-08-22", + "version": "1.12.3" }, { - "children": { - "fixes": ["Fix plugin settings error."] - }, - "date": "2023-08-26", - "version": "0.62.1" + "children": {}, + "date": "2024-08-22", + "version": "1.12.2" }, { "children": { - "features": ["支持超过 4k 的会话使用 16k 总结标题."], - "fixes": ["Fix plugin settings error."], - "improvements": ["优化清理会话的操作路径,优化默认角色的配置."] + "fixes": ["Fix embeddings multi-insert when there is issues with async task."] }, - "date": "2023-08-26", - "version": "0.62.0" + "date": "2024-08-21", + "version": "1.12.1" }, { "children": { - "features": ["新增自动滚动."] + "features": ["Files and knowledge base."] }, - "date": "2023-08-26", - "version": "0.61.0" + "date": "2024-08-21", + "version": "1.12.0" }, { "children": { - "improvements": ["优化文案."] + "fixes": ["Fixed bedrock llama model id."] }, - "date": "2023-08-26", - "version": "0.60.4" + "date": "2024-08-19", + "version": "1.11.9" }, { "children": { - "fixes": ["Fix global state merge error."] + "improvements": ["Update zhipu models."] }, - "date": "2023-08-26", - "version": "0.60.3" + "date": "2024-08-19", + "version": "1.11.8" }, { "children": { - "fixes": ["Fix fetch plugin header error."] + "fixes": ["Fix topic scroll issue."] }, - "date": "2023-08-26", - "version": "0.60.2" + "date": "2024-08-18", + "version": "1.11.7" }, { "children": { - "fixes": ["Fix settings storage error."] + "improvements": ["Refactor the SITE_URL to APP_URL."] }, - "date": "2023-08-26", - "version": "0.60.1" + "date": "2024-08-18", + "version": "1.11.6" }, { "children": { - "improvements": ["Refactor with new market index url."], - "features": ["支持插件 manifest 加载失败后重试."] + "improvements": ["Refactor the fetch method to fix response.undefined."] }, - "date": "2023-08-26", - "version": "0.60.0" + "date": "2024-08-18", + "version": "1.11.5" }, { "children": { - "features": ["支持展示插件插件状态,支持插件 i18n 模式展示."] + "improvements": [ + "Add SILICONCLOUD_MODEL_LIST & SILICONCLOUD_PROXY_URL support for SiliconCloud." + ] }, - "date": "2023-08-26", - "version": "0.59.0" + "date": "2024-08-18", + "version": "1.11.4" }, { "children": { - "features": ["Implement responsive design for mobile devices."] + "improvements": ["Refactor PanelTitle and move commit from file uploading."] }, - "date": "2023-08-26", - "version": "0.58.0" + "date": "2024-08-17", + "version": "1.11.3" }, { - "children": { - "improvements": ["Refactor to ChatErrorType."], - "features": [ - "完善插件请求的错误处理,支持修改与记录插件的配置,支持发送插件配置信息,支持渲染 manifest 中的 settings, 支持设置不正确时进行插件的配置,新增插件请求状态的错误处理." - ], - "fixes": ["修正缓存旧数据的报错问题."] - }, - "date": "2023-08-26", - "version": "0.57.0" + "children": {}, + "date": "2024-08-17", + "version": "1.11.2" }, { "children": { - "features": ["Use new plugin manifest to support plugin’s multi api."] + "fixes": ["Make S3 upload ACL setting optional."] }, - "date": "2023-08-24", - "version": "0.56.0" + "date": "2024-08-15", + "version": "1.11.1" }, { "children": { - "improvements": ["Refactor plugin api with @lobehub/chat-plugins-gateway."] + "features": ["Add 2 new models to openai provider."] }, - "date": "2023-08-22", - "version": "0.55.1" + "date": "2024-08-14", + "version": "1.11.0" }, { - "children": { - "improvements": ["完成插件市场 loading 态样式."], - "features": ["初步完成插件市场动态加载全链路,实现插件组件的动态加载."], - "fixes": ["Fix error, 修正无法正常开启插件的问题,修正测试,补充插件 store 的水合逻辑."] - }, - "date": "2023-08-22", - "version": "0.55.0" + "children": {}, + "date": "2024-08-14", + "version": "1.10.1" }, { "children": { - "fixes": ["Fix not cannot change setting error."] + "features": ["Add SiliconCloud model provider."] }, - "date": "2023-08-21", - "version": "0.54.4" + "date": "2024-08-14", + "version": "1.10.0" }, { "children": { - "improvements": ["Refactor plugin request."] + "improvements": ["Resize the image size in chat message."] }, - "date": "2023-08-21", - "version": "0.54.3" + "date": "2024-08-13", + "version": "1.9.8" + }, + { + "children": {}, + "date": "2024-08-13", + "version": "1.9.7" + }, + { + "children": {}, + "date": "2024-08-09", + "version": "1.9.6" }, { "children": { - "improvements": ["修正图片选项的样式问题."] + "improvements": ["Updated AWS bedrock model list."] }, - "date": "2023-08-16", - "version": "0.54.2" + "date": "2024-08-08", + "version": "1.9.5" }, { "children": { - "fixes": ["修正 i18n 失效的问题."] + "fixes": ["Fix import clerk AuthObject from public api."] }, - "date": "2023-08-16", - "version": "0.54.1" + "date": "2024-08-06", + "version": "1.9.4" }, { "children": { - "features": ["Add new features and improve user interface and functionality."] + "improvements": ["Refactor server db schema for better code organize."] }, - "date": "2023-08-15", - "version": "0.54.0" + "date": "2024-08-06", + "version": "1.9.3" }, { "children": {}, - "date": "2023-08-15", - "version": "0.53.0" + "date": "2024-08-05", + "version": "1.9.2" }, { "children": { - "improvements": ["Replace cdn."] + "fixes": ["Azure modelTag icon display."] }, - "date": "2023-08-15", - "version": "0.52.1" + "date": "2024-08-05", + "version": "1.9.1" }, { "children": { - "features": ["Add avatar compress."] + "features": ["Skip login page if only one provider exists."] }, - "date": "2023-08-15", - "version": "0.52.0" + "date": "2024-08-05", + "version": "1.9.0" }, { "children": { - "features": ["Add Footer component and modify Token and index files."] + "fixes": ["Add PROXY_URL in docker with proxychains-ng."] }, - "date": "2023-08-15", - "version": "0.51.0" + "date": "2024-08-03", + "version": "1.8.2" }, { "children": { - "features": [ - "Update messages, settings, error codes, plugin names, weather data display, and UI." - ] + "improvements": ["Fix aya, mathstral model tag icon & update ollama model info."] }, - "date": "2023-08-15", - "version": "0.50.0" + "date": "2024-08-03", + "version": "1.8.1" }, { "children": { - "features": ["Add BackToBottom to conversation, Update icons and text in various components."] + "features": ["Add NextAuth as authentication service in server database."] }, - "date": "2023-08-15", - "version": "0.49.0" + "date": "2024-08-02", + "version": "1.8.0" }, { "children": { - "features": ["Import SiOpenai icon and replace 'Tag' component in chat feature."] + "improvements": ["Add Gemini 1.5 Pro Exp model."] }, - "date": "2023-08-15", - "version": "0.48.0" + "date": "2024-08-02", + "version": "1.7.10" }, { "children": { - "features": ["Add and update UI elements and agent configuration."] + "fixes": ["Fix Mistral models calling & update model info."], + "improvements": ["Fix stepfun & baichuan model tag icon missing, update Perplexity models."] }, - "date": "2023-08-15", - "version": "0.47.0" + "date": "2024-08-01", + "version": "1.7.9" }, { - "children": { - "improvements": ["Fix SystemRole Skeleton padding."] - }, - "date": "2023-08-14", - "version": "0.46.1" + "children": {}, + "date": "2024-07-30", + "version": "1.7.8" }, { "children": { - "features": [ - "Update styling and functionality of AgentPrompt and EditableMessage components, 支持停止生成消息." - ], - "fixes": ["Remove input highlight."] + "improvements": ["Improve tools calling UI."] }, - "date": "2023-08-14", - "version": "0.46.0" + "date": "2024-07-30", + "version": "1.7.7" }, { "children": { - "features": ["优化每个角色的初始引导."], - "improvements": ["优化初始化加载状态,等到会话加载完毕再显示内容."] + "fixes": ["Disable anthropic browser request."] }, - "date": "2023-08-14", - "version": "0.45.0" + "date": "2024-07-29", + "version": "1.7.6" }, { "children": { - "improvements": ["优化 Chat Skeleton 样式,优化 Inbox 样式."] + "fixes": ["Fix create_session edit_agent feature flags and add more flags."], + "improvements": ["Update 360GPT model (360GPT2 Pro)."] }, - "date": "2023-08-13", - "version": "0.44.4" + "date": "2024-07-29", + "version": "1.7.5" }, { "children": { - "improvements": ["修正话题列表无法滚动的问题."], - "fixes": ["修正 inbox 点击重新生成会报错的问题."] + "fixes": ["Improve remote model list fetching for Novita AI."] }, - "date": "2023-08-13", - "version": "0.44.3" + "date": "2024-07-29", + "version": "1.7.4" }, { "children": { - "fixes": ["修正重新生成时切分历史消息的逻辑."] + "fixes": ["Update minimax models."] }, - "date": "2023-08-13", - "version": "0.44.2" + "date": "2024-07-28", + "version": "1.7.3" }, { - "children": {}, - "date": "2023-08-12", - "version": "0.44.1" + "children": { + "fixes": ["Avoid baseURL being an empty string, resulting in incorrect client fetch."] + }, + "date": "2024-07-26", + "version": "1.7.2" }, { "children": { - "improvements": [ - "Fix Inbox defaultMessage avatar, 优化 header 的 setting 展示,优化门禁下默认的解锁方式,补充 ChatList 的 Loading 态." - ], - "features": [ - "支持 inbox 消息导出,支持 inbox 的会话功能,新增 inbox 数据模型,新增 inbox 模块入口." - ] + "fixes": ["Fix dalle tools calling prompts to avoid content risk."] }, - "date": "2023-08-12", - "version": "0.44.0" + "date": "2024-07-26", + "version": "1.7.1" }, { "children": { - "features": ["支持切换语言."] + "features": ["Enabled function calling on Deepseek models."] }, - "date": "2023-08-12", - "version": "0.43.0" + "date": "2024-07-26", + "version": "1.7.0" }, { "children": { - "improvements": ["暂时隐藏 Hero 模板."] + "improvements": ["Fix file upload height."] }, - "date": "2023-08-12", - "version": "0.42.3" + "date": "2024-07-26", + "version": "1.6.15" }, { "children": { - "improvements": [ - "将 useSettings 更名为 useGlobalStore, 将原本的 settings 更名为 global, 收敛切换 SideBar 方法为 useSwitchSideBarOnInit, 重构需本地缓存的状态为 preference." - ], - "fixes": ["修正移除 session 时的路由跳转逻辑."] + "improvements": ["Improve input file upload."] }, - "date": "2023-08-12", - "version": "0.42.2" + "date": "2024-07-26", + "version": "1.6.14" }, { "children": { - "improvements": ["优化 App 首页 Loading 态."] + "improvements": ["Updated Groq model list to include llama-3.1 and llama3-Groq."] }, - "date": "2023-08-12", - "version": "0.42.1" + "date": "2024-07-25", + "version": "1.6.13" }, { "children": { - "features": ["Add Welcome page."] + "improvements": ["Add new models to groq which are llama 3.1."] }, - "date": "2023-08-11", - "version": "0.42.0" + "date": "2024-07-25", + "version": "1.6.12" }, { "children": { - "improvements": [ - "将 sessionStore 默认 equalFn 改为 shallow, 将 settingStore 默认 equalFn 改为 shallow." - ] + "fixes": ["Fix UNAUTHORIZED issue with clerk auth provider."] }, - "date": "2023-08-10", - "version": "0.41.2" + "date": "2024-07-24", + "version": "1.6.11" }, { "children": { - "improvements": ["重构 settings store 代码写法."] + "improvements": ["Fix the scrolling of the return result area of function calling."] }, - "date": "2023-08-10", - "version": "0.41.1" + "date": "2024-07-23", + "version": "1.6.10" + }, + { + "children": {}, + "date": "2024-07-23", + "version": "1.6.9" }, { "children": { - "features": ["支持持久化隐藏 Topic 功能."], - "improvements": ["优化第一次水合逻辑."] + "improvements": ["Move server modules."] }, - "date": "2023-08-10", - "version": "0.41.0" + "date": "2024-07-23", + "version": "1.6.8" }, { "children": { - "improvements": ["优化 Topic 的水合加载效果."] + "improvements": ["Add new model provider Novita AI."] }, - "date": "2023-08-10", - "version": "0.40.7" + "date": "2024-07-23", + "version": "1.6.7" + }, + { + "children": {}, + "date": "2024-07-22", + "version": "1.6.6" }, { "children": { - "improvements": ["优化水合前的加载效果."] + "fixes": ["Content lost unexpectedly on Qwen provider when finish_reason is stop."] }, - "date": "2023-08-10", - "version": "0.40.6" + "date": "2024-07-22", + "version": "1.6.5" }, { "children": { - "improvements": ["增加未初始化的 loading 态."] + "improvements": ["Add trpc query client with react-query."] }, - "date": "2023-08-10", - "version": "0.40.5" + "date": "2024-07-21", + "version": "1.6.4" }, { "children": { - "improvements": ["优化 Header 样式."] + "improvements": ["Update Zhipu models (GLM-4-AllTools & CodeGeeX-4)."] }, - "date": "2023-08-10", - "version": "0.40.4" + "date": "2024-07-21", + "version": "1.6.3" }, { "children": { - "fixes": ["修正没有 prompt 的编辑与保存按钮的问题."] + "fixes": ["Fix dayjs render on server."] }, - "date": "2023-08-10", - "version": "0.40.3" + "date": "2024-07-21", + "version": "1.6.2" }, { "children": { - "fixes": ["修正 defaults 造成的 config 报错."] + "improvements": ["Refactor the DragUpload."] }, - "date": "2023-08-08", - "version": "0.40.2" + "date": "2024-07-19", + "version": "1.6.1" }, { "children": { - "fixes": ["优化 openai 接口的错误处理逻辑."] + "features": ["Add gpt-4o-mini in OpenAI Provider and set it as the default model."] }, - "date": "2023-08-06", - "version": "0.40.1" + "date": "2024-07-19", + "version": "1.6.0" }, { "children": { - "features": ["Add new dependency, add Tag and PluginTag components, update HeaderTitle."] + "improvements": ["Added Gemma2 instead of outdated Gemma."] }, - "date": "2023-08-05", - "version": "0.40.0" + "date": "2024-07-19", + "version": "1.5.5" }, { "children": { - "improvements": ["修正 assistant 消息没有 background 的问题."] + "fixes": ["Fix delete session group."] }, - "date": "2023-08-05", - "version": "0.39.4" + "date": "2024-07-17", + "version": "1.5.4" }, { "children": { - "fixes": ["优化 405 报错返回内容,并优化 openai 服务端超时处理逻辑."] + "fixes": ["Fix OpenAI deployment restrictions, fix cant duplicate assistant."] }, - "date": "2023-08-04", - "version": "0.39.3" + "date": "2024-07-17", + "version": "1.5.3" }, { "children": { - "improvements": ["优化 topic 样式."] + "fixes": ["Fix session not reorder after send message."] }, - "date": "2023-08-04", - "version": "0.39.2" + "date": "2024-07-17", + "version": "1.5.2" }, { "children": { - "fixes": ["修正 basePath 在生产环境下不生效的问题."] + "improvements": ["Improve brand url."] }, - "date": "2023-08-04", - "version": "0.39.1" + "date": "2024-07-17", + "version": "1.5.1" }, { "children": { - "features": ["支持多轮的插件意图识别,支持自定义 OpenAI 代理地址."], - "improvements": ["优化插件的展示逻辑."] + "features": ["Spport qwen-vl and tool call for qwen."] }, - "date": "2023-08-04", - "version": "0.39.0" + "date": "2024-07-17", + "version": "1.5.0" }, { "children": { - "features": ["Add topic empty."] + "fixes": ["Save assistant info on blur."] }, - "date": "2023-08-04", - "version": "0.38.0" + "date": "2024-07-15", + "version": "1.4.3" }, { "children": { - "features": ["支持使用全局助手的设置作为默认助手的创建角色."] + "improvements": ["Remove code related to gemini-pro-vision."] }, - "date": "2023-08-03", - "version": "0.37.0" + "date": "2024-07-13", + "version": "1.4.2" }, { "children": { - "improvements": ["Refactor zustand usage with v4.4."] + "improvements": ["Add cloud promotion banner."] }, - "date": "2023-08-03", - "version": "0.36.1" + "date": "2024-07-13", + "version": "1.4.1" }, { "children": { - "features": ["实现自定义历史消息数功能."], - "fixes": ["Fix setting type."], - "improvements": ["Fix session item height."] + "features": ["Add 360AI model provider."] }, - "date": "2023-08-03", - "version": "0.36.0" + "date": "2024-07-12", + "version": "1.4.0" }, { "children": { - "improvements": ["Update doc mode and token tags."] + "improvements": ["Improve agent runtime code."] }, - "date": "2023-07-31", - "version": "0.35.1" + "date": "2024-07-11", + "version": "1.3.6" }, { "children": { - "features": [ - "Add agent settings functionality, new components, and features for AgentMeta, Add and modify translations for various keys in JSON code files." - ] + "fixes": ["Fix assistant meta change race issue."] }, - "date": "2023-07-31", - "version": "0.35.0" + "date": "2024-07-10", + "version": "1.3.5" }, { "children": { - "features": [ - "Add agent settings functionality, Add new components and features for AgentMeta, Improve organization and functionality of settings and configuration features." - ] + "improvements": ["Support disable clientFetch by default."] }, - "date": "2023-07-31", - "version": "0.34.0" + "date": "2024-07-09", + "version": "1.3.4" }, { "children": { - "features": ["支持输入模板预处理."] + "fixes": ["Allow user to use their own WebRTC signaling."] }, - "date": "2023-07-30", - "version": "0.33.0" + "date": "2024-07-09", + "version": "1.3.3" }, { "children": { - "features": ["支持会话置顶."] + "fixes": ["Automatic refresh when change language."] }, - "date": "2023-07-30", - "version": "0.32.0" + "date": "2024-07-09", + "version": "1.3.2" }, { - "children": { - "features": ["支持展示 token 使用量."] - }, - "date": "2023-07-30", - "version": "0.31.0" + "children": {}, + "date": "2024-07-09", + "version": "1.3.1" }, { "children": { - "improvements": ["优化搜索引擎插件展示."] + "features": ["Add Taichu model provider."] }, - "date": "2023-07-30", - "version": "0.30.1" + "date": "2024-07-09", + "version": "1.3.0" }, { "children": { - "features": [ - "优化保存为话题功能,实现 Topic 重命名功能,实现话题删除功能,支持缓存角色面板的展开折叠状态." - ] + "improvements": ["Provider changes with model in model settings."] }, - "date": "2023-07-30", - "version": "0.30.0" + "date": "2024-07-08", + "version": "1.2.14" }, { "children": { - "features": ["实现单个会话和角色的导出功能,实现清空所有会话消息."] + "fixes": ["Fix tool message order."] }, - "date": "2023-07-30", - "version": "0.29.0" + "date": "2024-07-07", + "version": "1.2.13" }, { "children": { - "improvements": ["重构 settings 相关类型."], - "features": [ - "优化 SideBar 实现,激活态指示更加明确,实现 session 导入功能,实现配置导出功能." + "fixes": [ + "Fixed mobile web page navigation issue with inbox assistant, support to disable clerk signup with feature flag." ] }, - "date": "2023-07-30", - "version": "0.28.0" + "date": "2024-07-07", + "version": "1.2.12" }, { "children": { - "fixes": ["修正日志超过 4096 长度的问题."] + "improvements": ["Update deepseek max token."] }, - "date": "2023-07-29", - "version": "0.27.4" + "date": "2024-07-07", + "version": "1.2.11" }, { "children": { - "fixes": ["修正返回结果导致插件无法正常识别的问题."], - "improvements": ["优化样式."] + "fixes": ["Improve tools calling error feedback when arguments are not correctly."] }, - "date": "2023-07-29", - "version": "0.27.3" + "date": "2024-07-05", + "version": "1.2.10" }, { "children": { - "improvements": ["重构并优化文档抓取插件能力."] + "improvements": ["Fix tool message suspense loading."] }, - "date": "2023-07-29", - "version": "0.27.2" + "date": "2024-07-04", + "version": "1.2.9" }, { "children": { - "improvements": ["优化搜索引擎样式."] + "fixes": ["Allow builtin tools to trigger AI message."] }, - "date": "2023-07-29", - "version": "0.27.1" + "date": "2024-07-03", + "version": "1.2.8" }, { "children": { - "features": ["优化搜索引擎插件交互展示."], - "improvements": ["优化兜底结果展示."] + "improvements": ["Improve delete assistant message with tools."] }, - "date": "2023-07-29", - "version": "0.27.0" + "date": "2024-07-03", + "version": "1.2.7" }, { "children": { - "improvements": ["优化 setting Layout 实现."] + "fixes": ["Clerk provider refreshes continously."] }, - "date": "2023-07-29", - "version": "0.26.1" + "date": "2024-07-03", + "version": "1.2.6" }, { "children": { - "features": ["support password auth and error."] + "fixes": ["Fix clerk appearance is not applied correctly."] }, - "date": "2023-07-28", - "version": "0.26.0" - }, - { - "children": {}, - "date": "2023-07-26", - "version": "0.25.0" + "date": "2024-07-02", + "version": "1.2.5" }, { "children": { - "features": [ - "Add new translations, update existing translations, add functionality to components, modify styling, and adjust placeholder text" - ] + "improvements": ["Update ProviderAvatar for Baichuan & Stepfun."] }, - "date": "2023-07-26", - "version": "0.24.0" + "date": "2024-07-02", + "version": "1.2.4" }, { "children": { - "features": ["Add new features, update URLs, customize appearance, and implement components"] + "improvements": ["Refactor the portal implement."] }, - "date": "2023-07-26", - "version": "0.23.0" + "date": "2024-07-01", + "version": "1.2.3" }, { "children": { - "improvements": ["优化 tooltip 显示."] + "fixes": ["Display issue when select default model in System Agent."] }, - "date": "2023-07-26", - "version": "0.22.2" + "date": "2024-07-01", + "version": "1.2.2" }, { "children": { - "fixes": ["修正自定义 OpenAI API Key 的使用问题."] + "improvements": ["Refactor the portal file."] }, - "date": "2023-07-25", - "version": "0.22.1" + "date": "2024-07-01", + "version": "1.2.1" }, { "children": { - "features": ["支持使用自定义 OpenAI Key."] + "features": ["Add Baichuan model provider."] }, - "date": "2023-07-25", - "version": "0.22.0" + "date": "2024-07-01", + "version": "1.2.0" }, { "children": { - "improvements": ["Move component folder."], - "features": ["支持快捷配置模型、温度."] + "improvements": ["Add Gemini 1.5 stable version model & 2M context window."] }, - "date": "2023-07-25", - "version": "0.21.0" + "date": "2024-07-01", + "version": "1.1.18" }, { "children": { - "features": ["实现话题模块."] + "improvements": ["Refactor to prepare for Chat Portal."] }, - "date": "2023-07-25", - "version": "0.20.0" + "date": "2024-07-01", + "version": "1.1.17" }, { "children": { - "improvements": ["将 message reducer 提取到独立文件夹中,清理无用代码实现."], - "features": ["数据结构层完成 topic 模型改造."] + "fixes": ["Fix clerk UNAUTHORIZED error after long-time hang-up."] }, - "date": "2023-07-24", - "version": "0.19.0" + "date": "2024-06-29", + "version": "1.1.16" }, { "children": { - "improvements": ["修正 markdown li 丢失的问题."] + "fixes": ["Pin @azure/core-rest-pipeline@1.16.0 to fix azure openai issue."] }, - "date": "2023-07-24", - "version": "0.18.2" + "date": "2024-06-28", + "version": "1.1.15" }, { "children": { - "improvements": ["优化新会话的创建逻辑 session."] + "improvements": ["Refactor plugin state."] }, - "date": "2023-07-24", - "version": "0.18.1" + "date": "2024-06-27", + "version": "1.1.14" }, { "children": { - "features": ["实现会话展示模式切换,并优化默认创建角色的配置."] + "fixes": ["-check_updates cannot be set by FEATURE_FLAGS."] }, - "date": "2023-07-24", - "version": "0.18.0" + "date": "2024-06-27", + "version": "1.1.13" }, { "children": { - "features": ["表单配置支持设定各项高级参数."] + "fixes": ["Fix azure tools calling."] }, - "date": "2023-07-24", - "version": "0.17.0" + "date": "2024-06-26", + "version": "1.1.12" }, { "children": { - "improvements": ["优化 document title."] + "improvements": ["Refactor format utils."] }, - "date": "2023-07-24", - "version": "0.16.1" + "date": "2024-06-25", + "version": "1.1.11" }, { - "children": { - "features": ["支持自动跳转到第一条会话."], - "improvements": ["修正插件的展示文案."] - }, - "date": "2023-07-24", - "version": "0.16.0" + "children": {}, + "date": "2024-06-24", + "version": "1.1.10" }, { "children": { - "improvements": ["更新插件文案."] + "fixes": ["Fix agent tags."], + "improvements": ["Always show action on mobile."] }, - "date": "2023-07-24", - "version": "0.15.1" + "date": "2024-06-24", + "version": "1.1.9" }, { "children": { - "features": [ - "Add new features and improve user experience, Import and use constants from \"meta.ts\" instead of \"agentConfig\"." - ] + "fixes": ["Fix anthropic parallel tools calling."] }, - "date": "2023-07-24", - "version": "0.15.0" + "date": "2024-06-24", + "version": "1.1.8" }, { "children": { - "features": ["支持网页抓取."] + "improvements": ["Add AES-GCM footer and tooltip."] }, - "date": "2023-07-24", - "version": "0.14.0" + "date": "2024-06-24", + "version": "1.1.7" }, { "children": { - "fixes": ["修正搜索引擎插件的实现问题."] + "improvements": ["Refactor the server db implement."], + "fixes": ["Fix incorrect baseURL for Groq in client mode."] }, - "date": "2023-07-23", - "version": "0.13.1" + "date": "2024-06-23", + "version": "1.1.6" }, { "children": { - "features": ["优化插件模式下的用户体验."] + "improvements": ["Remove deprecated env."] }, - "date": "2023-07-23", - "version": "0.13.0" + "date": "2024-06-23", + "version": "1.1.5" }, { "children": { - "fixes": ["修正 message parentId 不正确的问题."] + "fixes": ["Create first-time user on server db."] }, - "date": "2023-07-23", - "version": "0.12.1" + "date": "2024-06-22", + "version": "1.1.4" }, { "children": { - "features": ["支持插件列表 与 基于 Serpapi 的搜索引擎插件."] + "fixes": ["Ollama not enabled client fetch by default."] }, - "date": "2023-07-23", - "version": "0.12.0" + "date": "2024-06-21", + "version": "1.1.3" }, { - "children": { - "improvements": ["Update manifest, 增加国际化文案."], - "features": ["支持查询天气."] - }, - "date": "2023-07-23", - "version": "0.11.0" + "children": {}, + "date": "2024-06-20", + "version": "1.1.2" }, { "children": { - "improvements": ["优化模型在 list 中的展示逻辑."] + "improvements": ["Fixed System Agent missing in mobile layout."] }, - "date": "2023-07-23", - "version": "0.10.2" + "date": "2024-06-20", + "version": "1.1.1" }, { "children": { - "improvements": ["修正对话中用户头像的问题."] + "features": ["Supoort Claude 3.5 Sonnet."] }, - "date": "2023-07-22", - "version": "0.10.1" + "date": "2024-06-20", + "version": "1.1.0" }, { "children": { - "features": ["支持复制与编辑会话消息."] + "fixes": ["Fix to send image without text."] }, - "date": "2023-07-22", - "version": "0.10.0" + "date": "2024-06-20", + "version": "1.0.14" }, { "children": { - "features": ["展示模型类型."] + "fixes": ["Fix and improve tool calling."] }, - "date": "2023-07-22", - "version": "0.9.0" + "date": "2024-06-19", + "version": "1.0.13" }, { "children": { - "fixes": ["Fix miss manifest.json link, 优化 model tag 展示逻辑."] + "fixes": ["Fix auto avatar."] }, - "date": "2023-07-22", - "version": "0.8.2" + "date": "2024-06-19", + "version": "1.0.12" }, { "children": { - "fixes": ["Fix import."] + "improvements": ["Fix phi3 icon display under OpenRouter."] }, - "date": "2023-07-22", - "version": "0.8.1" + "date": "2024-06-19", + "version": "1.0.11" }, { "children": { - "features": ["支持 pwa 模式."] + "fixes": ["Typo in reset settings."] }, - "date": "2023-07-22", - "version": "0.8.0" + "date": "2024-06-19", + "version": "1.0.10" }, { - "children": { - "features": ["支持展示来自模型的标记信息."] - }, - "date": "2023-07-22", - "version": "0.7.0" + "children": {}, + "date": "2024-06-19", + "version": "1.0.9" }, { "children": { - "fixes": ["Add deps."] + "fixes": ["Fix hydration mismatch on macOS."], + "improvements": ["Update brand and improve docs."] }, - "date": "2023-07-22", - "version": "0.6.1" + "date": "2024-06-18", + "version": "1.0.8" }, { "children": { - "improvements": ["重构 selector 文件组织."], - "features": ["补充 token 详情."] + "improvements": ["Remove some i18n text."] }, - "date": "2023-07-22", - "version": "0.6.0" + "date": "2024-06-18", + "version": "1.0.7" }, { "children": { - "features": ["支持选择 Emoji."], - "fixes": ["修正 total token 计算不正确的问题."] + "improvements": ["Fix footer being cut on wide screen."] }, - "date": "2023-07-22", - "version": "0.5.0" + "date": "2024-06-17", + "version": "1.0.6" }, { "children": { - "improvements": ["优化 edit 代码结构."] + "fixes": ["Fix onboard and auth for community version."] }, - "date": "2023-07-22", - "version": "0.4.3" + "date": "2024-06-17", + "version": "1.0.5" }, { "children": { - "improvements": ["Fix input style, fix layout."] + "improvements": ["Add stepfun as a new provider."] }, - "date": "2023-07-22", - "version": "0.4.2" + "date": "2024-06-17", + "version": "1.0.4" }, { "children": { - "fixes": ["Fix SSR style error."] + "fixes": ["Fix clerk UNAUTHORIZED auth error."] }, - "date": "2023-07-22", - "version": "0.4.1" + "date": "2024-06-17", + "version": "1.0.3" }, { "children": { - "features": [ - "Add styles and modify layout of FolderPanel, SliderWithInput, SessionList, EditPage, ChatLayout, and SettingLayout components, Introduce FOLDER_WIDTH constant and update components." - ] + "fixes": ["Openai key and openai proxy are invalid in feature flags."] }, - "date": "2023-07-20", - "version": "0.4.0" + "date": "2024-06-17", + "version": "1.0.2" }, { "children": { - "features": ["Add new files, modify components, and adjust layout and styling."] + "improvements": ["Remove r2 cdn url."] }, - "date": "2023-07-18", - "version": "0.3.0" + "date": "2024-06-17", + "version": "1.0.1" }, { "children": { - "features": ["Add import statement and define CSS styles for Avatar component."] + "features": ["Release LobeChat 1.0."], + "improvements": [ + "Bump version.", + "Release LobeChat 1.0, closes #2897 (97bb377)", + "Bump version, closes #2896 (6e2d03b)", + "Update LICENSE to Apache 2.0", + "update to 1.0.0" + ] }, - "date": "2023-07-18", - "version": "0.2.0" - }, - { - "children": {}, - "date": "2023-07-18", - "version": "0.1.6" + "date": "2024-06-17", + "version": "1.0.0" } ] diff --git a/src/app/(main)/changelog/page.tsx b/src/app/(main)/changelog/page.tsx index 861735fe65c6..fe8ae4edeb97 100644 --- a/src/app/(main)/changelog/page.tsx +++ b/src/app/(main)/changelog/page.tsx @@ -38,6 +38,8 @@ const Page = async () => { const changelogService = new ChangelogService(); const data = await changelogService.getChangelogIndex(); + if (!data) return notFound(); + const ld = ldModule.generate({ description: t('changelog.description', { appName: BRANDING_NAME }), title: t('changelog.title', { appName: BRANDING_NAME }), @@ -48,7 +50,7 @@ const Page = async () => { <> - {data.map((item) => ( + {data?.map((item) => ( (({ identifier, providerData, data } tags: ( ), diff --git a/src/app/(main)/discover/(detail)/model/[...slugs]/features/Header.tsx b/src/app/(main)/discover/(detail)/model/[...slugs]/features/Header.tsx index eb22a4b2eaa7..c5cbab937ef7 100644 --- a/src/app/(main)/discover/(detail)/model/[...slugs]/features/Header.tsx +++ b/src/app/(main)/discover/(detail)/model/[...slugs]/features/Header.tsx @@ -75,7 +75,7 @@ const Header = memo(({ identifier, data, mobile }) => { {data.meta.description &&
{t(`${identifier}.description`, { ns: 'models' })}
}
diff --git a/src/app/(main)/discover/(detail)/model/[...slugs]/features/InfoSidebar/SuggestionItem.tsx b/src/app/(main)/discover/(detail)/model/[...slugs]/features/InfoSidebar/SuggestionItem.tsx index 2a4953322fba..4023bf44510f 100644 --- a/src/app/(main)/discover/(detail)/model/[...slugs]/features/InfoSidebar/SuggestionItem.tsx +++ b/src/app/(main)/discover/(detail)/model/[...slugs]/features/InfoSidebar/SuggestionItem.tsx @@ -45,7 +45,7 @@ export interface SuggestionItemProps FlexboxProps {} const SuggestionItem = memo(({ className, meta, identifier, ...rest }) => { - const { title, description, tokens, vision, functionCall } = meta; + const { title, description, contextWindowTokens, vision, functionCall } = meta; const { t } = useTranslation('models'); const { cx, styles } = useStyles(); @@ -67,7 +67,7 @@ const SuggestionItem = memo(({ className, meta, identifier, {t(`${identifier}.description`)} )} - + ); }); diff --git a/src/app/(main)/discover/(detail)/model/[...slugs]/features/ProviderList/ProviderItem.tsx b/src/app/(main)/discover/(detail)/model/[...slugs]/features/ProviderList/ProviderItem.tsx index fb5aedc06946..735a7487fae0 100644 --- a/src/app/(main)/discover/(detail)/model/[...slugs]/features/ProviderList/ProviderItem.tsx +++ b/src/app/(main)/discover/(detail)/model/[...slugs]/features/ProviderList/ProviderItem.tsx @@ -45,7 +45,7 @@ const ProviderItem = memo(({ mobile, modelId, identifier }) = const items: StatisticProps[] = [ { title: t('models.contentLength'), - value: model?.tokens ? formatTokenNumber(model.tokens) : '--', + value: model?.contextWindowTokens ? formatTokenNumber(model.contextWindowTokens) : '--', }, { title: t('models.providerInfo.maxOutput'), diff --git a/src/app/(main)/discover/(detail)/provider/[slug]/features/ModelList/ModelItem.tsx b/src/app/(main)/discover/(detail)/provider/[slug]/features/ModelList/ModelItem.tsx index f91a6a75e43f..3e9114c4a50c 100644 --- a/src/app/(main)/discover/(detail)/provider/[slug]/features/ModelList/ModelItem.tsx +++ b/src/app/(main)/discover/(detail)/provider/[slug]/features/ModelList/ModelItem.tsx @@ -47,7 +47,7 @@ export interface SuggestionItemProps } const ModelItem = memo(({ mobile, meta, identifier }) => { - const { title, tokens, vision, functionCall } = meta; + const { title, contextWindowTokens, vision, functionCall } = meta; const { xl = true } = useResponsive(); const { t } = useTranslation('discover'); const { styles, theme } = useStyles(); @@ -57,7 +57,7 @@ const ModelItem = memo(({ mobile, meta, identifier }) => { const items: StatisticProps[] = [ { title: t('models.contentLength'), - value: meta?.tokens ? formatTokenNumber(meta.tokens) : '--', + value: meta?.contextWindowTokens ? formatTokenNumber(meta.contextWindowTokens) : '--', }, { title: t('models.providerInfo.maxOutput'), @@ -98,7 +98,7 @@ const ModelItem = memo(({ mobile, meta, identifier }) => { - + ); diff --git a/src/app/(main)/discover/(list)/models/features/Card.tsx b/src/app/(main)/discover/(list)/models/features/Card.tsx index 446057e5bb15..4e9295ff3326 100644 --- a/src/app/(main)/discover/(list)/models/features/Card.tsx +++ b/src/app/(main)/discover/(list)/models/features/Card.tsx @@ -72,7 +72,7 @@ export interface ModelCardProps extends DiscoverModelItem, FlexboxProps { } const ModelCard = memo(({ className, meta, identifier, ...rest }) => { - const { description, title, functionCall, vision, tokens } = meta; + const { description, title, functionCall, vision, contextWindowTokens } = meta; const { t } = useTranslation('models'); const { cx, styles } = useStyles(); @@ -107,7 +107,11 @@ const ModelCard = memo(({ className, meta, identifier, ...rest } )} - + ); diff --git a/src/app/(main)/settings/llm/ProviderList/Cloudflare/index.tsx b/src/app/(main)/settings/llm/ProviderList/Cloudflare/index.tsx index f4cf3e72774d..bb5ac7afcf8b 100644 --- a/src/app/(main)/settings/llm/ProviderList/Cloudflare/index.tsx +++ b/src/app/(main)/settings/llm/ProviderList/Cloudflare/index.tsx @@ -29,11 +29,7 @@ export const useCloudflareProvider = (): ProviderItem => { name: [KeyVaultsConfigKey, providerKey, 'apiKey'], }, { - children: ( - - ), + children: , desc: t(`${providerKey}.baseURLOrAccountID.desc`), label: t(`${providerKey}.baseURLOrAccountID.title`), name: [KeyVaultsConfigKey, providerKey, 'baseURLOrAccountID'], diff --git a/src/app/(main)/settings/llm/components/ProviderModelList/ModelFetcher.tsx b/src/app/(main)/settings/llm/components/ProviderModelList/ModelFetcher.tsx index 30eebd7ae230..c8228d549b84 100644 --- a/src/app/(main)/settings/llm/components/ProviderModelList/ModelFetcher.tsx +++ b/src/app/(main)/settings/llm/components/ProviderModelList/ModelFetcher.tsx @@ -80,8 +80,8 @@ const ModelFetcher = memo(({ provider }) => { title={ latestFetchTime ? t('llm.fetcher.latestTime', { - time: dayjs(latestFetchTime).format('YYYY-MM-DD HH:mm:ss'), - }) + time: dayjs(latestFetchTime).format('YYYY-MM-DD HH:mm:ss'), + }) : t('llm.fetcher.noLatestTime') } > diff --git a/src/app/@modal/(.)changelog/modal/page.tsx b/src/app/@modal/(.)changelog/modal/page.tsx index 57cc070898d4..7114643a1a41 100644 --- a/src/app/@modal/(.)changelog/modal/page.tsx +++ b/src/app/@modal/(.)changelog/modal/page.tsx @@ -20,9 +20,11 @@ const Page = async () => { const changelogService = new ChangelogService(); const data = await changelogService.getChangelogIndex(); + if (!data) return notFound(); + return ( <> - {data.map((item) => ( + {data?.map((item) => ( } key={item.id}> diff --git a/src/components/ModelSelect/index.tsx b/src/components/ModelSelect/index.tsx index 7820c00b20af..069c35e2d7ae 100644 --- a/src/components/ModelSelect/index.tsx +++ b/src/components/ModelSelect/index.tsx @@ -102,19 +102,22 @@ export const ModelInfoTags = memo( )} - {model.tokens !== undefined && ( + {model.contextWindowTokens !== undefined && (
- {model.tokens === 0 ? ( + {model.contextWindowTokens === 0 ? ( ) : ( - formatTokenNumber(model.tokens) + formatTokenNumber(model.contextWindowTokens) )}
diff --git a/src/config/__tests__/app.test.ts b/src/config/__tests__/app.test.ts index 68e7cc33677e..d454e58d03fc 100644 --- a/src/config/__tests__/app.test.ts +++ b/src/config/__tests__/app.test.ts @@ -24,7 +24,9 @@ describe('getServerConfig', () => { describe('index url', () => { it('should return default URLs when no environment variables are set', () => { const config = getAppConfig(); - expect(config.AGENTS_INDEX_URL).toBe('https://registry.npmmirror.com/@lobehub/agents-index/v1/files/public'); + expect(config.AGENTS_INDEX_URL).toBe( + 'https://registry.npmmirror.com/@lobehub/agents-index/v1/files/public', + ); expect(config.PLUGINS_INDEX_URL).toBe('https://chat-plugins.lobehub.com'); }); @@ -41,7 +43,9 @@ describe('getServerConfig', () => { process.env.PLUGINS_INDEX_URL = ''; const config = getAppConfig(); - expect(config.AGENTS_INDEX_URL).toBe('https://registry.npmmirror.com/@lobehub/agents-index/v1/files/public'); + expect(config.AGENTS_INDEX_URL).toBe( + 'https://registry.npmmirror.com/@lobehub/agents-index/v1/files/public', + ); expect(config.PLUGINS_INDEX_URL).toBe('https://chat-plugins.lobehub.com'); }); }); diff --git a/src/config/app.ts b/src/config/app.ts index dd9ee235bf61..fef3ca3a4b37 100644 --- a/src/config/app.ts +++ b/src/config/app.ts @@ -23,8 +23,7 @@ if (typeof window === 'undefined' && isServerMode && !APP_URL) { throw new Error('`APP_URL` is required in server mode'); } -const ASSISTANT_INDEX_URL = - 'https://registry.npmmirror.com/@lobehub/agents-index/v1/files/public'; +const ASSISTANT_INDEX_URL = 'https://registry.npmmirror.com/@lobehub/agents-index/v1/files/public'; const PLUGINS_INDEX_URL = 'https://chat-plugins.lobehub.com'; diff --git a/src/config/modelProviders/ai21.ts b/src/config/modelProviders/ai21.ts index 2ac3ac912252..055f37dc6dc3 100644 --- a/src/config/modelProviders/ai21.ts +++ b/src/config/modelProviders/ai21.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const Ai21: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 256_000, displayName: 'Jamba 1.5 Mini', enabled: true, functionCall: true, @@ -12,9 +13,9 @@ const Ai21: ModelProviderCard = { input: 0.2, output: 0.4, }, - tokens: 256_000, }, { + contextWindowTokens: 256_000, displayName: 'Jamba 1.5 Large', enabled: true, functionCall: true, @@ -23,7 +24,6 @@ const Ai21: ModelProviderCard = { input: 2, output: 8, }, - tokens: 256_000, }, ], checkModel: 'jamba-1.5-mini', diff --git a/src/config/modelProviders/ai360.ts b/src/config/modelProviders/ai360.ts index 2f2f5fab7e78..2b4fdf308ef1 100644 --- a/src/config/modelProviders/ai360.ts +++ b/src/config/modelProviders/ai360.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const Ai360: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 8192, description: '360GPT2 Pro 是 360 公司推出的高级自然语言处理模型,具备卓越的文本生成和理解能力,尤其在生成与创作领域表现出色,能够处理复杂的语言转换和角色演绎任务。', displayName: '360GPT2 Pro', @@ -15,9 +16,9 @@ const Ai360: ModelProviderCard = { input: 5, output: 5, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '360GPT Pro 作为 360 AI 模型系列的重要成员,以高效的文本处理能力满足多样化的自然语言应用场景,支持长文本理解和多轮对话等功能。', displayName: '360GPT Pro', @@ -30,9 +31,9 @@ const Ai360: ModelProviderCard = { input: 5, output: 5, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '360GPT Turbo 提供强大的计算和对话能力,具备出色的语义理解和生成效率,是企业和开发者理想的智能助理解决方案。', displayName: '360GPT Turbo', @@ -44,9 +45,9 @@ const Ai360: ModelProviderCard = { input: 2, output: 2, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '360GPT Turbo Responsibility 8K 强调语义安全和责任导向,专为对内容安全有高度要求的应用场景设计,确保用户体验的准确性与稳健性。', displayName: '360GPT Turbo Responsibility 8K', @@ -58,7 +59,6 @@ const Ai360: ModelProviderCard = { input: 2, output: 2, }, - tokens: 8192, }, ], checkModel: '360gpt-turbo', diff --git a/src/config/modelProviders/anthropic.ts b/src/config/modelProviders/anthropic.ts index 1a5e4092145a..4ceb7a55b067 100644 --- a/src/config/modelProviders/anthropic.ts +++ b/src/config/modelProviders/anthropic.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const Anthropic: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 200_000, description: 'Claude 3.5 Haiku 是 Anthropic 最快的下一代模型。与 Claude 3 Haiku 相比,Claude 3.5 Haiku 在各项技能上都有所提升,并在许多智力基准测试中超越了上一代最大的模型 Claude 3 Opus。', displayName: 'Claude 3.5 Haiku', @@ -18,9 +19,9 @@ const Anthropic: ModelProviderCard = { writeCacheInput: 1.25, }, releasedAt: '2024-11-05', - tokens: 200_000, }, { + contextWindowTokens: 200_000, description: 'Claude 3.5 Sonnet 提供了超越 Opus 的能力和比 Sonnet 更快的速度,同时保持与 Sonnet 相同的价格。Sonnet 特别擅长编程、数据科学、视觉处理、代理任务。', displayName: 'Claude 3.5 Sonnet', @@ -35,10 +36,10 @@ const Anthropic: ModelProviderCard = { writeCacheInput: 3.75, }, releasedAt: '2024-10-22', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3.5 Sonnet 提供了超越 Opus 的能力和比 Sonnet 更快的速度,同时保持与 Sonnet 相同的价格。Sonnet 特别擅长编程、数据科学、视觉处理、代理任务。', displayName: 'Claude 3.5 Sonnet 0620', @@ -52,10 +53,10 @@ const Anthropic: ModelProviderCard = { writeCacheInput: 3.75, }, releasedAt: '2024-06-20', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Haiku 是 Anthropic 的最快且最紧凑的模型,旨在实现近乎即时的响应。它具有快速且准确的定向性能。', displayName: 'Claude 3 Haiku', @@ -67,10 +68,10 @@ const Anthropic: ModelProviderCard = { output: 1.25, }, releasedAt: '2024-03-07', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Sonnet 在智能和速度方面为企业工作负载提供了理想的平衡。它以更低的价格提供最大效用,可靠且适合大规模部署。', displayName: 'Claude 3 Sonnet', @@ -82,10 +83,10 @@ const Anthropic: ModelProviderCard = { output: 15, }, releasedAt: '2024-02-29', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Opus 是 Anthropic 用于处理高度复杂任务的最强大模型。它在性能、智能、流畅性和理解力方面表现卓越。', displayName: 'Claude 3 Opus', @@ -98,10 +99,10 @@ const Anthropic: ModelProviderCard = { output: 75, }, releasedAt: '2024-02-29', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 2 为企业提供了关键能力的进步,包括业界领先的 200K token 上下文、大幅降低模型幻觉的发生率、系统提示以及一个新的测试功能:工具调用。', displayName: 'Claude 2.1', @@ -112,9 +113,9 @@ const Anthropic: ModelProviderCard = { output: 24, }, releasedAt: '2023-11-21', - tokens: 200_000, }, { + contextWindowTokens: 100_000, description: 'Claude 2 为企业提供了关键能力的进步,包括业界领先的 200K token 上下文、大幅降低模型幻觉的发生率、系统提示以及一个新的测试功能:工具调用。', displayName: 'Claude 2.0', @@ -125,7 +126,6 @@ const Anthropic: ModelProviderCard = { output: 24, }, releasedAt: '2023-07-11', - tokens: 100_000, }, ], checkModel: 'claude-3-haiku-20240307', diff --git a/src/config/modelProviders/azure.ts b/src/config/modelProviders/azure.ts index ce831b712ba6..a8d3c1157b90 100644 --- a/src/config/modelProviders/azure.ts +++ b/src/config/modelProviders/azure.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const Azure: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 16_385, deploymentName: 'gpt-35-turbo', description: 'GPT 3.5 Turbo,OpenAI提供的高效模型,适用于聊天和文本生成任务,支持并行函数调用。', @@ -12,44 +13,43 @@ const Azure: ModelProviderCard = { functionCall: true, id: 'gpt-35-turbo', maxOutput: 4096, - tokens: 16_385, }, { + contextWindowTokens: 16_384, deploymentName: 'gpt-35-turbo-16k', description: 'GPT 3.5 Turbo 16k,高容量文本生成模型,适合复杂任务。', displayName: 'GPT 3.5 Turbo', functionCall: true, id: 'gpt-35-turbo-16k', - tokens: 16_384, }, { + contextWindowTokens: 128_000, deploymentName: 'gpt-4-turbo', description: 'GPT 4 Turbo,多模态模型,提供杰出的语言理解和生成能力,同时支持图像输入。', displayName: 'GPT 4 Turbo', enabled: true, functionCall: true, id: 'gpt-4', - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, deploymentName: 'gpt-4o-mini', description: 'GPT-4o Mini,小型高效模型,具备与GPT-4o相似的卓越性能。', displayName: 'GPT 4o Mini', enabled: true, functionCall: true, id: 'gpt-4o-mini', - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, deploymentName: 'gpt-4o', description: 'GPT-4o 是最新的多模态模型,结合高级文本和图像处理能力。', displayName: 'GPT 4o', enabled: true, functionCall: true, id: 'gpt-4o', - tokens: 128_000, vision: true, }, ], diff --git a/src/config/modelProviders/baichuan.ts b/src/config/modelProviders/baichuan.ts index 988ee6448a79..4fb50285915d 100644 --- a/src/config/modelProviders/baichuan.ts +++ b/src/config/modelProviders/baichuan.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const Baichuan: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 32_768, description: '模型能力国内第一,在知识百科、长文本、生成创作等中文任务上超越国外主流模型。还具备行业领先的多模态能力,多项权威评测基准表现优异。', displayName: 'Baichuan 4', @@ -16,9 +17,9 @@ const Baichuan: ModelProviderCard = { input: 100, output: 100, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: '模型能力国内第一,在知识百科、长文本、生成创作等中文任务上超越国外主流模型。还具备行业领先的多模态能力,多项权威评测基准表现优异。', displayName: 'Baichuan 4 Turbo', @@ -31,9 +32,9 @@ const Baichuan: ModelProviderCard = { input: 15, output: 15, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: '模型能力国内第一,在知识百科、长文本、生成创作等中文任务上超越国外主流模型。还具备行业领先的多模态能力,多项权威评测基准表现优异。', displayName: 'Baichuan 4 Air', @@ -46,9 +47,9 @@ const Baichuan: ModelProviderCard = { input: 0.98, output: 0.98, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: '针对企业高频场景优化,效果大幅提升,高性价比。相对于Baichuan2模型,内容创作提升20%,知识问答提升17%, 角色扮演能力提升40%。整体效果比GPT3.5更优。', displayName: 'Baichuan 3 Turbo', @@ -60,9 +61,9 @@ const Baichuan: ModelProviderCard = { input: 12, output: 12, }, - tokens: 32_768, }, { + contextWindowTokens: 128_000, description: '具备 128K 超长上下文窗口,针对企业高频场景优化,效果大幅提升,高性价比。相对于Baichuan2模型,内容创作提升20%,知识问答提升17%, 角色扮演能力提升40%。整体效果比GPT3.5更优。', displayName: 'Baichuan 3 Turbo 128k', @@ -73,9 +74,9 @@ const Baichuan: ModelProviderCard = { input: 24, output: 24, }, - tokens: 128_000, }, { + contextWindowTokens: 32_768, description: '采用搜索增强技术实现大模型与领域知识、全网知识的全面链接。支持PDF、Word等多种文档上传及网址输入,信息获取及时、全面,输出结果准确、专业。', displayName: 'Baichuan 2 Turbo', @@ -86,7 +87,6 @@ const Baichuan: ModelProviderCard = { input: 8, output: 8, }, - tokens: 32_768, }, ], checkModel: 'Baichuan3-Turbo', diff --git a/src/config/modelProviders/bedrock.ts b/src/config/modelProviders/bedrock.ts index 8980d1439be9..5f249ed7b96b 100644 --- a/src/config/modelProviders/bedrock.ts +++ b/src/config/modelProviders/bedrock.ts @@ -27,6 +27,7 @@ const Bedrock: ModelProviderCard = { }, */ { + contextWindowTokens: 200_000, description: 'Claude 3.5 Sonnet 提升了行业标准,性能超过竞争对手模型和 Claude 3 Opus,在广泛的评估中表现出色,同时具有我们中等层级模型的速度和成本。', displayName: 'Claude 3.5 Sonnet', @@ -37,10 +38,10 @@ const Bedrock: ModelProviderCard = { input: 3, output: 15, }, - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3.5 Sonnet 提升了行业标准,性能超过竞争对手模型和 Claude 3 Opus,在广泛的评估中表现出色,同时具有我们中等层级模型的速度和成本。', displayName: 'Claude 3.5 Sonnet v2 (Inference profile)', @@ -51,10 +52,10 @@ const Bedrock: ModelProviderCard = { input: 3, output: 15, }, - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3.5 Sonnet 提升了行业标准,性能超过竞争对手模型和 Claude 3 Opus,在广泛的评估中表现出色,同时具有我们中等层级模型的速度和成本。', displayName: 'Claude 3.5 Sonnet 0620', @@ -65,10 +66,10 @@ const Bedrock: ModelProviderCard = { input: 3, output: 15, }, - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Haiku 是 Anthropic 最快、最紧凑的模型,提供近乎即时的响应速度。它可以快速回答简单的查询和请求。客户将能够构建模仿人类互动的无缝 AI 体验。Claude 3 Haiku 可以处理图像并返回文本输出,具有 200K 的上下文窗口。', displayName: 'Claude 3 Haiku', @@ -79,10 +80,10 @@ const Bedrock: ModelProviderCard = { input: 0.25, output: 1.25, }, - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Anthropic 的 Claude 3 Sonnet 在智能和速度之间达到了理想的平衡——特别适合企业工作负载。它以低于竞争对手的价格提供最大的效用,并被设计成为可靠的、高耐用的主力机,适用于规模化的 AI 部署。Claude 3 Sonnet 可以处理图像并返回文本输出,具有 200K 的上下文窗口。', displayName: 'Claude 3 Sonnet', @@ -93,10 +94,10 @@ const Bedrock: ModelProviderCard = { input: 3, output: 15, }, - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Opus 是 Anthropic 最强大的 AI 模型,具有在高度复杂任务上的最先进性能。它可以处理开放式提示和未见过的场景,具有出色的流畅性和类人的理解能力。Claude 3 Opus 展示了生成 AI 可能性的前沿。Claude 3 Opus 可以处理图像并返回文本输出,具有 200K 的上下文窗口。', displayName: 'Claude 3 Opus', @@ -107,10 +108,10 @@ const Bedrock: ModelProviderCard = { input: 15, output: 75, }, - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 2 的更新版,具有双倍的上下文窗口,以及在长文档和 RAG 上下文中的可靠性、幻觉率和基于证据的准确性的改进。', displayName: 'Claude 2.1', @@ -120,9 +121,9 @@ const Bedrock: ModelProviderCard = { input: 8, output: 24, }, - tokens: 200_000, }, { + contextWindowTokens: 100_000, description: 'Anthropic 在从复杂对话和创意内容生成到详细指令跟随的广泛任务中都表现出高度能力的模型。', displayName: 'Claude 2.0', @@ -131,9 +132,9 @@ const Bedrock: ModelProviderCard = { input: 8, output: 24, }, - tokens: 100_000, }, { + contextWindowTokens: 100_000, description: '一款快速、经济且仍然非常有能力的模型,可以处理包括日常对话、文本分析、总结和文档问答在内的一系列任务。', displayName: 'Claude Instant', @@ -142,9 +143,9 @@ const Bedrock: ModelProviderCard = { input: 0.8, output: 2.4, }, - tokens: 100_000, }, { + contextWindowTokens: 128_000, description: 'Meta Llama 3.1 8B Instruct 的更新版,包括扩展的 128K 上下文长度、多语言性和改进的推理能力。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B、70B 和 405B 大小 (文本输入/输出)。Llama 3.1 指令调整的文本模型 (8B、70B、405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。', displayName: 'Llama 3.1 8B Instruct', @@ -155,9 +156,9 @@ const Bedrock: ModelProviderCard = { input: 0.22, output: 0.22, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Meta Llama 3.1 70B Instruct 的更新版,包括扩展的 128K 上下文长度、多语言性和改进的推理能力。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B、70B 和 405B 大小 (文本输入/输出)。Llama 3.1 指令调整的文本模型 (8B、70B、405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。', displayName: 'Llama 3.1 70B Instruct', @@ -168,9 +169,9 @@ const Bedrock: ModelProviderCard = { input: 0.99, output: 0.99, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Meta Llama 3.1 405B Instruct 是 Llama 3.1 Instruct 模型中最大、最强大的模型,是一款高度先进的对话推理和合成数据生成模型,也可以用作在特定领域进行专业持续预训练或微调的基础。Llama 3.1 提供的多语言大型语言模型 (LLMs) 是一组预训练的、指令调整的生成模型,包括 8B、70B 和 405B 大小 (文本输入/输出)。Llama 3.1 指令调整的文本模型 (8B、70B、405B) 专为多语言对话用例进行了优化,并在常见的行业基准测试中超过了许多可用的开源聊天模型。Llama 3.1 旨在用于多种语言的商业和研究用途。指令调整的文本模型适用于类似助手的聊天,而预训练模型可以适应各种自然语言生成任务。Llama 3.1 模型还支持利用其模型的输出来改进其他模型,包括合成数据生成和精炼。Llama 3.1 是使用优化的变压器架构的自回归语言模型。调整版本使用监督微调 (SFT) 和带有人类反馈的强化学习 (RLHF) 来符合人类对帮助性和安全性的偏好。', displayName: 'Llama 3.1 405B Instruct', @@ -181,9 +182,9 @@ const Bedrock: ModelProviderCard = { input: 5.32, output: 16, }, - tokens: 128_000, }, { + contextWindowTokens: 8000, description: 'Meta Llama 3 是一款面向开发者、研究人员和企业的开放大型语言模型 (LLM),旨在帮助他们构建、实验并负责任地扩展他们的生成 AI 想法。作为全球社区创新的基础系统的一部分,它非常适合计算能力和资源有限、边缘设备和更快的训练时间。', displayName: 'Llama 3 8B Instruct', @@ -192,9 +193,9 @@ const Bedrock: ModelProviderCard = { input: 0.3, output: 0.6, }, - tokens: 8000, }, { + contextWindowTokens: 8000, description: 'Meta Llama 3 是一款面向开发者、研究人员和企业的开放大型语言模型 (LLM),旨在帮助他们构建、实验并负责任地扩展他们的生成 AI 想法。作为全球社区创新的基础系统的一部分,它非常适合内容创建、对话 AI、语言理解、研发和企业应用。', displayName: 'Llama 3 70B Instruct', @@ -203,7 +204,6 @@ const Bedrock: ModelProviderCard = { input: 2.65, output: 3.5, }, - tokens: 8000, }, /* // TODO: Not support for now diff --git a/src/config/modelProviders/cloudflare.ts b/src/config/modelProviders/cloudflare.ts index 2db437c5e393..2d1207be8610 100644 --- a/src/config/modelProviders/cloudflare.ts +++ b/src/config/modelProviders/cloudflare.ts @@ -5,68 +5,69 @@ import { ModelProviderCard } from '@/types/llm'; const Cloudflare: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 16_384, displayName: 'deepseek-coder-6.7b-instruct-awq', enabled: true, id: '@hf/thebloke/deepseek-coder-6.7b-instruct-awq', - tokens: 16_384, }, { + contextWindowTokens: 2048, displayName: 'gemma-7b-it', enabled: true, id: '@hf/google/gemma-7b-it', - tokens: 2048, }, { + contextWindowTokens: 4096, displayName: 'hermes-2-pro-mistral-7b', + enabled: true, // functionCall: true, id: '@hf/nousresearch/hermes-2-pro-mistral-7b', - tokens: 4096, }, { + contextWindowTokens: 8192, displayName: 'llama-3-8b-instruct-awq', id: '@cf/meta/llama-3-8b-instruct-awq', - tokens: 8192, }, { + contextWindowTokens: 4096, displayName: 'mistral-7b-instruct-v0.2', id: '@hf/mistral/mistral-7b-instruct-v0.2', - tokens: 4096, }, { + contextWindowTokens: 32_768, displayName: 'neural-chat-7b-v3-1-awq', enabled: true, id: '@hf/thebloke/neural-chat-7b-v3-1-awq', - tokens: 32_768, }, { + contextWindowTokens: 8192, displayName: 'openchat-3.5-0106', id: '@cf/openchat/openchat-3.5-0106', - tokens: 8192, }, { + contextWindowTokens: 32_768, displayName: 'openhermes-2.5-mistral-7b-awq', enabled: true, id: '@hf/thebloke/openhermes-2.5-mistral-7b-awq', - tokens: 32_768, }, { + contextWindowTokens: 32_768, displayName: 'qwen1.5-14b-chat-awq', enabled: true, id: '@cf/qwen/qwen1.5-14b-chat-awq', - tokens: 32_768, }, { + contextWindowTokens: 4096, displayName: 'starling-lm-7b-beta', enabled: true, id: '@hf/nexusflow/starling-lm-7b-beta', - tokens: 4096, }, { + contextWindowTokens: 32_768, displayName: 'zephyr-7b-beta-awq', enabled: true, id: '@hf/thebloke/zephyr-7b-beta-awq', - tokens: 32_768, }, { displayName: 'meta-llama-3-8b-instruct', diff --git a/src/config/modelProviders/deepseek.ts b/src/config/modelProviders/deepseek.ts index 027b1a9fb4a0..12d190385bfb 100644 --- a/src/config/modelProviders/deepseek.ts +++ b/src/config/modelProviders/deepseek.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const DeepSeek: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 65_536, description: '融合通用与代码能力的全新开源模型, 不仅保留了原有 Chat 模型的通用对话能力和 Coder 模型的强大代码处理能力,还更好地对齐了人类偏好。此外,DeepSeek-V2.5 在写作任务、指令跟随等多个方面也实现了大幅提升。', displayName: 'DeepSeek V2.5', @@ -17,7 +18,6 @@ const DeepSeek: ModelProviderCard = { output: 2, }, releasedAt: '2024-09-05', - tokens: 65_536, }, ], checkModel: 'deepseek-chat', diff --git a/src/config/modelProviders/fireworksai.ts b/src/config/modelProviders/fireworksai.ts index f8301682bdf6..a60261d1769a 100644 --- a/src/config/modelProviders/fireworksai.ts +++ b/src/config/modelProviders/fireworksai.ts @@ -5,208 +5,210 @@ import { ModelProviderCard } from '@/types/llm'; const FireworksAI: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 8192, description: 'Fireworks 公司最新推出的 Firefunction-v2 是一款性能卓越的函数调用模型,基于 Llama-3 开发,并通过大量优化,特别适用于函数调用、对话及指令跟随等场景。', displayName: 'Firefunction V2', enabled: true, functionCall: true, id: 'accounts/fireworks/models/firefunction-v2', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Fireworks 开源函数调用模型,提供卓越的指令执行能力和开放可定制的特性。', displayName: 'Firefunction V1', functionCall: true, id: 'accounts/fireworks/models/firefunction-v1', - tokens: 32_768, }, { + contextWindowTokens: 4096, description: 'fireworks-ai/FireLLaVA-13b 是一款视觉语言模型,可以同时接收图像和文本输入,经过高质量数据训练,适合多模态任务。', displayName: 'FireLLaVA-13B', enabled: true, functionCall: false, id: 'accounts/fireworks/models/firellava-13b', - tokens: 4096, vision: true, }, { + contextWindowTokens: 131_072, description: 'Llama 3.2 1B 指令模型是Meta推出的一款轻量级多语言模型。该模型旨在提高效率,与更大型的模型相比,在延迟和成本方面提供了显著的改进。该模型的示例用例包括检索和摘要。', displayName: 'Llama 3.2 1B', enabled: true, id: 'accounts/fireworks/models/llama-v3p2-1b-instruct', - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.2 3B 指令模型是Meta推出的一款轻量级多语言模型。该模型旨在提高效率,与更大型的模型相比,在延迟和成本方面提供了显著的改进。该模型的示例用例包括查询和提示重写以及写作辅助。', displayName: 'Llama 3.2 3B', enabled: true, id: 'accounts/fireworks/models/llama-v3p2-3b-instruct', - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Meta的11B参数指令调整图像推理模型。该模型针对视觉识别、图像推理、图像描述和回答关于图像的一般性问题进行了优化。该模型能够理解视觉数据,如图表和图形,并通过生成文本描述图像细节来弥合视觉与语言之间的差距。', displayName: 'Llama 3.2 11B Vision', enabled: true, id: 'accounts/fireworks/models/llama-v3p2-11b-vision-instruct', - tokens: 131_072, vision: true, }, { + contextWindowTokens: 131_072, description: 'Meta的90B参数指令调整图像推理模型。该模型针对视觉识别、图像推理、图像描述和回答关于图像的一般性问题进行了优化。该模型能够理解视觉数据,如图表和图形,并通过生成文本描述图像细节来弥合视觉与语言之间的差距。', displayName: 'Llama 3.2 90B Vision', enabled: true, id: 'accounts/fireworks/models/llama-v3p2-90b-vision-instruct', - tokens: 131_072, vision: true, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 8B 指令模型,专为多语言对话优化,能够在常见行业基准上超越多数开源及闭源模型。', displayName: 'Llama 3.1 8B Instruct', enabled: true, functionCall: false, id: 'accounts/fireworks/models/llama-v3p1-8b-instruct', - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 70B 指令模型,提供卓越的自然语言理解和生成能力,是对话及分析任务的理想选择。', displayName: 'Llama 3.1 70B Instruct', enabled: true, functionCall: false, id: 'accounts/fireworks/models/llama-v3p1-70b-instruct', - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 405B 指令模型,具备超大规模参数,适合复杂任务和高负载场景下的指令跟随。', displayName: 'Llama 3.1 405B Instruct', enabled: true, functionCall: false, id: 'accounts/fireworks/models/llama-v3p1-405b-instruct', - tokens: 131_072, }, { + contextWindowTokens: 8192, description: 'Llama 3 8B 指令模型,优化用于对话及多语言任务,表现卓越且高效。', displayName: 'Llama 3 8B Instruct', functionCall: false, id: 'accounts/fireworks/models/llama-v3-8b-instruct', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 70B 指令模型,专为多语言对话和自然语言理解优化,性能优于多数竞争模型。', displayName: 'Llama 3 70B Instruct', functionCall: false, id: 'accounts/fireworks/models/llama-v3-70b-instruct', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 8B 指令模型(HF 版本),与官方实现结果一致,具备高度一致性和跨平台兼容性。', displayName: 'Llama 3 8B Instruct (HF version)', functionCall: false, id: 'accounts/fireworks/models/llama-v3-8b-instruct-hf', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 70B 指令模型(HF 版本),与官方实现结果保持一致,适合高质量的指令跟随任务。', displayName: 'Llama 3 70B Instruct (HF version)', functionCall: false, id: 'accounts/fireworks/models/llama-v3-70b-instruct-hf', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Mixtral MoE 8x7B 指令模型,多专家架构提供高效的指令跟随及执行。', displayName: 'Mixtral MoE 8x7B Instruct', enabled: true, functionCall: false, id: 'accounts/fireworks/models/mixtral-8x7b-instruct', - tokens: 32_768, }, { + contextWindowTokens: 65_536, description: 'Mixtral MoE 8x22B 指令模型,大规模参数和多专家架构,全方位支持复杂任务的高效处理。', displayName: 'Mixtral MoE 8x22B Instruct', enabled: true, functionCall: false, id: 'accounts/fireworks/models/mixtral-8x22b-instruct', - tokens: 65_536, }, { + contextWindowTokens: 32_768, description: 'Mixtral MoE 8x7B 指令模型(HF 版本),性能与官方实现一致,适合多种高效任务场景。', displayName: 'Mixtral MoE 8x7B Instruct (HF version)', functionCall: false, id: 'accounts/fireworks/models/mixtral-8x7b-instruct-hf', - tokens: 32_768, }, { + contextWindowTokens: 32_064, description: 'Phi-3-Vision-128K-Instruct 是一个轻量级的、最先进的开放多模态模型,它基于包括合成数据和经过筛选的公开网站在内的数据集构建,专注于非常高质量、推理密集型的数据,这些数据既包括文本也包括视觉。该模型属于 Phi-3 模型系列,其多模态版本支持 128K 的上下文长度(以标记为单位)。该模型经过严格的增强过程,结合了监督微调和直接偏好优化,以确保精确遵循指令和强大的安全措施。', displayName: 'Phi 3.5 Vision Instruct', enabled: true, functionCall: false, id: 'accounts/fireworks/models/phi-3-vision-128k-instruct', - tokens: 32_064, vision: true, }, { + contextWindowTokens: 32_768, description: 'QwQ模型是由 Qwen 团队开发的实验性研究模型,专注于增强 AI 推理能力。', displayName: 'QwQ 32B Preview', enabled: true, id: 'accounts/fireworks/models/qwen-qwq-32b-preview', - tokens: 32_768, }, { - description: 'Qwen2.5 是由阿里云 Qwen 团队开发的一系列仅包含解码器的语言模型。这些模型提供不同的大小,包括 0.5B、1.5B、3B、7B、14B、32B 和 72B,并且有基础版(base)和指令版(instruct)两种变体。', + contextWindowTokens: 32_768, + description: + 'Qwen2.5 是由阿里云 Qwen 团队开发的一系列仅包含解码器的语言模型。这些模型提供不同的大小,包括 0.5B、1.5B、3B、7B、14B、32B 和 72B,并且有基础版(base)和指令版(instruct)两种变体。', displayName: 'Qwen2.5 72B Instruct', enabled: true, functionCall: false, id: 'accounts/fireworks/models/qwen2p5-72b-instruct', - tokens: 32_768, }, { - description: 'Qwen2.5 Coder 32B Instruct 是阿里云发布的代码特定大语言模型系列的最新版本。该模型在 Qwen2.5 的基础上,通过 5.5 万亿个 tokens 的训练,显著提升了代码生成、推理和修复能力。它不仅增强了编码能力,还保持了数学和通用能力的优势。模型为代码智能体等实际应用提供了更全面的基础', + contextWindowTokens: 32_768, + description: + 'Qwen2.5 Coder 32B Instruct 是阿里云发布的代码特定大语言模型系列的最新版本。该模型在 Qwen2.5 的基础上,通过 5.5 万亿个 tokens 的训练,显著提升了代码生成、推理和修复能力。它不仅增强了编码能力,还保持了数学和通用能力的优势。模型为代码智能体等实际应用提供了更全面的基础', displayName: 'Qwen2.5 Coder 32B Instruct', enabled: false, id: 'accounts/fireworks/models/qwen2p5-coder-32b-instruct', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Yi-Large 模型,具备卓越的多语言处理能力,可用于各类语言生成和理解任务。', displayName: 'Yi-Large', enabled: true, functionCall: false, id: 'accounts/yi-01-ai/models/yi-large', - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'StarCoder 7B 模型,针对80多种编程语言训练,拥有出色的编程填充能力和语境理解。', displayName: 'StarCoder 7B', functionCall: false, id: 'accounts/fireworks/models/starcoder-7b', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'StarCoder 15.5B 模型,支持高级编程任务,多语言能力增强,适合复杂代码生成和理解。', displayName: 'StarCoder 15.5B', functionCall: false, id: 'accounts/fireworks/models/starcoder-16b', - tokens: 8192, }, { + contextWindowTokens: 4096, description: 'MythoMax L2 13B 模型,结合新颖的合并技术,擅长叙事和角色扮演。', displayName: 'MythoMax L2 13b', functionCall: false, id: 'accounts/fireworks/models/mythomax-l2-13b', - tokens: 4096, }, ], checkModel: 'accounts/fireworks/models/firefunction-v2', diff --git a/src/config/modelProviders/giteeai.ts b/src/config/modelProviders/giteeai.ts index 349a7c5e3edb..1c88def17aaa 100644 --- a/src/config/modelProviders/giteeai.ts +++ b/src/config/modelProviders/giteeai.ts @@ -4,13 +4,13 @@ import { ModelProviderCard } from '@/types/llm'; const GiteeAI: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 16_000, description: 'Qwen2.5-72B-Instruct 支持 16k 上下文, 生成长文本超过 8K 。支持 function call 与外部系统无缝交互,极大提升了灵活性和扩展性。模型知识明显增加,并且大大提高了编码和数学能力, 多语言支持超过 29 种', displayName: 'Qwen2.5 72B Instruct', enabled: true, functionCall: true, id: 'Qwen2.5-72B-Instruct', - tokens: 16_000, }, { description: @@ -42,18 +42,18 @@ const GiteeAI: ModelProviderCard = { id: 'Qwen2.5-14B-Instruct', }, { + contextWindowTokens: 6000, description: 'Qwen2 是 Qwen 模型的最新系列,支持 128k 上下文,对比当前最优的开源模型,Qwen2-72B 在自然语言理解、知识、代码、数学及多语言等多项能力上均显著超越当前领先的模型。', displayName: 'Qwen2 72B Instruct', id: 'Qwen2-72B-Instruct', - tokens: 6000, }, { + contextWindowTokens: 32_000, description: 'Qwen2 是 Qwen 模型的最新系列,能够超越同等规模的最优开源模型甚至更大规模的模型,Qwen2 7B 在多个评测上取得显著的优势,尤其是代码及中文理解上。', displayName: 'Qwen2 7B Instruct', id: 'Qwen2-7B-Instruct', - tokens: 32_000, }, { description: @@ -72,28 +72,28 @@ const GiteeAI: ModelProviderCard = { vision: true, }, { + contextWindowTokens: 32_000, description: 'GLM-4-9B-Chat 在语义、数学、推理、代码和知识等多方面均表现出较高性能。还具备网页浏览、代码执行、自定义工具调用和长文本推理。 支持包括日语,韩语,德语在内的 26 种语言。', displayName: 'GLM4 9B Chat', enabled: true, id: 'glm-4-9b-chat', - tokens: 32_000, }, { + contextWindowTokens: 4000, description: 'Yi-1.5-34B 在保持原系列模型优秀的通用语言能力的前提下,通过增量训练 5 千亿高质量 token,大幅提高了数学逻辑、代码能力。', displayName: 'Yi 34B Chat', enabled: true, id: 'Yi-34B-Chat', - tokens: 4000, }, { + contextWindowTokens: 8000, description: 'DeepSeek Coder 33B 是一个代码语言模型, 基于 2 万亿数据训练而成,其中 87% 为代码, 13% 为中英文语言。模型引入 16K 窗口大小和填空任务,提供项目级别的代码补全和片段填充功能。', displayName: 'DeepSeek Coder 33B Instruct', enabled: true, id: 'deepseek-coder-33B-instruct', - tokens: 8000, }, { description: @@ -103,12 +103,12 @@ const GiteeAI: ModelProviderCard = { id: 'code-raccoon-v1', }, { + contextWindowTokens: 40_000, description: 'CodeGeeX4-ALL-9B 是一个多语言代码生成模型,支持包括代码补全和生成、代码解释器、网络搜索、函数调用、仓库级代码问答在内的全面功能,覆盖软件开发的各种场景。是参数少于 10B 的顶尖代码生成模型。', displayName: 'CodeGeeX4 All 9B', enabled: true, id: 'codegeex4-all-9b', - tokens: 40_000, }, ], checkModel: 'Qwen2.5-72B-Instruct', diff --git a/src/config/modelProviders/github.ts b/src/config/modelProviders/github.ts index 0c8745d70162..ac871b4402bf 100644 --- a/src/config/modelProviders/github.ts +++ b/src/config/modelProviders/github.ts @@ -5,26 +5,28 @@ import { ModelProviderCard } from '@/types/llm'; const Github: ModelProviderCard = { chatModels: [ { - description: '专注于高级推理和解决复杂问题,包括数学和科学任务。非常适合需要深入上下文理解和代理工作流程的应用程序。', + contextWindowTokens: 200_000, + description: + '专注于高级推理和解决复杂问题,包括数学和科学任务。非常适合需要深入上下文理解和代理工作流程的应用程序。', displayName: 'OpenAI o1', enabled: true, functionCall: false, id: 'o1', maxOutput: 100_000, - tokens: 200_000, vision: true, }, { + contextWindowTokens: 128_000, description: '比 o1-preview 更小、更快,成本低80%,在代码生成和小上下文操作方面表现良好。', displayName: 'OpenAI o1-mini', enabled: true, functionCall: false, id: 'o1-mini', maxOutput: 65_536, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '专注于高级推理和解决复杂问题,包括数学和科学任务。非常适合需要深度上下文理解和自主工作流程的应用。', displayName: 'OpenAI o1-preview', @@ -32,195 +34,194 @@ const Github: ModelProviderCard = { functionCall: false, id: 'o1-preview', maxOutput: 32_768, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '一种经济高效的AI解决方案,适用于多种文本和图像任务。', displayName: 'OpenAI GPT-4o mini', enabled: true, functionCall: true, id: 'gpt-4o-mini', maxOutput: 4096, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'OpenAI GPT-4系列中最先进的多模态模型,可以处理文本和图像输入。', displayName: 'OpenAI GPT-4o', enabled: true, functionCall: true, id: 'gpt-4o', maxOutput: 4096, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 262_144, description: '一个52B参数(12B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。', displayName: 'AI21 Jamba 1.5 Mini', functionCall: true, id: 'ai21-jamba-1.5-mini', maxOutput: 4096, - tokens: 262_144, }, { + contextWindowTokens: 262_144, description: '一个398B参数(94B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。', displayName: 'AI21 Jamba 1.5 Large', functionCall: true, id: 'ai21-jamba-1.5-large', maxOutput: 4096, - tokens: 262_144, }, { + contextWindowTokens: 131_072, description: 'Command R是一个可扩展的生成模型,旨在针对RAG和工具使用,使企业能够实现生产级AI。', displayName: 'Cohere Command R', id: 'cohere-command-r', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Command R+是一个最先进的RAG优化模型,旨在应对企业级工作负载。', displayName: 'Cohere Command R+', id: 'cohere-command-r-plus', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Mistral Nemo是一种尖端的语言模型(LLM),在其尺寸类别中拥有最先进的推理、世界知识和编码能力。', displayName: 'Mistral Nemo', id: 'mistral-nemo', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Mistral Small可用于任何需要高效率和低延迟的基于语言的任务。', displayName: 'Mistral Small', id: 'mistral-small', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Mistral的旗舰模型,适合需要大规模推理能力或高度专业化的复杂任务(合成文本生成、代码生成、RAG或代理)。', displayName: 'Mistral Large', id: 'mistral-large', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '在高分辨率图像上表现出色的图像推理能力,适用于视觉理解应用。', displayName: 'Llama 3.2 11B Vision', id: 'llama-3.2-11b-vision-instruct', maxOutput: 4096, - tokens: 131_072, vision: true, }, { + contextWindowTokens: 131_072, description: '适用于视觉理解代理应用的高级图像推理能力。', displayName: 'Llama 3.2 90B Vision', id: 'llama-3.2-90b-vision-instruct', maxOutput: 4096, - tokens: 131_072, vision: true, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。', displayName: 'Meta Llama 3.1 8B', id: 'meta-llama-3.1-8b-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。', displayName: 'Meta Llama 3.1 70B', id: 'meta-llama-3.1-70b-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。', displayName: 'Meta Llama 3.1 405B', id: 'meta-llama-3.1-405b-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 8192, description: '一个多功能的80亿参数模型,针对对话和文本生成任务进行了优化。', displayName: 'Meta Llama 3 8B', id: 'meta-llama-3-8b-instruct', maxOutput: 4096, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '一个强大的700亿参数模型,在推理、编码和广泛的语言应用方面表现出色。', displayName: 'Meta Llama 3 70B', id: 'meta-llama-3-70b-instruct', maxOutput: 4096, - tokens: 8192, }, { + contextWindowTokens: 131_072, description: 'Phi-3-mini模型的更新版。', displayName: 'Phi-3.5-mini 128K', id: 'Phi-3.5-mini-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Phi-3-vision模型的更新版。', displayName: 'Phi-3.5-vision 128K', id: 'Phi-3.5-vision-instrust', maxOutput: 4096, - tokens: 131_072, vision: true, }, { + contextWindowTokens: 4096, description: 'Phi-3家族中最小的成员,针对质量和低延迟进行了优化。', displayName: 'Phi-3-mini 4K', id: 'Phi-3-mini-4k-instruct', maxOutput: 4096, - tokens: 4096, }, { + contextWindowTokens: 131_072, description: '相同的Phi-3-mini模型,但具有更大的上下文大小,适用于RAG或少量提示。', displayName: 'Phi-3-mini 128K', id: 'Phi-3-mini-128k-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 8192, description: '一个70亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。', displayName: 'Phi-3-small 8K', id: 'Phi-3-small-8k-instruct', maxOutput: 4096, - tokens: 8192, }, { + contextWindowTokens: 131_072, description: '相同的Phi-3-small模型,但具有更大的上下文大小,适用于RAG或少量提示。', displayName: 'Phi-3-small 128K', id: 'Phi-3-small-128k-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 4096, description: '一个140亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。', displayName: 'Phi-3-medium 4K', id: 'Phi-3-medium-4k-instruct', maxOutput: 4096, - tokens: 4096, }, { + contextWindowTokens: 131_072, description: '相同的Phi-3-medium模型,但具有更大的上下文大小,适用于RAG或少量提示。', displayName: 'Phi-3-medium 128K', id: 'Phi-3-medium-128k-instruct', maxOutput: 4096, - tokens: 131_072, }, ], checkModel: 'Phi-3-mini-4k-instruct', diff --git a/src/config/modelProviders/google.ts b/src/config/modelProviders/google.ts index e1faab534a99..755122c73d39 100644 --- a/src/config/modelProviders/google.ts +++ b/src/config/modelProviders/google.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const Google: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 32_767 + 8192, description: 'Gemini 2.0 Flash Exp 是 Google 最新的实验性多模态AI模型,拥有下一代特性,卓越的速度,原生工具调用以及多模态生成。', displayName: 'Gemini 2.0 Flash Thinking Experimental 1219', @@ -17,10 +18,10 @@ const Google: ModelProviderCard = { output: 0, }, releasedAt: '2024-12-19', - tokens: 32_767 + 8192, vision: true, }, { + contextWindowTokens: 1_048_576 + 8192, description: 'Gemini 2.0 Flash Exp 是 Google 最新的实验性多模态AI模型,拥有下一代特性,卓越的速度,原生工具调用以及多模态生成。', displayName: 'Gemini 2.0 Flash Experimental', @@ -34,10 +35,10 @@ const Google: ModelProviderCard = { output: 0, }, releasedAt: '2024-12-11', - tokens: 1_048_576 + 8192, vision: true, }, { + contextWindowTokens: 2_097_152 + 8192, description: 'Gemini Exp 1206 是 Google 最新的实验性多模态AI模型,与历史版本相比有一定的质量提升。', displayName: 'Gemini Experimental 1206', @@ -51,10 +52,10 @@ const Google: ModelProviderCard = { output: 0, }, releasedAt: '2024-12-06', - tokens: 2_097_152 + 8192, vision: true, }, { + contextWindowTokens: 32_767 + 8192, description: 'Gemini Exp 1121 是 Google 的实验性多模态AI模型,拥有改进的编码、推理和视觉能力。', displayName: 'Gemini Experimental 1121', @@ -67,12 +68,11 @@ const Google: ModelProviderCard = { output: 0, }, releasedAt: '2024-11-21', - tokens: 32_767 + 8192, vision: true, }, { - description: - 'Gemini Exp 1114 是 Google 的实验性多模态AI模型,对输出质量有一定改进。', + contextWindowTokens: 32_767 + 8192, + description: 'Gemini Exp 1114 是 Google 的实验性多模态AI模型,对输出质量有一定改进。', displayName: 'Gemini Experimental 1114', functionCall: true, id: 'gemini-exp-1114', @@ -83,10 +83,10 @@ const Google: ModelProviderCard = { output: 0, }, releasedAt: '2024-11-14', - tokens: 32_767 + 8192, vision: true, }, { + contextWindowTokens: 32_767 + 8192, description: 'LearnLM 是一个实验性的、特定于任务的语言模型,经过训练以符合学习科学原则,可在教学和学习场景中遵循系统指令,充当专家导师等。', displayName: 'LearnLM 1.5 Pro Experimental', @@ -99,10 +99,10 @@ const Google: ModelProviderCard = { output: 0, }, releasedAt: '2024-11-19', - tokens: 32_767 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 是 Google 最新的多模态AI模型,具备快速处理能力,支持文本、图像和视频输入,适用于多种任务的高效扩展。', displayName: 'Gemini 1.5 Flash', @@ -115,10 +115,10 @@ const Google: ModelProviderCard = { input: 0.075, output: 0.3, }, - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 002 是一款高效的多模态模型,支持广泛应用的扩展。', displayName: 'Gemini 1.5 Flash 002', enabled: true, @@ -131,10 +131,10 @@ const Google: ModelProviderCard = { output: 0.3, }, releasedAt: '2024-09-25', - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 001 是一款高效的多模态模型,支持广泛应用的扩展。', displayName: 'Gemini 1.5 Flash 001', functionCall: true, @@ -145,10 +145,10 @@ const Google: ModelProviderCard = { input: 0.075, output: 0.3, }, - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 支持高达200万个tokens,是中型多模态模型的理想选择,适用于复杂任务的多方面支持。', displayName: 'Gemini 1.5 Pro', @@ -162,10 +162,10 @@ const Google: ModelProviderCard = { output: 10.5, }, releasedAt: '2024-02-15', - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 002 是最新的生产就绪模型,提供更高质量的输出,特别在数学、长上下文和视觉任务方面有显著提升。', displayName: 'Gemini 1.5 Pro 002', @@ -179,10 +179,10 @@ const Google: ModelProviderCard = { output: 2.5, }, releasedAt: '2024-09-24', - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 001 是可扩展的多模态AI解决方案,支持广泛的复杂任务。', displayName: 'Gemini 1.5 Pro 001', functionCall: true, @@ -194,10 +194,10 @@ const Google: ModelProviderCard = { output: 10.5, }, releasedAt: '2024-02-15', - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 8B 是一款高效的多模态模型,支持广泛应用的扩展。', displayName: 'Gemini 1.5 Flash 8B', enabled: true, @@ -210,10 +210,10 @@ const Google: ModelProviderCard = { output: 0.3, }, releasedAt: '2024-10-03', - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 8B 0924 是最新的实验性模型,在文本和多模态用例中都有显著的性能提升。', displayName: 'Gemini 1.5 Flash 8B 0924', @@ -226,11 +226,11 @@ const Google: ModelProviderCard = { output: 0.3, }, releasedAt: '2024-09-24', - tokens: 1_000_000 + 8192, vision: true, }, // Gemini 1.0 Pro will be removed on 2025.02.15 { + contextWindowTokens: 30_720 + 2048, description: 'Gemini 1.0 Pro 是Google的高性能AI模型,专为广泛任务扩展而设计。', displayName: 'Gemini 1.0 Pro', id: 'gemini-1.0-pro-latest', @@ -240,9 +240,9 @@ const Google: ModelProviderCard = { output: 1.5, }, releasedAt: '2023-12-06', - tokens: 30_720 + 2048, }, { + contextWindowTokens: 30_720 + 2048, description: 'Gemini 1.0 Pro 001 (Tuning) 提供稳定并可调优的性能,是复杂任务解决方案的理想选择。', displayName: 'Gemini 1.0 Pro 001 (Tuning)', @@ -254,9 +254,9 @@ const Google: ModelProviderCard = { output: 1.5, }, releasedAt: '2023-12-06', - tokens: 30_720 + 2048, }, { + contextWindowTokens: 30_720 + 2048, description: 'Gemini 1.0 Pro 002 (Tuning) 提供出色的多模态支持,专注于复杂任务的有效解决。', displayName: 'Gemini 1.0 Pro 002 (Tuning)', id: 'gemini-1.0-pro-002', @@ -266,7 +266,6 @@ const Google: ModelProviderCard = { output: 1.5, }, releasedAt: '2023-12-06', - tokens: 30_720 + 2048, }, ], checkModel: 'gemini-1.5-flash-latest', diff --git a/src/config/modelProviders/groq.ts b/src/config/modelProviders/groq.ts index a7eee3c8f7e9..b0090f02d881 100644 --- a/src/config/modelProviders/groq.ts +++ b/src/config/modelProviders/groq.ts @@ -5,6 +5,7 @@ const Groq: ModelProviderCard = { chatModels: [ // TODO: During preview launch, Groq is limiting 3.2 models to max_tokens of 8k. { + contextWindowTokens: 131_072, description: 'Meta Llama 3.3 多语言大语言模型 ( LLM ) 是 70B(文本输入/文本输出)中的预训练和指令调整生成模型。 Llama 3.3 指令调整的纯文本模型针对多语言对话用例进行了优化,并且在常见行业基准上优于许多可用的开源和封闭式聊天模型。', displayName: 'Llama 3.3 70B', @@ -16,9 +17,9 @@ const Groq: ModelProviderCard = { input: 0.05, output: 0.08, }, - tokens: 131_072, }, { + contextWindowTokens: 8192, description: 'Llama 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 11B Vision', @@ -29,10 +30,10 @@ const Groq: ModelProviderCard = { input: 0.05, output: 0.08, }, - tokens: 8192, vision: true, }, { + contextWindowTokens: 8192, description: 'Llama 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 90B Vision', @@ -43,10 +44,10 @@ const Groq: ModelProviderCard = { input: 0.59, output: 0.79, }, - tokens: 8192, vision: true, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 8B 是一款高效能模型,提供了快速的文本生成能力,非常适合需要大规模效率和成本效益的应用场景。', displayName: 'Llama 3.1 8B', @@ -58,9 +59,9 @@ const Groq: ModelProviderCard = { input: 0.05, output: 0.08, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 70B 提供更强大的AI推理能力,适合复杂应用,支持超多的计算处理并保证高效和准确率。', displayName: 'Llama 3.1 70B', @@ -72,7 +73,6 @@ const Groq: ModelProviderCard = { input: 0.59, output: 0.79, }, - tokens: 131_072, }, { description: 'Llama 3.1 405B 指令模型,具备超大规模参数,适合复杂任务和高负载场景下的指令跟随。', @@ -83,6 +83,7 @@ const Groq: ModelProviderCard = { tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 Groq 8B Tool Use 是针对高效工具使用优化的模型,支持快速并行计算。', displayName: 'Llama 3 Groq 8B Tool Use (Preview)', functionCall: true, @@ -91,9 +92,9 @@ const Groq: ModelProviderCard = { input: 0.19, output: 0.19, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 Groq 70B Tool Use 提供强大的工具调用能力,支持复杂任务的高效处理。', displayName: 'Llama 3 Groq 70B Tool Use (Preview)', functionCall: true, @@ -102,9 +103,9 @@ const Groq: ModelProviderCard = { input: 0.89, output: 0.89, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Meta Llama 3 8B 带来优质的推理效能,适合多场景应用需求。', displayName: 'Meta Llama 3 8B', functionCall: true, @@ -113,9 +114,9 @@ const Groq: ModelProviderCard = { input: 0.05, output: 0.08, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Meta Llama 3 70B 提供无与伦比的复杂性处理能力,为高要求项目量身定制。', displayName: 'Meta Llama 3 70B', functionCall: true, @@ -124,9 +125,9 @@ const Groq: ModelProviderCard = { input: 0.59, output: 0.79, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma 2 9B 是一款优化用于特定任务和工具整合的模型。', displayName: 'Gemma 2 9B', enabled: true, @@ -136,9 +137,9 @@ const Groq: ModelProviderCard = { input: 0.2, output: 0.2, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma 7B 适合中小规模任务处理,兼具成本效益。', displayName: 'Gemma 7B', functionCall: true, @@ -147,9 +148,9 @@ const Groq: ModelProviderCard = { input: 0.07, output: 0.07, }, - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Mixtral 8x7B 提供高容错的并行计算能力,适合复杂任务。', displayName: 'Mixtral 8x7B', functionCall: true, @@ -158,13 +159,12 @@ const Groq: ModelProviderCard = { input: 0.24, output: 0.24, }, - tokens: 32_768, }, { + contextWindowTokens: 4096, description: 'LLaVA 1.5 7B 提供视觉处理能力融合,通过视觉信息输入生成复杂输出。', displayName: 'LLaVA 1.5 7B', id: 'llava-v1.5-7b-4096-preview', - tokens: 4096, vision: true, }, ], diff --git a/src/config/modelProviders/higress.ts b/src/config/modelProviders/higress.ts index a3b208d31c27..5269a6fb0eea 100644 --- a/src/config/modelProviders/higress.ts +++ b/src/config/modelProviders/higress.ts @@ -4,6 +4,7 @@ const Higress: ModelProviderCard = { chatModels: [ //qwen { + contextWindowTokens: 131_072, description: '通义千问超大规模语言模型,支持中文、英文等不同语言输入。', displayName: 'Qwen Turbo', enabled: true, @@ -14,9 +15,9 @@ const Higress: ModelProviderCard = { input: 0.3, output: 0.6, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问超大规模语言模型增强版,支持中文、英文等不同语言输入。', displayName: 'Qwen Plus', enabled: true, @@ -27,9 +28,9 @@ const Higress: ModelProviderCard = { input: 0.8, output: 2, }, - tokens: 131_072, }, { + contextWindowTokens: 32_768, description: '通义千问千亿级别超大规模语言模型,支持中文、英文等不同语言输入,当前通义千问2.5产品版本背后的API模型。', displayName: 'Qwen Max', @@ -41,9 +42,9 @@ const Higress: ModelProviderCard = { input: 20, output: 60, }, - tokens: 32_768, }, { + contextWindowTokens: 1_000_000, description: '通义千问超大规模语言模型,支持长文本上下文,以及基于长文档、多文档等多个场景的对话功能。', displayName: 'Qwen Long', @@ -53,10 +54,10 @@ const Higress: ModelProviderCard = { input: 0.5, output: 2, }, - tokens: 1_000_000, }, //后面几个qwen未知支持 { + contextWindowTokens: 32_000, description: '通义千问大规模视觉语言模型增强版。大幅提升细节识别能力和文字识别能力,支持超百万像素分辨率和任意长宽比规格的图像。', displayName: 'Qwen VL Plus', @@ -67,10 +68,10 @@ const Higress: ModelProviderCard = { input: 8, output: 8, }, - tokens: 32_000, vision: true, }, { + contextWindowTokens: 32_000, description: '通义千问超大规模视觉语言模型。相比增强版,再次提升视觉推理能力和指令遵循能力,提供更高的视觉感知和认知水平。', displayName: 'Qwen VL Max', @@ -81,10 +82,10 @@ const Higress: ModelProviderCard = { input: 20, output: 20, }, - tokens: 32_000, vision: true, }, { + contextWindowTokens: 4096, description: '通义千问数学模型是专门用于数学解题的语言模型。', displayName: 'Qwen Math Turbo', id: 'qwen-math-turbo-latest', @@ -93,9 +94,9 @@ const Higress: ModelProviderCard = { input: 2, output: 6, }, - tokens: 4096, }, { + contextWindowTokens: 4096, description: '通义千问数学模型是专门用于数学解题的语言模型。', displayName: 'Qwen Math Plus', id: 'qwen-math-plus-latest', @@ -104,9 +105,9 @@ const Higress: ModelProviderCard = { input: 4, output: 12, }, - tokens: 4096, }, { + contextWindowTokens: 131_072, description: '通义千问代码模型。', displayName: 'Qwen Coder Turbo', id: 'qwen-coder-turbo-latest', @@ -115,9 +116,9 @@ const Higress: ModelProviderCard = { input: 2, output: 6, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问2.5对外开源的7B规模的模型。', displayName: 'Qwen2.5 7B', functionCall: true, @@ -127,9 +128,9 @@ const Higress: ModelProviderCard = { input: 1, output: 2, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问2.5对外开源的14B规模的模型。', displayName: 'Qwen2.5 14B', functionCall: true, @@ -139,9 +140,9 @@ const Higress: ModelProviderCard = { input: 2, output: 6, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问2.5对外开源的32B规模的模型。', displayName: 'Qwen2.5 32B', functionCall: true, @@ -151,9 +152,9 @@ const Higress: ModelProviderCard = { input: 3.5, output: 7, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问2.5对外开源的72B规模的模型。', displayName: 'Qwen2.5 72B', functionCall: true, @@ -163,9 +164,9 @@ const Higress: ModelProviderCard = { input: 4, output: 12, }, - tokens: 131_072, }, { + contextWindowTokens: 4096, description: 'Qwen-Math 模型具有强大的数学解题能力。', displayName: 'Qwen2.5 Math 1.5B', id: 'qwen2.5-math-1.5b-instruct', @@ -174,9 +175,9 @@ const Higress: ModelProviderCard = { input: 0, output: 0, }, - tokens: 4096, }, { + contextWindowTokens: 4096, description: 'Qwen-Math 模型具有强大的数学解题能力。', displayName: 'Qwen2.5 Math 7B', id: 'qwen2.5-math-7b-instruct', @@ -185,9 +186,9 @@ const Higress: ModelProviderCard = { input: 1, output: 2, }, - tokens: 4096, }, { + contextWindowTokens: 4096, description: 'Qwen-Math 模型具有强大的数学解题能力。', displayName: 'Qwen2.5 Math 72B', id: 'qwen2.5-math-72b-instruct', @@ -196,9 +197,9 @@ const Higress: ModelProviderCard = { input: 4, output: 12, }, - tokens: 4096, }, { + contextWindowTokens: 131_072, description: '通义千问代码模型开源版。', displayName: 'Qwen2.5 Coder 1.5B', id: 'qwen2.5-coder-1.5b-instruct', @@ -207,9 +208,9 @@ const Higress: ModelProviderCard = { input: 0, output: 0, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问代码模型开源版。', displayName: 'Qwen2.5 Coder 7B', id: 'qwen2.5-coder-7b-instruct', @@ -218,9 +219,9 @@ const Higress: ModelProviderCard = { input: 1, output: 2, }, - tokens: 131_072, }, { + contextWindowTokens: 8000, description: '以 Qwen-7B 语言模型初始化,添加图像模型,图像输入分辨率为448的预训练模型。', displayName: 'Qwen VL', id: 'qwen-vl-v1', @@ -229,10 +230,10 @@ const Higress: ModelProviderCard = { input: 0, output: 0, }, - tokens: 8000, vision: true, }, { + contextWindowTokens: 8000, description: '通义千问VL支持灵活的交互方式,包括多图、多轮问答、创作等能力的模型。', displayName: 'Qwen VL Chat', id: 'qwen-vl-chat-v1', @@ -241,40 +242,40 @@ const Higress: ModelProviderCard = { input: 0, output: 0, }, - tokens: 8000, vision: true, }, //moonshot { + contextWindowTokens: 8192, description: 'Moonshot V1 8K 专为生成短文本任务设计,具有高效的处理性能,能够处理8,192个tokens,非常适合简短对话、速记和快速内容生成。', displayName: 'Moonshot V1 8K', enabled: true, functionCall: true, id: 'moonshot-v1-8k', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Moonshot V1 32K 提供中等长度的上下文处理能力,能够处理32,768个tokens,特别适合生成各种长文档和复杂对话,应用于内容创作、报告生成和对话系统等领域。', displayName: 'Moonshot V1 32K', enabled: true, functionCall: true, id: 'moonshot-v1-32k', - tokens: 32_768, }, { + contextWindowTokens: 128_000, description: 'Moonshot V1 128K 是一款拥有超长上下文处理能力的模型,适用于生成超长文本,满足复杂的生成任务需求,能够处理多达128,000个tokens的内容,非常适合科研、学术和大型文档生成等应用场景。', displayName: 'Moonshot V1 128K', enabled: true, functionCall: true, id: 'moonshot-v1-128k', - tokens: 128_000, }, //百川智能 { + contextWindowTokens: 32_768, description: '模型能力国内第一,在知识百科、长文本、生成创作等中文任务上超越国外主流模型。还具备行业领先的多模态能力,多项权威评测基准表现优异。', displayName: 'Baichuan 4', @@ -287,7 +288,6 @@ const Higress: ModelProviderCard = { input: 100, output: 100, }, - tokens: 32_768, }, { description: '', @@ -318,6 +318,7 @@ const Higress: ModelProviderCard = { // tokens: 32_768, }, { + contextWindowTokens: 32_768, description: '针对企业高频场景优化,效果大幅提升,高性价比。相对于Baichuan2模型,内容创作提升20%,知识问答提升17%, 角色扮演能力提升40%。整体效果比GPT3.5更优。', displayName: 'Baichuan 3 Turbo', @@ -330,9 +331,9 @@ const Higress: ModelProviderCard = { input: 12, output: 12, }, - tokens: 32_768, }, { + contextWindowTokens: 128_000, description: '具备 128K 超长上下文窗口,针对企业高频场景优化,效果大幅提升,高性价比。相对于Baichuan2模型,内容创作提升20%,知识问答提升17%, 角色扮演能力提升40%。整体效果比GPT3.5更优。', displayName: 'Baichuan 3 Turbo 128k', @@ -344,9 +345,9 @@ const Higress: ModelProviderCard = { input: 24, output: 24, }, - tokens: 128_000, }, { + contextWindowTokens: 32_768, description: '采用搜索增强技术实现大模型与领域知识、全网知识的全面链接。支持PDF、Word等多种文档上传及网址输入,信息获取及时、全面,输出结果准确、专业。', displayName: 'Baichuan 2 Turbo', @@ -357,10 +358,10 @@ const Higress: ModelProviderCard = { input: 8, output: 8, }, - tokens: 32_768, }, //零一万物 { + contextWindowTokens: 16_384, description: '最新高性能模型,保证高质量输出同时,推理速度大幅提升。', displayName: 'Yi Lightning', enabled: true, @@ -370,9 +371,9 @@ const Higress: ModelProviderCard = { input: 0.99, output: 0.99, }, - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: '小而精悍,轻量极速模型。提供强化数学运算和代码编写能力。', displayName: 'Yi Spark', enabled: true, @@ -382,9 +383,9 @@ const Higress: ModelProviderCard = { input: 1, output: 1, }, - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: '中型尺寸模型升级微调,能力均衡,性价比高。深度优化指令遵循能力。', displayName: 'Yi Medium', enabled: true, @@ -394,9 +395,9 @@ const Higress: ModelProviderCard = { input: 2.5, output: 2.5, }, - tokens: 16_384, }, { + contextWindowTokens: 200_000, description: '200K 超长上下文窗口,提供长文本深度理解和生成能力。', displayName: 'Yi Medium 200K', enabled: true, @@ -406,9 +407,9 @@ const Higress: ModelProviderCard = { input: 12, output: 12, }, - tokens: 200_000, }, { + contextWindowTokens: 16_384, description: '超高性价比、卓越性能。根据性能和推理速度、成本,进行平衡性高精度调优。', displayName: 'Yi Large Turbo', enabled: true, @@ -418,9 +419,9 @@ const Higress: ModelProviderCard = { input: 12, output: 12, }, - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: '基于 yi-large 超强模型的高阶服务,结合检索与生成技术提供精准答案,实时全网检索信息服务。', displayName: 'Yi Large RAG', @@ -431,9 +432,9 @@ const Higress: ModelProviderCard = { input: 25, output: 25, }, - tokens: 16_384, }, { + contextWindowTokens: 32_768, description: '在 yi-large 模型的基础上支持并强化了工具调用的能力,适用于各种需要搭建 agent 或 workflow 的业务场景。', displayName: 'Yi Large FC', @@ -445,9 +446,9 @@ const Higress: ModelProviderCard = { input: 20, output: 20, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: '全新千亿参数模型,提供超强问答及文本生成能力。', displayName: 'Yi Large', id: 'yi-large', @@ -456,9 +457,9 @@ const Higress: ModelProviderCard = { input: 20, output: 20, }, - tokens: 32_768, }, { + contextWindowTokens: 16_384, description: '复杂视觉任务模型,提供高性能图片理解、分析能力。', displayName: 'Yi Vision', enabled: true, @@ -468,10 +469,10 @@ const Higress: ModelProviderCard = { input: 6, output: 6, }, - tokens: 16_384, vision: true, }, { + contextWindowTokens: 16_384, description: '初期版本,推荐使用 yi-large(新版本)。', displayName: 'Yi Large Preview', id: 'yi-large-preview', @@ -480,9 +481,9 @@ const Higress: ModelProviderCard = { input: 20, output: 20, }, - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: '轻量化版本,推荐使用 yi-lightning。', displayName: 'Yi Lightning Lite', id: 'yi-lightning-lite', @@ -491,10 +492,10 @@ const Higress: ModelProviderCard = { input: 0.99, output: 0.99, }, - tokens: 16_384, }, //智谱AI { + contextWindowTokens: 128_000, description: 'GLM-4-Flash 是处理简单任务的理想选择,速度最快且免费。', displayName: 'GLM-4-Flash', enabled: true, @@ -505,9 +506,9 @@ const Higress: ModelProviderCard = { input: 0, output: 0, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4-FlashX 是Flash的增强版本,超快推理速度。', displayName: 'GLM-4-FlashX', enabled: true, @@ -518,9 +519,9 @@ const Higress: ModelProviderCard = { input: 0.1, output: 0.1, }, - tokens: 128_000, }, { + contextWindowTokens: 1_024_000, description: 'GLM-4-Long 支持超长文本输入,适合记忆型任务与大规模文档处理。', displayName: 'GLM-4-Long', functionCall: true, @@ -530,9 +531,9 @@ const Higress: ModelProviderCard = { input: 1, output: 1, }, - tokens: 1_024_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4-Air 是性价比高的版本,性能接近GLM-4,提供快速度和实惠的价格。', displayName: 'GLM-4-Air', enabled: true, @@ -543,9 +544,9 @@ const Higress: ModelProviderCard = { input: 1, output: 1, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: 'GLM-4-AirX 提供 GLM-4-Air 的高效版本,推理速度可达其2.6倍。', displayName: 'GLM-4-AirX', enabled: true, @@ -556,9 +557,9 @@ const Higress: ModelProviderCard = { input: 10, output: 10, }, - tokens: 8192, }, { + contextWindowTokens: 128_000, description: 'GLM-4-AllTools 是一个多功能智能体模型,优化以支持复杂指令规划与工具调用,如网络浏览、代码解释和文本生成,适用于多任务执行。', displayName: 'GLM-4-AllTools', @@ -569,9 +570,9 @@ const Higress: ModelProviderCard = { input: 100, output: 100, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4-Plus 作为高智能旗舰,具备强大的处理长文本和复杂任务的能力,性能全面提升。', displayName: 'GLM-4-Plus', @@ -583,9 +584,9 @@ const Higress: ModelProviderCard = { input: 50, output: 50, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4-0520 是最新模型版本,专为高度复杂和多样化任务设计,表现卓越。', displayName: 'GLM-4-0520', functionCall: true, @@ -595,9 +596,9 @@ const Higress: ModelProviderCard = { input: 100, output: 100, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4 是发布于2024年1月的旧旗舰版本,目前已被更强的 GLM-4-0520 取代。', displayName: 'GLM-4', functionCall: true, @@ -607,9 +608,9 @@ const Higress: ModelProviderCard = { input: 100, output: 100, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: 'GLM-4V-Plus 具备对视频内容及多图片的理解能力,适合多模态任务。', displayName: 'GLM-4V-Plus', enabled: true, @@ -619,10 +620,10 @@ const Higress: ModelProviderCard = { input: 10, output: 10, }, - tokens: 8192, vision: true, }, { + contextWindowTokens: 2048, description: 'GLM-4V 提供强大的图像理解与推理能力,支持多种视觉任务。', displayName: 'GLM-4V', id: 'glm-4v', @@ -631,10 +632,10 @@ const Higress: ModelProviderCard = { input: 50, output: 50, }, - tokens: 2048, vision: true, }, { + contextWindowTokens: 4096, description: 'CharGLM-3 专为角色扮演与情感陪伴设计,支持超长多轮记忆与个性化对话,应用广泛。', displayName: 'CharGLM-3', id: 'charglm-3', @@ -643,9 +644,9 @@ const Higress: ModelProviderCard = { input: 15, output: 15, }, - tokens: 4096, }, { + contextWindowTokens: 8192, description: 'Emohaa 是心理模型,具备专业咨询能力,帮助用户理解情感问题。', displayName: 'Emohaa', id: 'emohaa', @@ -654,10 +655,10 @@ const Higress: ModelProviderCard = { input: 15, output: 15, }, - tokens: 8192, }, //360智脑 { + contextWindowTokens: 8192, description: '360GPT2 Pro 是 360 公司推出的高级自然语言处理模型,具备卓越的文本生成和理解能力,尤其在生成与创作领域表现出色,能够处理复杂的语言转换和角色演绎任务。', displayName: '360GPT2 Pro', @@ -669,9 +670,9 @@ const Higress: ModelProviderCard = { input: 5, output: 5, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '360GPT Pro 作为 360 AI 模型系列的重要成员,以高效的文本处理能力满足多样化的自然语言应用场景,支持长文本理解和多轮对话等功能。', displayName: '360GPT Pro', @@ -684,9 +685,9 @@ const Higress: ModelProviderCard = { input: 5, output: 5, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '360GPT Turbo 提供强大的计算和对话能力,具备出色的语义理解和生成效率,是企业和开发者理想的智能助理解决方案。', displayName: '360GPT Turbo', @@ -698,9 +699,9 @@ const Higress: ModelProviderCard = { input: 2, output: 2, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '360GPT Turbo Responsibility 8K 强调语义安全和责任导向,专为对内容安全有高度要求的应用场景设计,确保用户体验的准确性与稳健性。', displayName: '360GPT Turbo Responsibility 8K', @@ -712,10 +713,10 @@ const Higress: ModelProviderCard = { input: 2, output: 2, }, - tokens: 8192, }, //文心一言 { + contextWindowTokens: 8192, description: '百度自研的旗舰级大规模⼤语⾔模型,覆盖海量中英文语料,具有强大的通用能力,可满足绝大部分对话问答、创作生成、插件应用场景要求;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 3.5 8K', @@ -726,9 +727,9 @@ const Higress: ModelProviderCard = { input: 0.8, output: 2, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级大规模⼤语⾔模型,覆盖海量中英文语料,具有强大的通用能力,可满足绝大部分对话问答、创作生成、插件应用场景要求;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 3.5 8K Preview', @@ -738,9 +739,9 @@ const Higress: ModelProviderCard = { input: 0.8, output: 2, }, - tokens: 8192, }, { + contextWindowTokens: 128_000, description: '百度自研的旗舰级大规模⼤语⾔模型,覆盖海量中英文语料,具有强大的通用能力,可满足绝大部分对话问答、创作生成、插件应用场景要求;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 3.5 128K', @@ -751,9 +752,9 @@ const Higress: ModelProviderCard = { input: 0.8, output: 2, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级超大规模⼤语⾔模型,相较ERNIE 3.5实现了模型能力全面升级,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 4.0 8K', @@ -764,9 +765,9 @@ const Higress: ModelProviderCard = { input: 30, output: 90, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级超大规模⼤语⾔模型,相较ERNIE 3.5实现了模型能力全面升级,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 4.0 8K Preview', @@ -776,9 +777,9 @@ const Higress: ModelProviderCard = { input: 30, output: 90, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级超大规模⼤语⾔模型,综合效果表现出色,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。相较于ERNIE 4.0在性能表现上更优秀', displayName: 'ERNIE 4.0 Turbo 8K', @@ -789,9 +790,9 @@ const Higress: ModelProviderCard = { input: 20, output: 60, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级超大规模⼤语⾔模型,综合效果表现出色,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。相较于ERNIE 4.0在性能表现上更优秀', displayName: 'ERNIE 4.0 Turbo 8K Preview', @@ -801,9 +802,9 @@ const Higress: ModelProviderCard = { input: 20, output: 60, }, - tokens: 8192, }, { + contextWindowTokens: 128_000, description: '百度自研的轻量级大语言模型,兼顾优异的模型效果与推理性能,效果比ERNIE Lite更优,适合低算力AI加速卡推理使用。', displayName: 'ERNIE Lite Pro 128K', @@ -814,9 +815,9 @@ const Higress: ModelProviderCard = { input: 0.2, output: 0.4, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: '百度2024年最新发布的自研高性能大语言模型,通用能力优异,效果比ERNIE Speed更优,适合作为基座模型进行精调,更好地处理特定场景问题,同时具备极佳的推理性能。', displayName: 'ERNIE Speed Pro 128K', @@ -827,9 +828,9 @@ const Higress: ModelProviderCard = { input: 0.3, output: 0.6, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: '百度2024年最新发布的自研高性能大语言模型,通用能力优异,适合作为基座模型进行精调,更好地处理特定场景问题,同时具备极佳的推理性能。', displayName: 'ERNIE Speed 128K', @@ -839,9 +840,9 @@ const Higress: ModelProviderCard = { input: 0, output: 0, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: '百度自研的垂直场景大语言模型,适合游戏NPC、客服对话、对话角色扮演等应用场景,人设风格更为鲜明、一致,指令遵循能力更强,推理性能更优。', displayName: 'ERNIE Character 8K', @@ -851,10 +852,10 @@ const Higress: ModelProviderCard = { input: 4, output: 8, }, - tokens: 8192, }, //混元 { + contextWindowTokens: 256_000, description: '升级为 MOE 结构,上下文窗口为 256k ,在 NLP,代码,数学,行业等多项评测集上领先众多开源模型。', displayName: 'Hunyuan Lite', @@ -866,9 +867,9 @@ const Higress: ModelProviderCard = { input: 0, output: 0, }, - tokens: 256_000, }, { + contextWindowTokens: 32_000, description: '采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-32K 性价比相对更高,在平衡效果、价格的同时,可对实现对长文本输入的处理。', displayName: 'Hunyuan Standard', @@ -880,9 +881,9 @@ const Higress: ModelProviderCard = { input: 4.5, output: 5, }, - tokens: 32_000, }, { + contextWindowTokens: 256_000, description: '采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-256K 在长度和效果上进一步突破,极大的扩展了可输入长度。', displayName: 'Hunyuan Standard 256K', @@ -894,9 +895,9 @@ const Higress: ModelProviderCard = { input: 15, output: 60, }, - tokens: 256_000, }, { + contextWindowTokens: 32_000, description: '混元全新一代大语言模型的预览版,采用全新的混合专家模型(MoE)结构,相比hunyuan-pro推理效率更快,效果表现更强。', displayName: 'Hunyuan Turbo', @@ -909,9 +910,9 @@ const Higress: ModelProviderCard = { input: 15, output: 50, }, - tokens: 32_000, }, { + contextWindowTokens: 32_000, description: '万亿级参数规模 MOE-32K 长文模型。在各种 benchmark 上达到绝对领先的水平,复杂指令和推理,具备复杂数学能力,支持 functioncall,在多语言翻译、金融法律医疗等领域应用重点优化。', displayName: 'Hunyuan Pro', @@ -924,7 +925,6 @@ const Higress: ModelProviderCard = { input: 30, output: 100, }, - tokens: 32_000, }, { description: '', @@ -941,6 +941,7 @@ const Higress: ModelProviderCard = { // tokens: 32_000, }, { + contextWindowTokens: 8000, description: '混元最新多模态模型,支持图片+文本输入生成文本内容。', displayName: 'Hunyuan Vision', enabled: true, @@ -951,10 +952,10 @@ const Higress: ModelProviderCard = { input: 18, output: 18, }, - tokens: 8000, vision: true, }, { + contextWindowTokens: 8000, description: '混元最新代码生成模型,经过 200B 高质量代码数据增训基座模型,迭代半年高质量 SFT 数据训练,上下文长窗口长度增大到 8K,五大语言代码生成自动评测指标上位居前列;五大语言10项考量各方面综合代码任务人工高质量评测上,性能处于第一梯队', displayName: 'Hunyuan Code', @@ -965,9 +966,9 @@ const Higress: ModelProviderCard = { input: 4, output: 8, }, - tokens: 8000, }, { + contextWindowTokens: 32_000, description: '混元最新 MOE 架构 FunctionCall 模型,经过高质量的 FunctionCall 数据训练,上下文窗口达 32K,在多个维度的评测指标上处于领先。', displayName: 'Hunyuan FunctionCall', @@ -979,9 +980,9 @@ const Higress: ModelProviderCard = { input: 4, output: 8, }, - tokens: 32_000, }, { + contextWindowTokens: 8000, description: '混元最新版角色扮演模型,混元官方精调训练推出的角色扮演模型,基于混元模型结合角色扮演场景数据集进行增训,在角色扮演场景具有更好的基础效果。', displayName: 'Hunyuan Role', @@ -992,10 +993,10 @@ const Higress: ModelProviderCard = { input: 4, output: 8, }, - tokens: 8000, }, //阶跃星辰 { + contextWindowTokens: 8000, description: '高速模型,适合实时对话。', displayName: 'Step 1 Flash', enabled: true, @@ -1006,9 +1007,9 @@ const Higress: ModelProviderCard = { input: 1, output: 4, }, - tokens: 8000, }, { + contextWindowTokens: 8000, description: '小型模型,适合轻量级任务。', displayName: 'Step 1 8K', enabled: true, @@ -1019,9 +1020,9 @@ const Higress: ModelProviderCard = { input: 5, output: 20, }, - tokens: 8000, }, { + contextWindowTokens: 32_000, description: '支持中等长度的对话,适用于多种应用场景。', displayName: 'Step 1 32K', enabled: true, @@ -1032,9 +1033,9 @@ const Higress: ModelProviderCard = { input: 15, output: 70, }, - tokens: 32_000, }, { + contextWindowTokens: 128_000, description: '平衡性能与成本,适合一般场景。', displayName: 'Step 1 128K', enabled: true, @@ -1045,9 +1046,9 @@ const Higress: ModelProviderCard = { input: 40, output: 200, }, - tokens: 128_000, }, { + contextWindowTokens: 256_000, description: '具备超长上下文处理能力,尤其适合长文档分析。', displayName: 'Step 1 256K', functionCall: true, @@ -1057,9 +1058,9 @@ const Higress: ModelProviderCard = { input: 95, output: 300, }, - tokens: 256_000, }, { + contextWindowTokens: 16_000, description: '支持大规模上下文交互,适合复杂对话场景。', displayName: 'Step 2 16K', enabled: true, @@ -1070,9 +1071,9 @@ const Higress: ModelProviderCard = { input: 38, output: 120, }, - tokens: 16_000, }, { + contextWindowTokens: 8000, description: '小型视觉模型,适合基本的图文任务。', displayName: 'Step 1V 8K', enabled: true, @@ -1083,10 +1084,10 @@ const Higress: ModelProviderCard = { input: 5, output: 20, }, - tokens: 8000, vision: true, }, { + contextWindowTokens: 32_000, description: '支持视觉输入,增强多模态交互体验。', displayName: 'Step 1V 32K', enabled: true, @@ -1097,10 +1098,10 @@ const Higress: ModelProviderCard = { input: 15, output: 70, }, - tokens: 32_000, vision: true, }, { + contextWindowTokens: 32_000, description: '该模型拥有强大的视频理解能力。', displayName: 'Step 1.5V Mini', enabled: true, @@ -1110,10 +1111,10 @@ const Higress: ModelProviderCard = { input: 8, output: 35, }, - tokens: 32_000, vision: true, }, { + contextWindowTokens: 8192, description: 'Spark Lite 是一款轻量级大语言模型,具备极低的延迟与高效的处理能力,完全免费开放,支持实时在线搜索功能。其快速响应的特性使其在低算力设备上的推理应用和模型微调中表现出色,为用户带来出色的成本效益和智能体验,尤其在知识问答、内容生成及搜索场景下表现不俗。', displayName: 'Spark Lite', @@ -1121,9 +1122,9 @@ const Higress: ModelProviderCard = { functionCall: false, id: 'lite', maxOutput: 4096, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Spark Pro 是一款为专业领域优化的高性能大语言模型,专注数学、编程、医疗、教育等多个领域,并支持联网搜索及内置天气、日期等插件。其优化后模型在复杂知识问答、语言理解及高层次文本创作中展现出色表现和高效性能,是适合专业应用场景的理想选择。', displayName: 'Spark Pro', @@ -1131,9 +1132,9 @@ const Higress: ModelProviderCard = { functionCall: false, id: 'generalv3', maxOutput: 8192, - tokens: 8192, }, { + contextWindowTokens: 131_072, description: 'Spark Pro 128K 配置了特大上下文处理能力,能够处理多达128K的上下文信息,特别适合需通篇分析和长期逻辑关联处理的长文内容,可在复杂文本沟通中提供流畅一致的逻辑与多样的引用支持。', displayName: 'Spark Pro 128K', @@ -1141,9 +1142,9 @@ const Higress: ModelProviderCard = { functionCall: false, id: 'pro-128k', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 8192, description: 'Spark Max 为功能最为全面的版本,支持联网搜索及众多内置插件。其全面优化的核心能力以及系统角色设定和函数调用功能,使其在各种复杂应用场景中的表现极为优异和出色。', displayName: 'Spark Max', @@ -1151,9 +1152,9 @@ const Higress: ModelProviderCard = { functionCall: false, id: 'generalv3.5', maxOutput: 8192, - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Spark Max 32K 配置了大上下文处理能力,更强的上下文理解和逻辑推理能力,支持32K tokens的文本输入,适用于长文档阅读、私有知识问答等场景', displayName: 'Spark Max 32K', @@ -1161,9 +1162,9 @@ const Higress: ModelProviderCard = { functionCall: false, id: 'max-32k', maxOutput: 8192, - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'Spark Ultra 是星火大模型系列中最为强大的版本,在升级联网搜索链路同时,提升对文本内容的理解和总结能力。它是用于提升办公生产力和准确响应需求的全方位解决方案,是引领行业的智能产品。', displayName: 'Spark 4.0 Ultra', @@ -1171,10 +1172,10 @@ const Higress: ModelProviderCard = { functionCall: false, id: '4.0Ultra', maxOutput: 8192, - tokens: 8192, }, //openai { + contextWindowTokens: 128_000, description: 'o1-mini是一款针对编程、数学和科学应用场景而设计的快速、经济高效的推理模型。该模型具有128K上下文和2023年10月的知识截止日期。', displayName: 'OpenAI o1-mini', @@ -1186,9 +1187,9 @@ const Higress: ModelProviderCard = { output: 12, }, releasedAt: '2024-09-12', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'o1是OpenAI新的推理模型,适用于需要广泛通用知识的复杂任务。该模型具有128K上下文和2023年10月的知识截止日期。', displayName: 'OpenAI o1-preview', @@ -1200,9 +1201,9 @@ const Higress: ModelProviderCard = { output: 60, }, releasedAt: '2024-09-12', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GPT-4o mini是OpenAI在GPT-4 Omni之后推出的最新模型,支持图文输入并输出文本。作为他们最先进的小型模型,它比其他近期的前沿模型便宜很多,并且比GPT-3.5 Turbo便宜超过60%。它保持了最先进的智能,同时具有显著的性价比。GPT-4o mini在MMLU测试中获得了 82% 的得分,目前在聊天偏好上排名高于 GPT-4。', displayName: 'GPT-4o mini', @@ -1214,10 +1215,10 @@ const Higress: ModelProviderCard = { input: 0.15, output: 0.6, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'GPT-4o', @@ -1228,10 +1229,10 @@ const Higress: ModelProviderCard = { input: 2.5, output: 10, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'GPT-4o 0806', @@ -1241,10 +1242,10 @@ const Higress: ModelProviderCard = { input: 2.5, output: 10, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'GPT-4o 0513', @@ -1254,10 +1255,10 @@ const Higress: ModelProviderCard = { input: 5, output: 15, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'ChatGPT-4o', @@ -1267,10 +1268,10 @@ const Higress: ModelProviderCard = { input: 5, output: 15, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo', @@ -1280,10 +1281,10 @@ const Higress: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo Vision 0409', @@ -1293,10 +1294,10 @@ const Higress: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo Preview', @@ -1306,9 +1307,9 @@ const Higress: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo Preview 0125', @@ -1318,9 +1319,9 @@ const Higress: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo Preview 1106', @@ -1330,9 +1331,9 @@ const Higress: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', displayName: 'GPT-4', @@ -1342,9 +1343,9 @@ const Higress: ModelProviderCard = { input: 30, output: 60, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', displayName: 'GPT-4 0613', @@ -1354,11 +1355,12 @@ const Higress: ModelProviderCard = { input: 30, output: 60, }, - tokens: 8192, }, { + contextWindowTokens: 32_768, description: - 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', // Will be discontinued on June 6, 2025 + 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', + // Will be discontinued on June 6, 2025 displayName: 'GPT-4 32K', functionCall: true, id: 'gpt-4-32k', @@ -1366,9 +1368,9 @@ const Higress: ModelProviderCard = { input: 60, output: 120, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, // Will be discontinued on June 6, 2025 description: 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', @@ -1379,9 +1381,9 @@ const Higress: ModelProviderCard = { input: 60, output: 120, }, - tokens: 32_768, }, { + contextWindowTokens: 16_385, description: 'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', displayName: 'GPT-3.5 Turbo', @@ -1391,9 +1393,9 @@ const Higress: ModelProviderCard = { input: 0.5, output: 1.5, }, - tokens: 16_385, }, { + contextWindowTokens: 16_385, description: 'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', displayName: 'GPT-3.5 Turbo 0125', @@ -1403,9 +1405,9 @@ const Higress: ModelProviderCard = { input: 0.5, output: 1.5, }, - tokens: 16_385, }, { + contextWindowTokens: 16_385, description: 'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', displayName: 'GPT-3.5 Turbo 1106', @@ -1415,9 +1417,9 @@ const Higress: ModelProviderCard = { input: 1, output: 2, }, - tokens: 16_385, }, { + contextWindowTokens: 4096, description: 'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', displayName: 'GPT-3.5 Turbo Instruct', @@ -1426,10 +1428,10 @@ const Higress: ModelProviderCard = { input: 1.5, output: 2, }, - tokens: 4096, }, //azure { + contextWindowTokens: 16_385, deploymentName: 'gpt-35-turbo', description: 'GPT 3.5 Turbo,OpenAI提供的高效模型,适用于聊天和文本生成任务,支持并行函数调用。', @@ -1438,66 +1440,66 @@ const Higress: ModelProviderCard = { functionCall: true, id: 'gpt-35-turbo', maxOutput: 4096, - tokens: 16_385, }, { + contextWindowTokens: 16_384, deploymentName: 'gpt-35-turbo-16k', description: 'GPT 3.5 Turbo 16k,高容量文本生成模型,适合复杂任务。', displayName: 'GPT 3.5 Turbo', functionCall: true, id: 'gpt-35-turbo-16k', - tokens: 16_384, }, { + contextWindowTokens: 128_000, deploymentName: 'gpt-4-turbo', description: 'GPT 4 Turbo,多模态模型,提供杰出的语言理解和生成能力,同时支持图像输入。', displayName: 'GPT 4 Turbo', enabled: true, functionCall: true, id: 'gpt-4', - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, deploymentName: 'gpt-4-vision', description: 'GPT-4 视觉预览版,专为图像分析和处理任务设计。', displayName: 'GPT 4 Turbo with Vision Preview', id: 'gpt-4-vision-preview', - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, deploymentName: 'gpt-4o-mini', description: 'GPT-4o Mini,小型高效模型,具备与GPT-4o相似的卓越性能。', displayName: 'GPT 4o Mini', enabled: true, functionCall: true, id: 'gpt-4o-mini', - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, deploymentName: 'gpt-4o', description: 'GPT-4o 是最新的多模态模型,结合高级文本和图像处理能力。', displayName: 'GPT 4o', enabled: true, functionCall: true, id: 'gpt-4o', - tokens: 128_000, vision: true, }, //github { + contextWindowTokens: 128_000, description: '比 o1-preview 更小、更快,成本低80%,在代码生成和小上下文操作方面表现良好。', displayName: 'OpenAI o1-mini', enabled: true, functionCall: false, id: 'o1-mini', maxOutput: 65_536, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '专注于高级推理和解决复杂问题,包括数学和科学任务。非常适合需要深度上下文理解和自主工作流程的应用。', displayName: 'OpenAI o1-preview', @@ -1505,199 +1507,199 @@ const Higress: ModelProviderCard = { functionCall: false, id: 'o1-preview', maxOutput: 32_768, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '一种经济高效的AI解决方案,适用于多种文本和图像任务。', displayName: 'OpenAI GPT-4o mini', enabled: true, functionCall: true, id: 'gpt-4o-mini', maxOutput: 4096, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'OpenAI GPT-4系列中最先进的多模态模型,可以处理文本和图像输入。', displayName: 'OpenAI GPT-4o', enabled: true, functionCall: true, id: 'gpt-4o', maxOutput: 4096, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 262_144, description: '一个52B参数(12B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。', displayName: 'AI21 Jamba 1.5 Mini', functionCall: true, id: 'ai21-jamba-1.5-mini', maxOutput: 4096, - tokens: 262_144, }, { + contextWindowTokens: 262_144, description: '一个398B参数(94B活跃)的多语言模型,提供256K长上下文窗口、函数调用、结构化输出和基于事实的生成。', displayName: 'AI21 Jamba 1.5 Large', functionCall: true, id: 'ai21-jamba-1.5-large', maxOutput: 4096, - tokens: 262_144, }, { + contextWindowTokens: 131_072, description: 'Command R是一个可扩展的生成模型,旨在针对RAG和工具使用,使企业能够实现生产级AI。', displayName: 'Cohere Command R', id: 'cohere-command-r', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Command R+是一个最先进的RAG优化模型,旨在应对企业级工作负载。', displayName: 'Cohere Command R+', id: 'cohere-command-r-plus', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Mistral Nemo是一种尖端的语言模型(LLM),在其尺寸类别中拥有最先进的推理、世界知识和编码能力。', displayName: 'Mistral Nemo', id: 'mistral-nemo', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Mistral Small可用于任何需要高效率和低延迟的基于语言的任务。', displayName: 'Mistral Small', id: 'mistral-small', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Mistral的旗舰模型,适合需要大规模推理能力或高度专业化的复杂任务(合成文本生成、代码生成、RAG或代理)。', displayName: 'Mistral Large', id: 'mistral-large', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '在高分辨率图像上表现出色的图像推理能力,适用于视觉理解应用。', displayName: 'Llama 3.2 11B Vision', id: 'llama-3.2-11b-vision-instruct', maxOutput: 4096, - tokens: 131_072, vision: true, }, { + contextWindowTokens: 131_072, description: '适用于视觉理解代理应用的高级图像推理能力。', displayName: 'Llama 3.2 90B Vision', id: 'llama-3.2-90b-vision-instruct', maxOutput: 4096, - tokens: 131_072, vision: true, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。', displayName: 'Meta Llama 3.1 8B', id: 'meta-llama-3.1-8b-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。', displayName: 'Meta Llama 3.1 70B', id: 'meta-llama-3.1-70b-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1指令调优的文本模型,针对多语言对话用例进行了优化,在许多可用的开源和封闭聊天模型中,在常见行业基准上表现优异。', displayName: 'Meta Llama 3.1 405B', id: 'meta-llama-3.1-405b-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 8192, description: '一个多功能的80亿参数模型,针对对话和文本生成任务进行了优化。', displayName: 'Meta Llama 3 8B', id: 'meta-llama-3-8b-instruct', maxOutput: 4096, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '一个强大的700亿参数模型,在推理、编码和广泛的语言应用方面表现出色。', displayName: 'Meta Llama 3 70B', id: 'meta-llama-3-70b-instruct', maxOutput: 4096, - tokens: 8192, }, { + contextWindowTokens: 131_072, description: 'Phi-3-mini模型的更新版。', displayName: 'Phi-3.5-mini 128K', id: 'Phi-3.5-mini-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Phi-3-vision模型的更新版。', displayName: 'Phi-3.5-vision 128K', id: 'Phi-3.5-vision-instrust', maxOutput: 4096, - tokens: 131_072, vision: true, }, { + contextWindowTokens: 4096, description: 'Phi-3家族中最小的成员,针对质量和低延迟进行了优化。', displayName: 'Phi-3-mini 4K', id: 'Phi-3-mini-4k-instruct', maxOutput: 4096, - tokens: 4096, }, { + contextWindowTokens: 131_072, description: '相同的Phi-3-mini模型,但具有更大的上下文大小,适用于RAG或少量提示。', displayName: 'Phi-3-mini 128K', id: 'Phi-3-mini-128k-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 8192, description: '一个70亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。', displayName: 'Phi-3-small 8K', id: 'Phi-3-small-8k-instruct', maxOutput: 4096, - tokens: 8192, }, { + contextWindowTokens: 131_072, description: '相同的Phi-3-small模型,但具有更大的上下文大小,适用于RAG或少量提示。', displayName: 'Phi-3-small 128K', id: 'Phi-3-small-128k-instruct', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 4096, description: '一个140亿参数模型,质量优于Phi-3-mini,重点关注高质量、推理密集型数据。', displayName: 'Phi-3-medium 4K', id: 'Phi-3-medium-4k-instruct', maxOutput: 4096, - tokens: 4096, }, { + contextWindowTokens: 131_072, description: '相同的Phi-3-medium模型,但具有更大的上下文大小,适用于RAG或少量提示。', displayName: 'Phi-3-medium 128K', id: 'Phi-3-medium-128k-instruct', maxOutput: 4096, - tokens: 131_072, }, //groq { + contextWindowTokens: 8192, description: 'Llama 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 11B Vision (Preview)', @@ -1708,10 +1710,10 @@ const Higress: ModelProviderCard = { input: 0.05, output: 0.08, }, - tokens: 8192, vision: true, }, { + contextWindowTokens: 8192, description: 'Llama 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 90B Vision (Preview)', @@ -1722,10 +1724,10 @@ const Higress: ModelProviderCard = { input: 0.59, output: 0.79, }, - tokens: 8192, vision: true, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 8B 是一款高效能模型,提供了快速的文本生成能力,非常适合需要大规模效率和成本效益的应用场景。', displayName: 'Llama 3.1 8B', @@ -1737,9 +1739,9 @@ const Higress: ModelProviderCard = { input: 0.05, output: 0.08, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 70B 提供更强大的AI推理能力,适合复杂应用,支持超多的计算处理并保证高效和准确率。', displayName: 'Llama 3.1 70B', @@ -1751,7 +1753,6 @@ const Higress: ModelProviderCard = { input: 0.59, output: 0.79, }, - tokens: 131_072, }, /* // Offline due to overwhelming demand! Stay tuned for updates. @@ -1763,6 +1764,7 @@ const Higress: ModelProviderCard = { }, */ { + contextWindowTokens: 8192, description: 'Llama 3 Groq 8B Tool Use 是针对高效工具使用优化的模型,支持快速并行计算。', displayName: 'Llama 3 Groq 8B Tool Use (Preview)', functionCall: true, @@ -1771,9 +1773,9 @@ const Higress: ModelProviderCard = { input: 0.19, output: 0.19, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 Groq 70B Tool Use 提供强大的工具调用能力,支持复杂任务的高效处理。', displayName: 'Llama 3 Groq 70B Tool Use (Preview)', functionCall: true, @@ -1782,9 +1784,9 @@ const Higress: ModelProviderCard = { input: 0.89, output: 0.89, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Meta Llama 3 8B 带来优质的推理效能,适合多场景应用需求。', displayName: 'Meta Llama 3 8B', functionCall: true, @@ -1793,9 +1795,9 @@ const Higress: ModelProviderCard = { input: 0.05, output: 0.08, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Meta Llama 3 70B 提供无与伦比的复杂性处理能力,为高要求项目量身定制。', displayName: 'Meta Llama 3 70B', functionCall: true, @@ -1804,9 +1806,9 @@ const Higress: ModelProviderCard = { input: 0.59, output: 0.79, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma 2 9B 是一款优化用于特定任务和工具整合的模型。', displayName: 'Gemma 2 9B', enabled: true, @@ -1816,9 +1818,9 @@ const Higress: ModelProviderCard = { input: 0.2, output: 0.2, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma 7B 适合中小规模任务处理,兼具成本效益。', displayName: 'Gemma 7B', functionCall: true, @@ -1827,9 +1829,9 @@ const Higress: ModelProviderCard = { input: 0.07, output: 0.07, }, - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Mixtral 8x7B 提供高容错的并行计算能力,适合复杂任务。', displayName: 'Mixtral 8x7B', functionCall: true, @@ -1838,17 +1840,17 @@ const Higress: ModelProviderCard = { input: 0.24, output: 0.24, }, - tokens: 32_768, }, { + contextWindowTokens: 4096, description: 'LLaVA 1.5 7B 提供视觉处理能力融合,通过视觉信息输入生成复杂输出。', displayName: 'LLaVA 1.5 7B', id: 'llava-v1.5-7b-4096-preview', - tokens: 4096, vision: true, }, //deepseek { + contextWindowTokens: 128_000, description: '融合通用与代码能力的全新开源模型, 不仅保留了原有 Chat 模型的通用对话能力和 Coder 模型的强大代码处理能力,还更好地对齐了人类偏好。此外,DeepSeek-V2.5 在写作任务、指令跟随等多个方面也实现了大幅提升。', displayName: 'DeepSeek V2.5', @@ -1861,10 +1863,10 @@ const Higress: ModelProviderCard = { output: 0.28, }, releasedAt: '2024-09-05', - tokens: 128_000, }, //claude { + contextWindowTokens: 200_000, description: 'Claude 3.5 Haiku 是 Anthropic 最快的下一代模型。与 Claude 3 Haiku 相比,Claude 3.5 Haiku 在各项技能上都有所提升,并在许多智力基准测试中超越了上一代最大的模型 Claude 3 Opus。', displayName: 'Claude 3.5 Haiku', @@ -1879,9 +1881,9 @@ const Higress: ModelProviderCard = { writeCacheInput: 1.25, }, releasedAt: '2024-11-05', - tokens: 200_000, }, { + contextWindowTokens: 200_000, description: 'Claude 3.5 Sonnet 提供了超越 Opus 的能力和比 Sonnet 更快的速度,同时保持与 Sonnet 相同的价格。Sonnet 特别擅长编程、数据科学、视觉处理、代理任务。', displayName: 'Claude 3.5 Sonnet', @@ -1896,10 +1898,10 @@ const Higress: ModelProviderCard = { writeCacheInput: 3.75, }, releasedAt: '2024-10-22', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3.5 Sonnet 提供了超越 Opus 的能力和比 Sonnet 更快的速度,同时保持与 Sonnet 相同的价格。Sonnet 特别擅长编程、数据科学、视觉处理、代理任务。', displayName: 'Claude 3.5 Sonnet 0620', @@ -1913,10 +1915,10 @@ const Higress: ModelProviderCard = { writeCacheInput: 3.75, }, releasedAt: '2024-06-20', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Haiku 是 Anthropic 的最快且最紧凑的模型,旨在实现近乎即时的响应。它具有快速且准确的定向性能。', displayName: 'Claude 3 Haiku', @@ -1928,10 +1930,10 @@ const Higress: ModelProviderCard = { output: 1.25, }, releasedAt: '2024-03-07', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Sonnet 在智能和速度方面为企业工作负载提供了理想的平衡。它以更低的价格提供最大效用,可靠且适合大规模部署。', displayName: 'Claude 3 Sonnet', @@ -1943,10 +1945,10 @@ const Higress: ModelProviderCard = { output: 15, }, releasedAt: '2024-02-29', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Opus 是 Anthropic 用于处理高度复杂任务的最强大模型。它在性能、智能、流畅性和理解力方面表现卓越。', displayName: 'Claude 3 Opus', @@ -1959,10 +1961,10 @@ const Higress: ModelProviderCard = { output: 75, }, releasedAt: '2024-02-29', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 2 为企业提供了关键能力的进步,包括业界领先的 200K token 上下文、大幅降低模型幻觉的发生率、系统提示以及一个新的测试功能:工具调用。', displayName: 'Claude 2.1', @@ -1973,9 +1975,9 @@ const Higress: ModelProviderCard = { output: 24, }, releasedAt: '2023-11-21', - tokens: 200_000, }, { + contextWindowTokens: 100_000, description: 'Claude 2 为企业提供了关键能力的进步,包括业界领先的 200K token 上下文、大幅降低模型幻觉的发生率、系统提示以及一个新的测试功能:工具调用。', displayName: 'Claude 2.0', @@ -1986,10 +1988,10 @@ const Higress: ModelProviderCard = { output: 24, }, releasedAt: '2023-07-11', - tokens: 100_000, }, //gemini { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 是Google最新的多模态AI模型,具备快速处理能力,支持文本、图像和视频输入,适用于多种任务的高效扩展。', displayName: 'Gemini 1.5 Flash', @@ -2002,10 +2004,10 @@ const Higress: ModelProviderCard = { input: 0.075, output: 0.3, }, - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 002 是一款高效的多模态模型,支持广泛应用的扩展。', displayName: 'Gemini 1.5 Flash 002', enabled: true, @@ -2018,10 +2020,10 @@ const Higress: ModelProviderCard = { output: 0.3, }, releasedAt: '2024-09-25', - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 001 是一款高效的多模态模型,支持广泛应用的扩展。', displayName: 'Gemini 1.5 Flash 001', functionCall: true, @@ -2032,10 +2034,10 @@ const Higress: ModelProviderCard = { input: 0.075, output: 0.3, }, - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 0827 提供了优化后的多模态处理能力,适用多种复杂任务场景。', displayName: 'Gemini 1.5 Flash 0827', functionCall: true, @@ -2047,10 +2049,10 @@ const Higress: ModelProviderCard = { output: 0.3, }, releasedAt: '2024-08-27', - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 8B 是一款高效的多模态模型,支持广泛应用的扩展。', displayName: 'Gemini 1.5 Flash 8B', enabled: true, @@ -2063,10 +2065,10 @@ const Higress: ModelProviderCard = { output: 0.3, }, releasedAt: '2024-10-03', - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 8B 0924 是最新的实验性模型,在文本和多模态用例中都有显著的性能提升。', displayName: 'Gemini 1.5 Flash 8B 0924', @@ -2079,10 +2081,10 @@ const Higress: ModelProviderCard = { output: 0.3, }, releasedAt: '2024-09-24', - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 支持高达200万个tokens,是中型多模态模型的理想选择,适用于复杂任务的多方面支持。', displayName: 'Gemini 1.5 Pro', @@ -2096,10 +2098,10 @@ const Higress: ModelProviderCard = { output: 10.5, }, releasedAt: '2024-02-15', - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 002 是最新的生产就绪模型,提供更高质量的输出,特别在数学、长上下文和视觉任务方面有显著提升。', displayName: 'Gemini 1.5 Pro 002', @@ -2113,10 +2115,10 @@ const Higress: ModelProviderCard = { output: 2.5, }, releasedAt: '2024-09-24', - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 001 是可扩展的多模态AI解决方案,支持广泛的复杂任务。', displayName: 'Gemini 1.5 Pro 001', functionCall: true, @@ -2128,10 +2130,10 @@ const Higress: ModelProviderCard = { output: 10.5, }, releasedAt: '2024-02-15', - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 0827 结合最新优化技术,带来更高效的多模态数据处理能力。', displayName: 'Gemini 1.5 Pro 0827', functionCall: true, @@ -2143,10 +2145,10 @@ const Higress: ModelProviderCard = { output: 10.5, }, releasedAt: '2024-08-27', - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 0801 提供出色的多模态处理能力,为应用开发带来更大灵活性。', displayName: 'Gemini 1.5 Pro 0801', functionCall: true, @@ -2158,10 +2160,10 @@ const Higress: ModelProviderCard = { output: 10.5, }, releasedAt: '2024-08-01', - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 30_720 + 2048, description: 'Gemini 1.0 Pro 是Google的高性能AI模型,专为广泛任务扩展而设计。', displayName: 'Gemini 1.0 Pro', id: 'gemini-1.0-pro-latest', @@ -2171,9 +2173,9 @@ const Higress: ModelProviderCard = { output: 1.5, }, releasedAt: '2023-12-06', - tokens: 30_720 + 2048, }, { + contextWindowTokens: 30_720 + 2048, description: 'Gemini 1.0 Pro 001 (Tuning) 提供稳定并可调优的性能,是复杂任务解决方案的理想选择。', displayName: 'Gemini 1.0 Pro 001 (Tuning)', @@ -2185,9 +2187,9 @@ const Higress: ModelProviderCard = { output: 1.5, }, releasedAt: '2023-12-06', - tokens: 30_720 + 2048, }, { + contextWindowTokens: 30_720 + 2048, description: 'Gemini 1.0 Pro 002 (Tuning) 提供出色的多模态支持,专注于复杂任务的有效解决。', displayName: 'Gemini 1.0 Pro 002 (Tuning)', id: 'gemini-1.0-pro-002', @@ -2197,11 +2199,11 @@ const Higress: ModelProviderCard = { output: 1.5, }, releasedAt: '2023-12-06', - tokens: 30_720 + 2048, }, //mistral { + contextWindowTokens: 128_000, description: 'Mistral Nemo是一个与Nvidia合作开发的12B模型,提供出色的推理和编码性能,易于集成和替换。', displayName: 'Mistral Nemo', @@ -2212,9 +2214,9 @@ const Higress: ModelProviderCard = { input: 0.15, output: 0.15, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Mistral Small是成本效益高、快速且可靠的选项,适用于翻译、摘要和情感分析等用例。', displayName: 'Mistral Small', @@ -2225,9 +2227,9 @@ const Higress: ModelProviderCard = { input: 0.2, output: 0.6, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Mistral Large是旗舰大模型,擅长多语言任务、复杂推理和代码生成,是高端应用的理想选择。', displayName: 'Mistral Large', @@ -2238,9 +2240,9 @@ const Higress: ModelProviderCard = { input: 2, output: 6, }, - tokens: 128_000, }, { + contextWindowTokens: 32_768, description: 'Codestral是专注于代码生成的尖端生成模型,优化了中间填充和代码补全任务。', displayName: 'Codestral', id: 'codestral-latest', @@ -2248,9 +2250,9 @@ const Higress: ModelProviderCard = { input: 0.2, output: 0.6, }, - tokens: 32_768, }, { + contextWindowTokens: 128_000, description: 'Pixtral 模型在图表和图理解、文档问答、多模态推理和指令遵循等任务上表现出强大的能力,能够以自然分辨率和宽高比摄入图像,还能够在长达 128K 令牌的长上下文窗口中处理任意数量的图像。', displayName: 'Pixtral 12B', @@ -2260,10 +2262,10 @@ const Higress: ModelProviderCard = { input: 0.15, output: 0.15, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'Ministral 3B 是Mistral的世界顶级边缘模型。', displayName: 'Ministral 3B', id: 'ministral-3b-latest', @@ -2271,9 +2273,9 @@ const Higress: ModelProviderCard = { input: 0.04, output: 0.04, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Ministral 8B 是Mistral的性价比极高的边缘模型。', displayName: 'Ministral 8B', id: 'ministral-8b-latest', @@ -2281,9 +2283,9 @@ const Higress: ModelProviderCard = { input: 0.1, output: 0.1, }, - tokens: 128_000, }, { + contextWindowTokens: 32_768, description: 'Mistral 7B是一款紧凑但高性能的模型,擅长批量处理和简单任务,如分类和文本生成,具有良好的推理能力。', displayName: 'Mistral 7B', @@ -2292,9 +2294,9 @@ const Higress: ModelProviderCard = { input: 0.25, output: 0.25, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Mixtral 8x7B是一个稀疏专家模型,利用多个参数提高推理速度,适合处理多语言和代码生成任务。', displayName: 'Mixtral 8x7B', @@ -2303,9 +2305,9 @@ const Higress: ModelProviderCard = { input: 0.7, output: 0.7, }, - tokens: 32_768, }, { + contextWindowTokens: 65_536, description: 'Mixtral 8x22B是一个更大的专家模型,专注于复杂任务,提供出色的推理能力和更高的吞吐量。', displayName: 'Mixtral 8x22B', @@ -2315,9 +2317,9 @@ const Higress: ModelProviderCard = { input: 2, output: 6, }, - tokens: 65_536, }, { + contextWindowTokens: 256_000, description: 'Codestral Mamba是专注于代码生成的Mamba 2语言模型,为先进的代码和推理任务提供强力支持。', displayName: 'Codestral Mamba', @@ -2326,44 +2328,43 @@ const Higress: ModelProviderCard = { input: 0.15, output: 0.15, }, - tokens: 256_000, }, //minimax { + contextWindowTokens: 245_760, description: '适用于广泛的自然语言处理任务,包括文本生成、对话系统等。', displayName: 'abab6.5s', enabled: true, functionCall: true, id: 'abab6.5s-chat', - tokens: 245_760, }, { + contextWindowTokens: 8192, description: '专为多语种人设对话设计,支持英文及其他多种语言的高质量对话生成。', displayName: 'abab6.5g', enabled: true, functionCall: true, id: 'abab6.5g-chat', - tokens: 8192, }, { + contextWindowTokens: 8192, description: '针对中文人设对话场景优化,提供流畅且符合中文表达习惯的对话生成能力。', displayName: 'abab6.5t', enabled: true, functionCall: true, id: 'abab6.5t-chat', - tokens: 8192, }, { + contextWindowTokens: 16_384, description: '面向生产力场景,支持复杂任务处理和高效文本生成,适用于专业领域应用。', displayName: 'abab5.5', id: 'abab5.5-chat', - tokens: 16_384, }, { + contextWindowTokens: 8192, description: '专为中文人设对话场景设计,提供高质量的中文对话生成能力,适用于多种应用场景。', displayName: 'abab5.5s', id: 'abab5.5s-chat', - tokens: 8192, }, //cohere { diff --git a/src/config/modelProviders/huggingface.ts b/src/config/modelProviders/huggingface.ts index 51f6ffb506ab..a24445cd235d 100644 --- a/src/config/modelProviders/huggingface.ts +++ b/src/config/modelProviders/huggingface.ts @@ -3,45 +3,45 @@ import { ModelProviderCard } from '@/types/llm'; const HuggingFace: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 32_768, description: 'Mistral AI的指令调优模型', displayName: 'Mistral 7B Instruct v0.3', enabled: true, id: 'mistralai/Mistral-7B-Instruct-v0.3', - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'Google的轻量级指令调优模型', displayName: 'Gemma 2 2B Instruct', id: 'google/gemma-2-2b-it', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: '阿里云通义千问团队开发的大型语言模型', displayName: 'Qwen 2.5 72B Instruct', id: 'Qwen/Qwen2.5-72B-Instruct', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Qwen2.5-Coder 专注于代码编写', displayName: 'Qwen 2.5 Coder 32B Instruct', id: 'Qwen/Qwen2.5-Coder-32B-Instruct', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Qwen QwQ 是由 Qwen 团队开发的实验研究模型,专注于提升AI推理能力。', displayName: 'QwQ 32B Preview', id: 'Qwen/QwQ-32B-Preview', - tokens: 32_768, }, { + contextWindowTokens: 32_768, displayName: 'Phi 3.5 mini instruct', id: 'microsoft/Phi-3.5-mini-instruct', - tokens: 32_768, }, { + contextWindowTokens: 16_384, displayName: 'Hermes 3 Llama 3.1 8B', id: 'NousResearch/Hermes-3-Llama-3.1-8B', - tokens: 16_384, }, ], checkModel: 'mistralai/Mistral-7B-Instruct-v0.2', diff --git a/src/config/modelProviders/hunyuan.ts b/src/config/modelProviders/hunyuan.ts index cde75f702d21..9de4440812ed 100644 --- a/src/config/modelProviders/hunyuan.ts +++ b/src/config/modelProviders/hunyuan.ts @@ -4,7 +4,9 @@ import { ModelProviderCard } from '@/types/llm'; const Hunyuan: ModelProviderCard = { chatModels: [ { - description: '升级为 MOE 结构,上下文窗口为 256k ,在 NLP,代码,数学,行业等多项评测集上领先众多开源模型。', + contextWindowTokens: 256_000, + description: + '升级为 MOE 结构,上下文窗口为 256k ,在 NLP,代码,数学,行业等多项评测集上领先众多开源模型。', displayName: 'Hunyuan Lite', enabled: true, id: 'hunyuan-lite', @@ -14,10 +16,11 @@ const Hunyuan: ModelProviderCard = { input: 0, output: 0, }, - tokens: 256_000, }, { - description: '采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-32K 性价比相对更高,在平衡效果、价格的同时,可对实现对长文本输入的处理。', + contextWindowTokens: 32_000, + description: + '采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-32K 性价比相对更高,在平衡效果、价格的同时,可对实现对长文本输入的处理。', displayName: 'Hunyuan Standard', enabled: true, id: 'hunyuan-standard', @@ -27,10 +30,11 @@ const Hunyuan: ModelProviderCard = { input: 4.5, output: 5, }, - tokens: 32_000, }, { - description: '采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-256K 在长度和效果上进一步突破,极大的扩展了可输入长度。', + contextWindowTokens: 256_000, + description: + '采用更优的路由策略,同时缓解了负载均衡和专家趋同的问题。长文方面,大海捞针指标达到99.9%。MOE-256K 在长度和效果上进一步突破,极大的扩展了可输入长度。', displayName: 'Hunyuan Standard 256K', enabled: true, id: 'hunyuan-standard-256K', @@ -40,10 +44,11 @@ const Hunyuan: ModelProviderCard = { input: 15, output: 60, }, - tokens: 256_000, }, { - description: '混元全新一代大语言模型的预览版,采用全新的混合专家模型(MoE)结构,相比hunyuan-pro推理效率更快,效果表现更强。', + contextWindowTokens: 32_000, + description: + '混元全新一代大语言模型的预览版,采用全新的混合专家模型(MoE)结构,相比hunyuan-pro推理效率更快,效果表现更强。', displayName: 'Hunyuan Turbo', enabled: true, functionCall: true, @@ -54,10 +59,11 @@ const Hunyuan: ModelProviderCard = { input: 15, output: 50, }, - tokens: 32_000, }, { - description: '万亿级参数规模 MOE-32K 长文模型。在各种 benchmark 上达到绝对领先的水平,复杂指令和推理,具备复杂数学能力,支持 functioncall,在多语言翻译、金融法律医疗等领域应用重点优化。', + contextWindowTokens: 32_000, + description: + '万亿级参数规模 MOE-32K 长文模型。在各种 benchmark 上达到绝对领先的水平,复杂指令和推理,具备复杂数学能力,支持 functioncall,在多语言翻译、金融法律医疗等领域应用重点优化。', displayName: 'Hunyuan Pro', enabled: true, functionCall: true, @@ -68,9 +74,9 @@ const Hunyuan: ModelProviderCard = { input: 30, output: 100, }, - tokens: 32_000, }, { + contextWindowTokens: 8000, description: '混元最新多模态模型,支持图片+文本输入生成文本内容。', displayName: 'Hunyuan Vision', enabled: true, @@ -81,11 +87,12 @@ const Hunyuan: ModelProviderCard = { input: 18, output: 18, }, - tokens: 8000, vision: true, }, { - description: '混元最新代码生成模型,经过 200B 高质量代码数据增训基座模型,迭代半年高质量 SFT 数据训练,上下文长窗口长度增大到 8K,五大语言代码生成自动评测指标上位居前列;五大语言10项考量各方面综合代码任务人工高质量评测上,性能处于第一梯队', + contextWindowTokens: 8000, + description: + '混元最新代码生成模型,经过 200B 高质量代码数据增训基座模型,迭代半年高质量 SFT 数据训练,上下文长窗口长度增大到 8K,五大语言代码生成自动评测指标上位居前列;五大语言10项考量各方面综合代码任务人工高质量评测上,性能处于第一梯队', displayName: 'Hunyuan Code', id: 'hunyuan-code', maxOutput: 4000, @@ -94,10 +101,11 @@ const Hunyuan: ModelProviderCard = { input: 4, output: 8, }, - tokens: 8000, }, { - description: '混元最新 MOE 架构 FunctionCall 模型,经过高质量的 FunctionCall 数据训练,上下文窗口达 32K,在多个维度的评测指标上处于领先。', + contextWindowTokens: 32_000, + description: + '混元最新 MOE 架构 FunctionCall 模型,经过高质量的 FunctionCall 数据训练,上下文窗口达 32K,在多个维度的评测指标上处于领先。', displayName: 'Hunyuan FunctionCall', functionCall: true, id: 'hunyuan-functioncall', @@ -107,10 +115,11 @@ const Hunyuan: ModelProviderCard = { input: 4, output: 8, }, - tokens: 32_000, }, { - description: '混元最新版角色扮演模型,混元官方精调训练推出的角色扮演模型,基于混元模型结合角色扮演场景数据集进行增训,在角色扮演场景具有更好的基础效果。', + contextWindowTokens: 8000, + description: + '混元最新版角色扮演模型,混元官方精调训练推出的角色扮演模型,基于混元模型结合角色扮演场景数据集进行增训,在角色扮演场景具有更好的基础效果。', displayName: 'Hunyuan Role', id: 'hunyuan-role', maxOutput: 4000, @@ -119,7 +128,6 @@ const Hunyuan: ModelProviderCard = { input: 4, output: 8, }, - tokens: 8000, }, ], checkModel: 'hunyuan-lite', diff --git a/src/config/modelProviders/internlm.ts b/src/config/modelProviders/internlm.ts index 07e07117ea47..b3de5a5cdfb1 100644 --- a/src/config/modelProviders/internlm.ts +++ b/src/config/modelProviders/internlm.ts @@ -3,7 +3,9 @@ import { ModelProviderCard } from '@/types/llm'; const InternLM: ModelProviderCard = { chatModels: [ { - description: '我们最新的模型系列,有着卓越的推理性能,支持 1M 的上下文长度以及更强的指令跟随和工具调用能力。', + contextWindowTokens: 32_768, + description: + '我们最新的模型系列,有着卓越的推理性能,支持 1M 的上下文长度以及更强的指令跟随和工具调用能力。', displayName: 'InternLM2.5', enabled: true, functionCall: true, @@ -13,9 +15,9 @@ const InternLM: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: '我们仍在维护的老版本模型,有 7B、20B 多种模型参数量可选。', displayName: 'InternLM2 Pro Chat', functionCall: true, @@ -25,7 +27,6 @@ const InternLM: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, ], checkModel: 'internlm2.5-latest', @@ -34,7 +35,8 @@ const InternLM: ModelProviderCard = { disableBrowserRequest: true, id: 'internlm', modelList: { showModelFetcher: true }, - modelsUrl: 'https://internlm.intern-ai.org.cn/doc/docs/Models#%E8%8E%B7%E5%8F%96%E6%A8%A1%E5%9E%8B%E5%88%97%E8%A1%A8', + modelsUrl: + 'https://internlm.intern-ai.org.cn/doc/docs/Models#%E8%8E%B7%E5%8F%96%E6%A8%A1%E5%9E%8B%E5%88%97%E8%A1%A8', name: 'InternLM', url: 'https://internlm.intern-ai.org.cn', }; diff --git a/src/config/modelProviders/minimax.ts b/src/config/modelProviders/minimax.ts index 4b8ac4da6694..e6c2af611b7c 100644 --- a/src/config/modelProviders/minimax.ts +++ b/src/config/modelProviders/minimax.ts @@ -4,40 +4,40 @@ import { ModelProviderCard } from '@/types/llm'; const Minimax: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 245_760, description: '适用于广泛的自然语言处理任务,包括文本生成、对话系统等。', displayName: 'abab6.5s', enabled: true, functionCall: true, id: 'abab6.5s-chat', - tokens: 245_760, }, { + contextWindowTokens: 8192, description: '专为多语种人设对话设计,支持英文及其他多种语言的高质量对话生成。', displayName: 'abab6.5g', enabled: true, functionCall: true, id: 'abab6.5g-chat', - tokens: 8192, }, { + contextWindowTokens: 8192, description: '针对中文人设对话场景优化,提供流畅且符合中文表达习惯的对话生成能力。', displayName: 'abab6.5t', enabled: true, functionCall: true, id: 'abab6.5t-chat', - tokens: 8192, }, { + contextWindowTokens: 16_384, description: '面向生产力场景,支持复杂任务处理和高效文本生成,适用于专业领域应用。', displayName: 'abab5.5', id: 'abab5.5-chat', - tokens: 16_384, }, { + contextWindowTokens: 8192, description: '专为中文人设对话场景设计,提供高质量的中文对话生成能力,适用于多种应用场景。', displayName: 'abab5.5s', id: 'abab5.5s-chat', - tokens: 8192, }, ], checkModel: 'abab6.5s-chat', diff --git a/src/config/modelProviders/mistral.ts b/src/config/modelProviders/mistral.ts index 39ab32f05684..0bb9871d5e33 100644 --- a/src/config/modelProviders/mistral.ts +++ b/src/config/modelProviders/mistral.ts @@ -5,6 +5,7 @@ import { ModelProviderCard } from '@/types/llm'; const Mistral: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 128_000, description: 'Mistral Nemo是一个与Nvidia合作开发的12B模型,提供出色的推理和编码性能,易于集成和替换。', displayName: 'Mistral Nemo', @@ -15,9 +16,9 @@ const Mistral: ModelProviderCard = { input: 0.15, output: 0.15, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Mistral Small是成本效益高、快速且可靠的选项,适用于翻译、摘要和情感分析等用例。', displayName: 'Mistral Small', @@ -28,9 +29,9 @@ const Mistral: ModelProviderCard = { input: 0.2, output: 0.6, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Mistral Large是旗舰大模型,擅长多语言任务、复杂推理和代码生成,是高端应用的理想选择。', displayName: 'Mistral Large', @@ -41,9 +42,9 @@ const Mistral: ModelProviderCard = { input: 2, output: 6, }, - tokens: 128_000, }, { + contextWindowTokens: 32_768, description: 'Codestral是专注于代码生成的尖端生成模型,优化了中间填充和代码补全任务。', displayName: 'Codestral', id: 'codestral-latest', @@ -51,9 +52,9 @@ const Mistral: ModelProviderCard = { input: 0.2, output: 0.6, }, - tokens: 32_768, }, { + contextWindowTokens: 128_000, description: 'Pixtral Large 是一款拥有 1240 亿参数的开源多模态模型,基于 Mistral Large 2 构建。这是我们多模态家族中的第二款模型,展现了前沿水平的图像理解能力。', displayName: 'Pixtral Large', @@ -64,10 +65,10 @@ const Mistral: ModelProviderCard = { input: 2, output: 6, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'Pixtral 模型在图表和图理解、文档问答、多模态推理和指令遵循等任务上表现出强大的能力,能够以自然分辨率和宽高比摄入图像,还能够在长达 128K 令牌的长上下文窗口中处理任意数量的图像。', displayName: 'Pixtral 12B', @@ -77,7 +78,6 @@ const Mistral: ModelProviderCard = { input: 0.15, output: 0.15, }, - tokens: 128_000, vision: true, }, { @@ -97,26 +97,27 @@ const Mistral: ModelProviderCard = { { description: 'Ministral 3B 是Mistral的世界顶级边缘模型。', + contextWindowTokens: 128_000, + description: 'Ministral 3B 是Mistral的世界顶级边缘模型。', displayName: 'Ministral 3B', id: 'ministral-3b-latest', pricing: { input: 0.04, output: 0.04, }, - tokens: 128_000, }, { - description: - 'Ministral 8B 是Mistral的性价比极高的边缘模型。', + contextWindowTokens: 128_000, + description: 'Ministral 8B 是Mistral的性价比极高的边缘模型。', displayName: 'Ministral 8B', id: 'ministral-8b-latest', pricing: { input: 0.1, output: 0.1, }, - tokens: 128_000, }, { + contextWindowTokens: 32_768, description: 'Mistral 7B是一款紧凑但高性能的模型,擅长批量处理和简单任务,如分类和文本生成,具有良好的推理能力。', displayName: 'Mistral 7B', @@ -125,9 +126,9 @@ const Mistral: ModelProviderCard = { input: 0.25, output: 0.25, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Mixtral 8x7B是一个稀疏专家模型,利用多个参数提高推理速度,适合处理多语言和代码生成任务。', displayName: 'Mixtral 8x7B', @@ -136,9 +137,9 @@ const Mistral: ModelProviderCard = { input: 0.7, output: 0.7, }, - tokens: 32_768, }, { + contextWindowTokens: 65_536, description: 'Mixtral 8x22B是一个更大的专家模型,专注于复杂任务,提供出色的推理能力和更高的吞吐量。', displayName: 'Mixtral 8x22B', @@ -148,9 +149,9 @@ const Mistral: ModelProviderCard = { input: 2, output: 6, }, - tokens: 65_536, }, { + contextWindowTokens: 256_000, description: 'Codestral Mamba是专注于代码生成的Mamba 2语言模型,为先进的代码和推理任务提供强力支持。', displayName: 'Codestral Mamba', @@ -159,7 +160,6 @@ const Mistral: ModelProviderCard = { input: 0.15, output: 0.15, }, - tokens: 256_000, }, ], checkModel: 'ministral-3b-latest', diff --git a/src/config/modelProviders/moonshot.ts b/src/config/modelProviders/moonshot.ts index 313870d2ece3..f7d8b3fdcbb2 100644 --- a/src/config/modelProviders/moonshot.ts +++ b/src/config/modelProviders/moonshot.ts @@ -4,31 +4,31 @@ import { ModelProviderCard } from '@/types/llm'; const Moonshot: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 8192, description: 'Moonshot V1 8K 专为生成短文本任务设计,具有高效的处理性能,能够处理8,192个tokens,非常适合简短对话、速记和快速内容生成。', displayName: 'Moonshot V1 8K', enabled: true, functionCall: true, id: 'moonshot-v1-8k', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Moonshot V1 32K 提供中等长度的上下文处理能力,能够处理32,768个tokens,特别适合生成各种长文档和复杂对话,应用于内容创作、报告生成和对话系统等领域。', displayName: 'Moonshot V1 32K', enabled: true, functionCall: true, id: 'moonshot-v1-32k', - tokens: 32_768, }, { + contextWindowTokens: 128_000, description: 'Moonshot V1 128K 是一款拥有超长上下文处理能力的模型,适用于生成超长文本,满足复杂的生成任务需求,能够处理多达128,000个tokens的内容,非常适合科研、学术和大型文档生成等应用场景。', displayName: 'Moonshot V1 128K', enabled: true, functionCall: true, id: 'moonshot-v1-128k', - tokens: 128_000, }, ], checkModel: 'moonshot-v1-8k', diff --git a/src/config/modelProviders/novita.ts b/src/config/modelProviders/novita.ts index 461a4dfd9257..1bb2f974ef48 100644 --- a/src/config/modelProviders/novita.ts +++ b/src/config/modelProviders/novita.ts @@ -4,107 +4,107 @@ import { ModelProviderCard } from '@/types/llm'; const Novita: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 8192, description: 'Llama 3.1 8B Instruct 是 Meta 推出的最新版本,优化了高质量对话场景,表现优于许多领先的闭源模型。', displayName: 'Llama 3.1 8B Instruct', enabled: true, id: 'meta-llama/llama-3.1-8b-instruct', - tokens: 8192, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 70B Instruct 专为高质量对话而设计,在人类评估中表现突出,特别适合高交互场景。', displayName: 'Llama 3.1 70B Instruct', enabled: true, id: 'meta-llama/llama-3.1-70b-instruct', - tokens: 131_072, }, { + contextWindowTokens: 32_768, description: 'Llama 3.1 405B Instruct 是 Meta最新推出的版本,优化用于生成高质量对话,超越了许多领导闭源模型。', displayName: 'Llama 3.1 405B Instruct', enabled: true, id: 'meta-llama/llama-3.1-405b-instruct', - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'Llama 3 8B Instruct 优化了高质量对话场景,性能优于许多闭源模型。', displayName: 'Llama 3 8B Instruct', id: 'meta-llama/llama-3-8b-instruct', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 70B Instruct 优化用于高质量对话场景,在各类人类评估中表现优异。', displayName: 'Llama 3 70B Instruct', id: 'meta-llama/llama-3-70b-instruct', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma 2 9B 是谷歌的一款开源语言模型,以其在效率和性能方面设立了新的标准。', displayName: 'Gemma 2 9B', enabled: true, id: 'google/gemma-2-9b-it', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Mistral Nemo 是多语言支持和高性能编程的7.3B参数模型。', displayName: 'Mistral Nemo', enabled: true, id: 'mistralai/mistral-nemo', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Mistral 7B Instruct 是一款兼有速度优化和长上下文支持的高性能行业标准模型。', displayName: 'Mistral 7B Instruct', enabled: true, id: 'mistralai/mistral-7b-instruct', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'WizardLM 2 7B 是微软AI最新的快速轻量化模型,性能接近于现有开源领导模型的10倍。', displayName: 'WizardLM 2 7B', enabled: true, id: 'microsoft/wizardlm 2-7b', - tokens: 32_768, }, { + contextWindowTokens: 65_535, description: 'WizardLM-2 8x22B 是微软AI最先进的Wizard模型,显示出极其竞争力的表现。', displayName: 'WizardLM-2 8x22B', enabled: true, id: 'microsoft/wizardlm-2-8x22b', - tokens: 65_535, }, { + contextWindowTokens: 16_000, description: 'Dolphin Mixtral 8x22B 是一款为指令遵循、对话和编程设计的模型。', displayName: 'Dolphin Mixtral 8x22B', id: 'cognitivecomputations/dolphin-mixtral-8x22b', - tokens: 16_000, }, { + contextWindowTokens: 8192, description: 'Hermes 2 Pro Llama 3 8B 是 Nous Hermes 2的升级版本,包含最新的内部开发的数据集。', displayName: 'Hermes 2 Pro Llama 3 8B', id: 'nousresearch/hermes-2-pro-llama-3-8b', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Hermes 2 Mixtral 8x7B DPO 是一款高度灵活的多模型合并,旨在提供卓越的创造性体验。', displayName: 'Hermes 2 Mixtral 8x7B DPO', id: 'Nous-Hermes-2-Mixtral-8x7B-DPO', - tokens: 32_768, }, { + contextWindowTokens: 4096, description: 'MythoMax l2 13B 是一款合并了多个顶尖模型的创意与智能相结合的语言模型。', displayName: 'MythoMax l2 13B', id: 'gryphe/mythomax-l2-13b', - tokens: 4096, }, { + contextWindowTokens: 4096, description: 'OpenChat 7B 是经过“C-RLFT(条件强化学习微调)”策略精调的开源语言模型库。', displayName: 'OpenChat 7B', id: 'openchat/openchat-7b', - tokens: 4096, }, ], checkModel: 'meta-llama/llama-3.1-8b-instruct', diff --git a/src/config/modelProviders/ollama.ts b/src/config/modelProviders/ollama.ts index 457765bbb030..ea7217c1b57e 100644 --- a/src/config/modelProviders/ollama.ts +++ b/src/config/modelProviders/ollama.ts @@ -3,334 +3,334 @@ import { ModelProviderCard } from '@/types/llm'; const Ollama: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 128_000, description: 'Llama 3.1 是 Meta 推出的领先模型,支持高达 405B 参数,可应用于复杂对话、多语言翻译和数据分析领域。', displayName: 'Llama 3.1 8B', enabled: true, functionCall: true, id: 'llama3.1', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Llama 3.1 是 Meta 推出的领先模型,支持高达 405B 参数,可应用于复杂对话、多语言翻译和数据分析领域。', displayName: 'Llama 3.1 70B', id: 'llama3.1:70b', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Llama 3.1 是 Meta 推出的领先模型,支持高达 405B 参数,可应用于复杂对话、多语言翻译和数据分析领域。', displayName: 'Llama 3.1 405B', id: 'llama3.1:405b', - tokens: 128_000, }, { + contextWindowTokens: 16_384, description: 'Code Llama 是一款专注于代码生成和讨论的 LLM,结合广泛的编程语言支持,适用于开发者环境。', displayName: 'Code Llama 7B', enabled: true, id: 'codellama', - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: 'Code Llama 是一款专注于代码生成和讨论的 LLM,结合广泛的编程语言支持,适用于开发者环境。', displayName: 'Code Llama 13B', id: 'codellama:13b', - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: 'Code Llama 是一款专注于代码生成和讨论的 LLM,结合广泛的编程语言支持,适用于开发者环境。', displayName: 'Code Llama 34B', id: 'codellama:34b', - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: 'Code Llama 是一款专注于代码生成和讨论的 LLM,结合广泛的编程语言支持,适用于开发者环境。', displayName: 'Code Llama 70B', id: 'codellama:70b', - tokens: 16_384, }, { + contextWindowTokens: 128_000, description: 'QwQ 是一个实验研究模型,专注于提高 AI 推理能力。', displayName: 'QwQ 32B', enabled: true, functionCall: true, id: 'qwq', releasedAt: '2024-11-28', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Qwen2.5 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2.5 0.5B', id: 'qwen2.5:0.5b', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Qwen2.5 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2.5 1.5B', id: 'qwen2.5:1.5b', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Qwen2.5 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2.5 7B', enabled: true, functionCall: true, id: 'qwen2.5', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Qwen2.5 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2.5 72B', id: 'qwen2.5:72b', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Qwen2.5 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2.5 7B', enabled: true, functionCall: true, id: 'qwen2.5', - tokens: 128_000, }, { + contextWindowTokens: 65_536, description: 'CodeQwen1.5 是基于大量代码数据训练的大型语言模型,专为解决复杂编程任务。', displayName: 'CodeQwen1.5 7B', functionCall: true, id: 'codeqwen', - tokens: 65_536, }, { + contextWindowTokens: 128_000, description: 'Qwen2 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2 0.5B', functionCall: true, id: 'qwen2:0.5b', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Qwen2 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2 1.5B', functionCall: true, id: 'qwen2:1.5b', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Qwen2 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2 7B', functionCall: true, id: 'qwen2', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Qwen2 是阿里巴巴的新一代大规模语言模型,以优异的性能支持多元化的应用需求。', displayName: 'Qwen2 72B', functionCall: true, id: 'qwen2:72b', - tokens: 128_000, }, { + contextWindowTokens: 8192, description: 'Gemma 2 是 Google 推出的高效模型,涵盖从小型应用到复杂数据处理的多种应用场景。', displayName: 'Gemma 2 2B', id: 'gemma2:2b', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma 2 是 Google 推出的高效模型,涵盖从小型应用到复杂数据处理的多种应用场景。', displayName: 'Gemma 2 9B', id: 'gemma2', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma 2 是 Google 推出的高效模型,涵盖从小型应用到复杂数据处理的多种应用场景。', displayName: 'Gemma 2 27B', id: 'gemma2:27b', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'CodeGemma 专用于不同编程任务的轻量级语言模型,支持快速迭代和集成。', displayName: 'CodeGemma 2B', id: 'codegemma:2b', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'CodeGemma 专用于不同编程任务的轻量级语言模型,支持快速迭代和集成。', displayName: 'CodeGemma 7B', id: 'codegemma', - tokens: 8192, }, { + contextWindowTokens: 128_000, description: 'Phi-3 是微软推出的轻量级开放模型,适用于高效集成和大规模知识推理。', displayName: 'Phi-3 3.8B', enabled: true, id: 'phi3', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Phi-3 是微软推出的轻量级开放模型,适用于高效集成和大规模知识推理。', displayName: 'Phi-3 14B', id: 'phi3:14b', - tokens: 128_000, }, { + contextWindowTokens: 32_768, description: 'WizardLM 2 是微软AI提供的语言模型,在复杂对话、多语言、推理和智能助手领域表现尤为出色。', displayName: 'WizardLM 2 7B', enabled: true, id: 'wizardlm2', - tokens: 32_768, }, { + contextWindowTokens: 65_536, description: 'WizardLM 2 是微软AI提供的语言模型,在复杂对话、多语言、推理和智能助手领域表现尤为出色。', displayName: 'WizardLM 2 8x22B', id: 'wizardlm2:8x22b', - tokens: 65_536, }, { + contextWindowTokens: 32_768, description: 'MathΣtral 专为科学研究和数学推理设计,提供有效的计算能力和结果解释。', displayName: 'MathΣtral 7B', enabled: true, id: 'mathstral', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Mistral 是 Mistral AI 发布的 7B 模型,适合多变的语言处理需求。', displayName: 'Mistral 7B', enabled: true, functionCall: true, id: 'mistral', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Mixtral 是 Mistral AI 的专家模型,具有开源权重,并在代码生成和语言理解方面提供支持。', displayName: 'Mixtral 8x7B', enabled: true, functionCall: true, id: 'mixtral', - tokens: 32_768, }, { + contextWindowTokens: 65_536, description: 'Mixtral 是 Mistral AI 的专家模型,具有开源权重,并在代码生成和语言理解方面提供支持。', displayName: 'Mixtral 8x22B', functionCall: true, id: 'mixtral:8x22b', - tokens: 65_536, }, { + contextWindowTokens: 128_000, description: 'Mixtral Large 是 Mistral 的旗舰模型,结合代码生成、数学和推理的能力,支持 128k 上下文窗口。', displayName: 'Mixtral Large 123B', enabled: true, id: 'mistral-large', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Mistral Nemo 由 Mistral AI 和 NVIDIA 合作推出,是高效性能的 12B 模型。', displayName: 'Mixtral Nemo 12B', enabled: true, functionCall: true, id: 'mistral-nemo', - tokens: 128_000, }, { + contextWindowTokens: 32_768, description: 'Codestral 是 Mistral AI 的首款代码模型,为代码生成任务提供优异支持。', displayName: 'Codestral 22B', enabled: true, id: 'codestral', - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'Aya 23 是 Cohere 推出的多语言模型,支持 23 种语言,为多元化语言应用提供便利。', displayName: 'Aya 23 8B', enabled: true, id: 'aya', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Aya 23 是 Cohere 推出的多语言模型,支持 23 种语言,为多元化语言应用提供便利。', displayName: 'Aya 23 35B', id: 'aya:35b', - tokens: 8192, }, { + contextWindowTokens: 131_072, description: 'Command R 是优化用于对话和长上下文任务的LLM,特别适合动态交互与知识管理。', displayName: 'Command R 35B', enabled: true, functionCall: true, id: 'command-r', - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Command R+ 是一款高性能的大型语言模型,专为真实企业场景和复杂应用而设计。', displayName: 'Command R+ 104B', enabled: true, functionCall: true, id: 'command-r-plus', - tokens: 131_072, }, { + contextWindowTokens: 32_768, description: 'DeepSeek V2 是高效的 Mixture-of-Experts 语言模型,适用于经济高效的处理需求。', displayName: 'DeepSeek V2 16B', enabled: true, id: 'deepseek-v2', - tokens: 32_768, }, { + contextWindowTokens: 128_000, description: 'DeepSeek V2 236B 是 DeepSeek 的设计代码模型,提供强大的代码生成能力。', displayName: 'DeepSeek V2 236B', id: 'deepseek-v2:236b', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'DeepSeek Coder V2 是开源的混合专家代码模型,在代码任务方面表现优异,与 GPT4-Turbo 相媲美。', displayName: 'DeepSeek Coder V2 16B', enabled: true, id: 'deepseek-coder-v2', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'DeepSeek Coder V2 是开源的混合专家代码模型,在代码任务方面表现优异,与 GPT4-Turbo 相媲美。', displayName: 'DeepSeek Coder V2 236B', id: 'deepseek-coder-v2:236b', - tokens: 128_000, }, { + contextWindowTokens: 4096, description: 'LLaVA 是结合视觉编码器和 Vicuna 的多模态模型,用于强大的视觉和语言理解。', displayName: 'LLaVA 7B', enabled: true, id: 'llava', - tokens: 4096, vision: true, }, { + contextWindowTokens: 4096, description: 'LLaVA 是结合视觉编码器和 Vicuna 的多模态模型,用于强大的视觉和语言理解。', displayName: 'LLaVA 13B', id: 'llava:13b', - tokens: 4096, vision: true, }, { + contextWindowTokens: 4096, description: 'LLaVA 是结合视觉编码器和 Vicuna 的多模态模型,用于强大的视觉和语言理解。', displayName: 'LLaVA 34B', id: 'llava:34b', - tokens: 4096, vision: true, }, { + contextWindowTokens: 128_000, description: 'MiniCPM-V 是 OpenBMB 推出的新一代多模态大模型,具备卓越的 OCR 识别和多模态理解能力,支持广泛的应用场景。', displayName: 'MiniCPM-V 8B', enabled: true, id: 'minicpm-v', - tokens: 128_000, vision: true, }, ], diff --git a/src/config/modelProviders/openai.ts b/src/config/modelProviders/openai.ts index 8b013e864752..48b40a80c0fd 100644 --- a/src/config/modelProviders/openai.ts +++ b/src/config/modelProviders/openai.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const OpenAI: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 128_000, description: 'o1-mini是一款针对编程、数学和科学应用场景而设计的快速、经济高效的推理模型。该模型具有128K上下文和2023年10月的知识截止日期。', displayName: 'OpenAI o1-mini', @@ -15,9 +16,9 @@ const OpenAI: ModelProviderCard = { output: 12, }, releasedAt: '2024-09-12', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'o1是OpenAI新的推理模型,适用于需要广泛通用知识的复杂任务。该模型具有128K上下文和2023年10月的知识截止日期。', displayName: 'OpenAI o1-preview', @@ -29,9 +30,9 @@ const OpenAI: ModelProviderCard = { output: 60, }, releasedAt: '2024-09-12', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GPT-4o mini是OpenAI在GPT-4 Omni之后推出的最新模型,支持图文输入并输出文本。作为他们最先进的小型模型,它比其他近期的前沿模型便宜很多,并且比GPT-3.5 Turbo便宜超过60%。它保持了最先进的智能,同时具有显著的性价比。GPT-4o mini在MMLU测试中获得了 82% 的得分,目前在聊天偏好上排名高于 GPT-4。', displayName: 'GPT-4o mini', @@ -43,10 +44,10 @@ const OpenAI: ModelProviderCard = { input: 0.15, output: 0.6, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'GPT-4o 1120', @@ -58,10 +59,10 @@ const OpenAI: ModelProviderCard = { output: 10, }, releasedAt: '2024-11-20', - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'GPT-4o', @@ -72,10 +73,10 @@ const OpenAI: ModelProviderCard = { input: 2.5, output: 10, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'GPT-4o 0806', @@ -86,10 +87,10 @@ const OpenAI: ModelProviderCard = { output: 10, }, releasedAt: '2024-08-06', - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'GPT-4o 0513', @@ -100,10 +101,10 @@ const OpenAI: ModelProviderCard = { output: 15, }, releasedAt: '2024-05-13', - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'ChatGPT-4o', @@ -113,10 +114,10 @@ const OpenAI: ModelProviderCard = { input: 5, output: 15, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo', @@ -126,10 +127,10 @@ const OpenAI: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo Vision 0409', @@ -139,10 +140,10 @@ const OpenAI: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo Preview', @@ -152,9 +153,9 @@ const OpenAI: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo Preview 0125', @@ -164,9 +165,9 @@ const OpenAI: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: '最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。', displayName: 'GPT-4 Turbo Preview 1106', @@ -176,9 +177,9 @@ const OpenAI: ModelProviderCard = { input: 10, output: 30, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', displayName: 'GPT-4', @@ -188,9 +189,9 @@ const OpenAI: ModelProviderCard = { input: 30, output: 60, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', displayName: 'GPT-4 0613', @@ -200,11 +201,12 @@ const OpenAI: ModelProviderCard = { input: 30, output: 60, }, - tokens: 8192, }, { + contextWindowTokens: 32_768, description: - 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', // Will be discontinued on June 6, 2025 + 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', + // Will be discontinued on June 6, 2025 displayName: 'GPT-4 32K', functionCall: true, id: 'gpt-4-32k', @@ -212,9 +214,9 @@ const OpenAI: ModelProviderCard = { input: 60, output: 120, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, // Will be discontinued on June 6, 2025 description: 'GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。', @@ -225,9 +227,9 @@ const OpenAI: ModelProviderCard = { input: 60, output: 120, }, - tokens: 32_768, }, { + contextWindowTokens: 16_385, description: 'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', displayName: 'GPT-3.5 Turbo', @@ -237,9 +239,9 @@ const OpenAI: ModelProviderCard = { input: 0.5, output: 1.5, }, - tokens: 16_385, }, { + contextWindowTokens: 16_385, description: 'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', displayName: 'GPT-3.5 Turbo 0125', @@ -249,9 +251,9 @@ const OpenAI: ModelProviderCard = { input: 0.5, output: 1.5, }, - tokens: 16_385, }, { + contextWindowTokens: 16_385, description: 'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', displayName: 'GPT-3.5 Turbo 1106', @@ -261,9 +263,9 @@ const OpenAI: ModelProviderCard = { input: 1, output: 2, }, - tokens: 16_385, }, { + contextWindowTokens: 4096, description: 'GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125', displayName: 'GPT-3.5 Turbo Instruct', @@ -272,7 +274,6 @@ const OpenAI: ModelProviderCard = { input: 1.5, output: 2, }, - tokens: 4096, }, ], checkModel: 'gpt-4o-mini', diff --git a/src/config/modelProviders/openrouter.ts b/src/config/modelProviders/openrouter.ts index 4b5e596a5d54..d0efaa821b37 100644 --- a/src/config/modelProviders/openrouter.ts +++ b/src/config/modelProviders/openrouter.ts @@ -4,16 +4,17 @@ import { ModelProviderCard } from '@/types/llm'; const OpenRouter: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 128_000, description: '根据上下文长度、主题和复杂性,你的请求将发送到 Llama 3 70B Instruct、Claude 3.5 Sonnet(自我调节)或 GPT-4o。', displayName: 'Auto (best for prompt)', enabled: true, functionCall: false, id: 'openrouter/auto', - tokens: 128_000, vision: false, }, { + contextWindowTokens: 128_000, description: 'o1-mini是一款针对编程、数学和科学应用场景而设计的快速、经济高效的推理模型。该模型具有128K上下文和2023年10月的知识截止日期。', displayName: 'OpenAI o1-mini', @@ -25,9 +26,9 @@ const OpenRouter: ModelProviderCard = { output: 12, }, releasedAt: '2024-09-12', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'o1是OpenAI新的推理模型,适用于需要广泛通用知识的复杂任务。该模型具有128K上下文和2023年10月的知识截止日期。', displayName: 'OpenAI o1-preview', @@ -39,9 +40,9 @@ const OpenRouter: ModelProviderCard = { output: 60, }, releasedAt: '2024-09-12', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GPT-4o mini是OpenAI在GPT-4 Omni之后推出的最新模型,支持图文输入并输出文本。作为他们最先进的小型模型,它比其他近期的前沿模型便宜很多,并且比GPT-3.5 Turbo便宜超过60%。它保持了最先进的智能,同时具有显著的性价比。GPT-4o mini在MMLU测试中获得了 82% 的得分,目前在聊天偏好上排名高于 GPT-4。', displayName: 'GPT-4o mini', @@ -53,10 +54,10 @@ const OpenRouter: ModelProviderCard = { input: 0.15, output: 0.6, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 128_000, description: 'ChatGPT-4o 是一款动态模型,实时更新以保持当前最新版本。它结合了强大的语言理解与生成能力,适合于大规模应用场景,包括客户服务、教育和技术支持。', displayName: 'GPT-4o', @@ -67,10 +68,10 @@ const OpenRouter: ModelProviderCard = { input: 2.5, output: 10, }, - tokens: 128_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Haiku 是 Anthropic 的最快且最紧凑的模型,旨在实现近乎即时的响应。它具有快速且准确的定向性能。', displayName: 'Claude 3 Haiku', @@ -85,10 +86,10 @@ const OpenRouter: ModelProviderCard = { writeCacheInput: 0.3125, }, releasedAt: '2024-03-07', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3.5 Sonnet 提供了超越 Opus 的能力和比 Sonnet 更快的速度,同时保持与 Sonnet 相同的价格。Sonnet 特别擅长编程、数据科学、视觉处理、代理任务。', displayName: 'Claude 3.5 Sonnet', @@ -103,10 +104,10 @@ const OpenRouter: ModelProviderCard = { writeCacheInput: 3.75, }, releasedAt: '2024-06-20', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 200_000, description: 'Claude 3 Opus 是 Anthropic 用于处理高度复杂任务的最强大模型。它在性能、智能、流畅性和理解力方面表现卓越。', displayName: 'Claude 3 Opus', @@ -121,10 +122,10 @@ const OpenRouter: ModelProviderCard = { writeCacheInput: 18.75, }, releasedAt: '2024-02-29', - tokens: 200_000, vision: true, }, { + contextWindowTokens: 1_000_000 + 8192, description: 'Gemini 1.5 Flash 提供了优化后的多模态处理能力,适用多种复杂任务场景。', displayName: 'Gemini 1.5 Flash', enabled: true, @@ -135,10 +136,10 @@ const OpenRouter: ModelProviderCard = { input: 0.075, output: 0.3, }, - tokens: 1_000_000 + 8192, vision: true, }, { + contextWindowTokens: 2_000_000 + 8192, description: 'Gemini 1.5 Pro 结合最新优化技术,带来更高效的多模态数据处理能力。', displayName: 'Gemini 1.5 Pro', enabled: true, @@ -149,10 +150,10 @@ const OpenRouter: ModelProviderCard = { input: 3.5, output: 10.5, }, - tokens: 2_000_000 + 8192, vision: true, }, { + contextWindowTokens: 128_000, description: '融合通用与代码能力的全新开源模型, 不仅保留了原有 Chat 模型的通用对话能力和 Coder 模型的强大代码处理能力,还更好地对齐了人类偏好。此外,DeepSeek-V2.5 在写作任务、指令跟随等多个方面也实现了大幅提升。', displayName: 'DeepSeek V2.5', @@ -164,10 +165,11 @@ const OpenRouter: ModelProviderCard = { output: 0.28, }, releasedAt: '2024-09-05', - tokens: 128_000, }, { - description: 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', + contextWindowTokens: 131_072, + description: + 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 11B Vision', enabled: true, id: 'meta-llama/llama-3.2-11b-vision-instruct', @@ -175,11 +177,12 @@ const OpenRouter: ModelProviderCard = { input: 0.162, output: 0.162, }, - tokens: 131_072, vision: true, }, { - description: 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', + contextWindowTokens: 131_072, + description: + 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 90B Vision', enabled: true, id: 'meta-llama/llama-3.2-90b-vision-instruct', @@ -187,29 +190,28 @@ const OpenRouter: ModelProviderCard = { input: 0.4, output: 0.4, }, - tokens: 131_072, vision: true, }, { + contextWindowTokens: 32_768, description: 'Qwen2 是全新的大型语言模型系列,具有更强的理解和生成能力。', displayName: 'Qwen2 7B (Free)', enabled: true, id: 'qwen/qwen-2-7b-instruct:free', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'LLaMA 3.1 提供多语言支持,是业界领先的生成模型之一。', displayName: 'Llama 3.1 8B (Free)', enabled: true, id: 'meta-llama/llama-3.1-8b-instruct:free', - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'Gemma 2 是Google轻量化的开源文本模型系列。', displayName: 'Gemma 2 9B (Free)', enabled: true, id: 'google/gemma-2-9b-it:free', - tokens: 8192, }, ], checkModel: 'google/gemma-2-9b-it:free', diff --git a/src/config/modelProviders/perplexity.ts b/src/config/modelProviders/perplexity.ts index 5c05f7734e1a..5cd434e9c9d2 100644 --- a/src/config/modelProviders/perplexity.ts +++ b/src/config/modelProviders/perplexity.ts @@ -4,56 +4,56 @@ import { ModelProviderCard } from '@/types/llm'; const Perplexity: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 128_000, description: 'Llama 3.1 Sonar Small Online 模型,具备8B参数,支持约127,000个标记的上下文长度,专为在线聊天设计,能高效处理各种文本交互。', displayName: 'Llama 3.1 Sonar Small Online', enabled: true, id: 'llama-3.1-sonar-small-128k-online', - tokens: 127_072, }, { + contextWindowTokens: 128_000, description: 'Llama 3.1 Sonar Large Online 模型,具备70B参数,支持约127,000个标记的上下文长度,适用于高容量和多样化聊天任务。', displayName: 'Llama 3.1 Sonar Large Online', enabled: true, id: 'llama-3.1-sonar-large-128k-online', - tokens: 127_072, }, { + contextWindowTokens: 128_000, description: 'Llama 3.1 Sonar Huge Online 模型,具备405B参数,支持约127,000个标记的上下文长度,设计用于复杂的在线聊天应用。', displayName: 'Llama 3.1 Sonar Huge Online', enabled: true, id: 'llama-3.1-sonar-huge-128k-online', - tokens: 127_072, }, { + contextWindowTokens: 128_000, description: 'Llama 3.1 Sonar Small Chat 模型,具备8B参数,专为离线聊天设计,支持约127,000个标记的上下文长度。', displayName: 'Llama 3.1 Sonar Small Chat', enabled: true, id: 'llama-3.1-sonar-small-128k-chat', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Llama 3.1 Sonar Large Chat 模型,具备70B参数,支持约127,000个标记的上下文长度,适合于复杂的离线聊天任务。', displayName: 'Llama 3.1 Sonar Large Chat', enabled: true, id: 'llama-3.1-sonar-large-128k-chat', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Llama 3.1 8B Instruct 模型,具备8B参数,支持画面指示任务的高效执行,提供优质的文本生成能力。', id: 'llama-3.1-8b-instruct', - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'Llama 3.1 70B Instruct 模型,具备70B参数,能在大型文本生成和指示任务中提供卓越性能。', id: 'llama-3.1-70b-instruct', - tokens: 128_000, }, ], checkModel: 'llama-3.1-8b-instruct', diff --git a/src/config/modelProviders/qwen.ts b/src/config/modelProviders/qwen.ts index 81ee094f0eba..b508b39599a1 100644 --- a/src/config/modelProviders/qwen.ts +++ b/src/config/modelProviders/qwen.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const Qwen: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 131_072, description: '通义千问超大规模语言模型,支持中文、英文等不同语言输入。', displayName: 'Qwen Turbo', enabled: true, @@ -14,9 +15,9 @@ const Qwen: ModelProviderCard = { input: 0.3, output: 0.6, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问超大规模语言模型增强版,支持中文、英文等不同语言输入。', displayName: 'Qwen Plus', enabled: true, @@ -27,9 +28,9 @@ const Qwen: ModelProviderCard = { input: 0.8, output: 2, }, - tokens: 131_072, }, { + contextWindowTokens: 32_768, description: '通义千问千亿级别超大规模语言模型,支持中文、英文等不同语言输入,当前通义千问2.5产品版本背后的API模型。', displayName: 'Qwen Max', @@ -41,9 +42,9 @@ const Qwen: ModelProviderCard = { input: 20, output: 60, }, - tokens: 32_768, }, { + contextWindowTokens: 1_000_000, description: '通义千问超大规模语言模型,支持长文本上下文,以及基于长文档、多文档等多个场景的对话功能。', displayName: 'Qwen Long', @@ -53,9 +54,9 @@ const Qwen: ModelProviderCard = { input: 0.5, output: 2, }, - tokens: 1_000_000, }, { + contextWindowTokens: 32_000, description: '通义千问大规模视觉语言模型增强版。大幅提升细节识别能力和文字识别能力,支持超百万像素分辨率和任意长宽比规格的图像。', displayName: 'Qwen VL Plus', @@ -66,10 +67,10 @@ const Qwen: ModelProviderCard = { input: 8, output: 8, }, - tokens: 32_000, vision: true, }, { + contextWindowTokens: 32_000, description: '通义千问超大规模视觉语言模型。相比增强版,再次提升视觉推理能力和指令遵循能力,提供更高的视觉感知和认知水平。', displayName: 'Qwen VL Max', @@ -80,12 +81,11 @@ const Qwen: ModelProviderCard = { input: 20, output: 20, }, - tokens: 32_000, vision: true, }, { - description: - '通义千问数学模型是专门用于数学解题的语言模型。', + contextWindowTokens: 4096, + description: '通义千问数学模型是专门用于数学解题的语言模型。', displayName: 'Qwen Math Turbo', id: 'qwen-math-turbo-latest', pricing: { @@ -93,11 +93,10 @@ const Qwen: ModelProviderCard = { input: 2, output: 6, }, - tokens: 4096, }, { - description: - '通义千问数学模型是专门用于数学解题的语言模型。', + contextWindowTokens: 4096, + description: '通义千问数学模型是专门用于数学解题的语言模型。', displayName: 'Qwen Math Plus', id: 'qwen-math-plus-latest', pricing: { @@ -105,9 +104,9 @@ const Qwen: ModelProviderCard = { input: 4, output: 12, }, - tokens: 4096, }, { + contextWindowTokens: 131_072, description: '通义千问代码模型。', displayName: 'Qwen Coder Turbo', id: 'qwen-coder-turbo-latest', @@ -116,9 +115,9 @@ const Qwen: ModelProviderCard = { input: 2, output: 6, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问代码模型。', displayName: 'Qwen Coder Plus', id: 'qwen-coder-plus-latest', @@ -127,9 +126,9 @@ const Qwen: ModelProviderCard = { input: 3.5, output: 7, }, - tokens: 131_072, }, { + contextWindowTokens: 32_768, description: 'QwQ模型是由 Qwen 团队开发的实验性研究模型,专注于增强 AI 推理能力。', displayName: 'QwQ 32B Preview', id: 'qwq-32b-preview', @@ -138,9 +137,9 @@ const Qwen: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { + contextWindowTokens: 131_072, description: '通义千问2.5对外开源的7B规模的模型。', displayName: 'Qwen2.5 7B', functionCall: true, @@ -150,9 +149,9 @@ const Qwen: ModelProviderCard = { input: 1, output: 2, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问2.5对外开源的14B规模的模型。', displayName: 'Qwen2.5 14B', functionCall: true, @@ -162,9 +161,9 @@ const Qwen: ModelProviderCard = { input: 2, output: 6, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问2.5对外开源的32B规模的模型。', displayName: 'Qwen2.5 32B', functionCall: true, @@ -174,9 +173,9 @@ const Qwen: ModelProviderCard = { input: 3.5, output: 7, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问2.5对外开源的72B规模的模型。', displayName: 'Qwen2.5 72B', functionCall: true, @@ -186,9 +185,9 @@ const Qwen: ModelProviderCard = { input: 4, output: 12, }, - tokens: 131_072, }, { + contextWindowTokens: 4096, description: 'Qwen-Math 模型具有强大的数学解题能力。', displayName: 'Qwen2.5 Math 7B', id: 'qwen2.5-math-7b-instruct', @@ -197,9 +196,9 @@ const Qwen: ModelProviderCard = { input: 1, output: 2, }, - tokens: 4096, }, { + contextWindowTokens: 4096, description: 'Qwen-Math 模型具有强大的数学解题能力。', displayName: 'Qwen2.5 Math 72B', id: 'qwen2.5-math-72b-instruct', @@ -208,9 +207,9 @@ const Qwen: ModelProviderCard = { input: 4, output: 12, }, - tokens: 4096, }, { + contextWindowTokens: 131_072, description: '通义千问代码模型开源版。', displayName: 'Qwen2.5 Coder 7B', id: 'qwen2.5-coder-7b-instruct', @@ -219,9 +218,9 @@ const Qwen: ModelProviderCard = { input: 1, output: 2, }, - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: '通义千问代码模型开源版。', displayName: 'Qwen2.5 Coder 32B', id: 'qwen2.5-coder-32b-instruct', @@ -230,9 +229,9 @@ const Qwen: ModelProviderCard = { input: 3.5, output: 7, }, - tokens: 131_072, }, { + contextWindowTokens: 8000, description: '以 Qwen-7B 语言模型初始化,添加图像模型,图像输入分辨率为448的预训练模型。', displayName: 'Qwen VL', id: 'qwen-vl-v1', @@ -241,10 +240,10 @@ const Qwen: ModelProviderCard = { input: 0, output: 0, }, - tokens: 8000, vision: true, }, { + contextWindowTokens: 8000, description: '通义千问VL支持灵活的交互方式,包括多图、多轮问答、创作等能力的模型。', displayName: 'Qwen VL Chat', id: 'qwen-vl-chat-v1', @@ -253,7 +252,6 @@ const Qwen: ModelProviderCard = { input: 0, output: 0, }, - tokens: 8000, vision: true, }, ], diff --git a/src/config/modelProviders/sensenova.ts b/src/config/modelProviders/sensenova.ts index ebfb3f29dd37..f0c21d18084f 100644 --- a/src/config/modelProviders/sensenova.ts +++ b/src/config/modelProviders/sensenova.ts @@ -5,6 +5,7 @@ import { ModelProviderCard } from '@/types/llm'; const SenseNova: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 131_072, description: '最新版本模型 (V5.5),128K上下文长度,在数学推理、英文对话、指令跟随以及长文本理解等领域能力显著提升,比肩GPT-4o', displayName: 'SenseChat 5.5', @@ -16,7 +17,6 @@ const SenseNova: ModelProviderCard = { input: 40, output: 100, }, - tokens: 131_072, }, /* // Not compatible with local mode (Not support Base64 Image) @@ -35,6 +35,7 @@ const SenseNova: ModelProviderCard = { }, */ { + contextWindowTokens: 32_768, description: '适用于快速问答、模型微调场景', displayName: 'SenseChat 5.0 Turbo', enabled: true, @@ -44,9 +45,9 @@ const SenseNova: ModelProviderCard = { input: 2, output: 5, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: '32K上下文长度,在粤语的对话理解上超越了GPT-4,在知识、推理、数学及代码编写等多个领域均能与GPT-4 Turbo相媲美', displayName: 'SenseChat 5.0 Cantonese', @@ -56,9 +57,9 @@ const SenseNova: ModelProviderCard = { input: 27, output: 27, }, - tokens: 32_768, }, { + contextWindowTokens: 131_072, description: '基础版本模型 (V4),128K上下文长度,在长文本理解及生成等任务中表现出色', displayName: 'SenseChat 4.0 128K', enabled: true, @@ -68,9 +69,9 @@ const SenseNova: ModelProviderCard = { input: 60, output: 60, }, - tokens: 131_072, }, { + contextWindowTokens: 32_768, description: '基础版本模型 (V4),32K上下文长度,灵活应用于各类场景', displayName: 'SenseChat 4.0 32K', enabled: true, @@ -80,9 +81,9 @@ const SenseNova: ModelProviderCard = { input: 36, output: 36, }, - tokens: 32_768, }, { + contextWindowTokens: 4096, description: '基础版本模型 (V4),4K上下文长度,通用能力强大', displayName: 'SenseChat 4.0 4K', enabled: true, @@ -92,9 +93,9 @@ const SenseNova: ModelProviderCard = { input: 12, output: 12, }, - tokens: 4096, }, { + contextWindowTokens: 8192, description: '标准版模型,8K上下文长度,高响应速度', displayName: 'SenseChat Character', id: 'SenseChat-Character', @@ -103,9 +104,9 @@ const SenseNova: ModelProviderCard = { input: 12, output: 12, }, - tokens: 8192, }, { + contextWindowTokens: 32_768, description: '高级版模型,32K上下文长度,能力全面提升,支持中/英文对话', displayName: 'SenseChat Character Pro', id: 'SenseChat-Character-Pro', @@ -114,7 +115,6 @@ const SenseNova: ModelProviderCard = { input: 15, output: 15, }, - tokens: 32_768, }, ], checkModel: 'SenseChat-Turbo', diff --git a/src/config/modelProviders/siliconcloud.ts b/src/config/modelProviders/siliconcloud.ts index 697880b45fb4..996ed80d6bbd 100644 --- a/src/config/modelProviders/siliconcloud.ts +++ b/src/config/modelProviders/siliconcloud.ts @@ -4,7 +4,9 @@ import { ModelProviderCard } from '@/types/llm'; const SiliconCloud: ModelProviderCard = { chatModels: [ { - description: 'Hunyuan-Large 是业界最大的开源 Transformer 架构 MoE 模型,拥有 3890 亿总参数量和 520 亿激活参数量。', + contextWindowTokens: 32_768, + description: + 'Hunyuan-Large 是业界最大的开源 Transformer 架构 MoE 模型,拥有 3890 亿总参数量和 520 亿激活参数量。', displayName: 'Hunyuan A52B Instruct', enabled: true, id: 'Tencent/Hunyuan-A52B-Instruct', @@ -13,10 +15,11 @@ const SiliconCloud: ModelProviderCard = { input: 21, output: 21, }, - tokens: 32_768, }, { - description: 'DeepSeek-V2.5 是 DeepSeek-V2-Chat 和 DeepSeek-Coder-V2-Instruct 的升级版本,集成了两个先前版本的通用和编码能力。该模型在多个方面进行了优化,包括写作和指令跟随能力,更好地与人类偏好保持一致。DeepSeek-V2.5 在各种评估基准上都取得了显著的提升,如 AlpacaEval 2.0、ArenaHard、AlignBench 和 MT-Bench 等', + contextWindowTokens: 32_768, + description: + 'DeepSeek-V2.5 是 DeepSeek-V2-Chat 和 DeepSeek-Coder-V2-Instruct 的升级版本,集成了两个先前版本的通用和编码能力。该模型在多个方面进行了优化,包括写作和指令跟随能力,更好地与人类偏好保持一致。DeepSeek-V2.5 在各种评估基准上都取得了显著的提升,如 AlpacaEval 2.0、ArenaHard、AlignBench 和 MT-Bench 等', displayName: 'DeepSeek V2.5', enabled: true, functionCall: true, @@ -26,10 +29,11 @@ const SiliconCloud: ModelProviderCard = { input: 1.33, output: 1.33, }, - tokens: 32_768, }, { - description: 'DeepSeek-V2 是一个强大、经济高效的混合专家(MoE)语言模型。它在 8.1 万亿个 token 的高质量语料库上进行了预训练,并通过监督微调(SFT)和强化学习(RL)进一步提升了模型能力。与 DeepSeek 67B 相比, DeepSeek-V2 在性能更强的同时,节省了 42.5% 的训练成本,减少了 93.3% 的 KV 缓存,并将最大生成吞吐量提高到了 5.76 倍。该模型支持 128k 的上下文长度,在标准基准测试和开放式生成评估中都表现出色', + contextWindowTokens: 32_768, + description: + 'DeepSeek-V2 是一个强大、经济高效的混合专家(MoE)语言模型。它在 8.1 万亿个 token 的高质量语料库上进行了预训练,并通过监督微调(SFT)和强化学习(RL)进一步提升了模型能力。与 DeepSeek 67B 相比, DeepSeek-V2 在性能更强的同时,节省了 42.5% 的训练成本,减少了 93.3% 的 KV 缓存,并将最大生成吞吐量提高到了 5.76 倍。该模型支持 128k 的上下文长度,在标准基准测试和开放式生成评估中都表现出色', displayName: 'DeepSeek V2 Chat', id: 'deepseek-ai/DeepSeek-V2-Chat', pricing: { @@ -37,10 +41,11 @@ const SiliconCloud: ModelProviderCard = { input: 1.33, output: 1.33, }, - tokens: 32_768, }, { - description: 'QwQ-32B-Preview是Qwen 最新的实验性研究模型,专注于提升AI推理能力。通过探索语言混合、递归推理等复杂机制,主要优势包括强大的推理分析能力、数学和编程能力。与此同时,也存在语言切换问题、推理循环、安全性考虑、其他能力方面的差异。', + contextWindowTokens: 32_768, + description: + 'QwQ-32B-Preview是Qwen 最新的实验性研究模型,专注于提升AI推理能力。通过探索语言混合、递归推理等复杂机制,主要优势包括强大的推理分析能力、数学和编程能力。与此同时,也存在语言切换问题、推理循环、安全性考虑、其他能力方面的差异。', displayName: 'QwQ 32B Preview', enabled: true, id: 'Qwen/QwQ-32B-Preview', @@ -49,10 +54,11 @@ const SiliconCloud: ModelProviderCard = { input: 1.26, output: 1.26, }, - tokens: 32_768, }, { - description: 'Qwen2.5-7B-Instruct 是阿里云发布的最新大语言模型系列之一。该 7B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-7B-Instruct 是阿里云发布的最新大语言模型系列之一。该 7B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 7B Instruct (Free)', enabled: true, functionCall: true, @@ -62,10 +68,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { - description: 'Qwen2.5-7B-Instruct 是阿里云发布的最新大语言模型系列之一。该 7B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-7B-Instruct 是阿里云发布的最新大语言模型系列之一。该 7B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 7B Instruct (LoRA)', id: 'LoRA/Qwen/Qwen2.5-7B-Instruct', pricing: { @@ -73,10 +80,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.53, output: 0.53, }, - tokens: 32_768, }, { - description: 'Qwen2.5-7B-Instruct 是阿里云发布的最新大语言模型系列之一。该 7B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-7B-Instruct 是阿里云发布的最新大语言模型系列之一。该 7B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 7B Instruct (Pro)', functionCall: true, id: 'Pro/Qwen/Qwen2.5-7B-Instruct', @@ -85,10 +93,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.35, output: 0.35, }, - tokens: 32_768, }, { - description: 'Qwen2.5-14B-Instruct 是阿里云发布的最新大语言模型系列之一。该 14B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-14B-Instruct 是阿里云发布的最新大语言模型系列之一。该 14B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 14B Instruct', functionCall: true, id: 'Qwen/Qwen2.5-14B-Instruct', @@ -97,10 +106,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.7, output: 0.7, }, - tokens: 32_768, }, { - description: 'Qwen2.5-32B-Instruct 是阿里云发布的最新大语言模型系列之一。该 32B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-32B-Instruct 是阿里云发布的最新大语言模型系列之一。该 32B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 32B Instruct', functionCall: true, id: 'Qwen/Qwen2.5-32B-Instruct', @@ -109,10 +119,11 @@ const SiliconCloud: ModelProviderCard = { input: 1.26, output: 1.26, }, - tokens: 32_768, }, { - description: 'Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 72B Instruct', functionCall: true, id: 'Qwen/Qwen2.5-72B-Instruct', @@ -121,10 +132,11 @@ const SiliconCloud: ModelProviderCard = { input: 4.13, output: 4.13, }, - tokens: 32_768, }, { - description: 'Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 72B Instruct (LoRA)', id: 'LoRA/Qwen/Qwen2.5-72B-Instruct', pricing: { @@ -132,10 +144,11 @@ const SiliconCloud: ModelProviderCard = { input: 6.2, output: 6.2, }, - tokens: 32_768, }, { - description: 'Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 72B Instruct (Vendor-A)', functionCall: true, id: 'Vendor-A/Qwen/Qwen2.5-72B-Instruct', @@ -144,10 +157,11 @@ const SiliconCloud: ModelProviderCard = { input: 1, output: 1, }, - tokens: 32_768, }, { - description: 'Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。它支持长达 128K tokens 的输入,可以生成超过 8K tokens 的长文本。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', + contextWindowTokens: 131_072, + description: + 'Qwen2.5-72B-Instruct 是阿里云发布的最新大语言模型系列之一。该 72B 模型在编码和数学等领域具有显著改进的能力。它支持长达 128K tokens 的输入,可以生成超过 8K tokens 的长文本。该模型还提供了多语言支持,覆盖超过 29 种语言,包括中文、英文等。模型在指令跟随、理解结构化数据以及生成结构化输出(尤其是 JSON)方面都有显著提升', displayName: 'Qwen2.5 72B Instruct 128K', enabled: true, functionCall: true, @@ -157,10 +171,11 @@ const SiliconCloud: ModelProviderCard = { input: 4.13, output: 4.13, }, - tokens: 131_072, }, { - description: 'Qwen2.5-Coder-7B-Instruct 是阿里云发布的代码特定大语言模型系列的最新版本。该模型在 Qwen2.5 的基础上,通过 5.5 万亿个 tokens 的训练,显著提升了代码生成、推理和修复能力。它不仅增强了编码能力,还保持了数学和通用能力的优势。模型为代码智能体等实际应用提供了更全面的基础', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-Coder-7B-Instruct 是阿里云发布的代码特定大语言模型系列的最新版本。该模型在 Qwen2.5 的基础上,通过 5.5 万亿个 tokens 的训练,显著提升了代码生成、推理和修复能力。它不仅增强了编码能力,还保持了数学和通用能力的优势。模型为代码智能体等实际应用提供了更全面的基础', displayName: 'Qwen2.5 Coder 7B Instruct (Free)', id: 'Qwen/Qwen2.5-Coder-7B-Instruct', pricing: { @@ -168,10 +183,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { - description: 'Qwen2.5-Coder-7B-Instruct 是阿里云发布的代码特定大语言模型系列的最新版本。该模型在 Qwen2.5 的基础上,通过 5.5 万亿个 tokens 的训练,显著提升了代码生成、推理和修复能力。它不仅增强了编码能力,还保持了数学和通用能力的优势。模型为代码智能体等实际应用提供了更全面的基础', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-Coder-7B-Instruct 是阿里云发布的代码特定大语言模型系列的最新版本。该模型在 Qwen2.5 的基础上,通过 5.5 万亿个 tokens 的训练,显著提升了代码生成、推理和修复能力。它不仅增强了编码能力,还保持了数学和通用能力的优势。模型为代码智能体等实际应用提供了更全面的基础', displayName: 'Qwen2.5 Coder 7B Instruct (Pro)', id: 'Pro/Qwen/Qwen2.5-Coder-7B-Instruct', pricing: { @@ -179,10 +195,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.35, output: 0.35, }, - tokens: 32_768, }, { - description: 'Qwen2.5-Coder-32B-Instruct 是基于 Qwen2.5 开发的代码特定大语言模型。该模型通过 5.5 万亿 tokens 的训练,在代码生成、代码推理和代码修复方面都取得了显著提升。它是当前最先进的开源代码语言模型,编码能力可与 GPT-4 相媲美。模型不仅增强了编码能力,还保持了在数学和通用能力方面的优势,并支持长文本处理', + contextWindowTokens: 32_768, + description: + 'Qwen2.5-Coder-32B-Instruct 是基于 Qwen2.5 开发的代码特定大语言模型。该模型通过 5.5 万亿 tokens 的训练,在代码生成、代码推理和代码修复方面都取得了显著提升。它是当前最先进的开源代码语言模型,编码能力可与 GPT-4 相媲美。模型不仅增强了编码能力,还保持了在数学和通用能力方面的优势,并支持长文本处理', displayName: 'Qwen2.5 Coder 32B Instruct', id: 'Qwen/Qwen2.5-Coder-32B-Instruct', pricing: { @@ -190,10 +207,11 @@ const SiliconCloud: ModelProviderCard = { input: 1.26, output: 1.26, }, - tokens: 32_768, }, { - description: 'Qwen2.5-Math-72B 是阿里云发布的 Qwen2.5-Math 系列数学大语言模型之一。该模型支持使用思维链(CoT)和工具集成推理(TIR)方法解决中文和英文数学问题。相比前代 Qwen2-Math 系列,Qwen2.5-Math 系列在中英文数学基准测试中取得了显著的性能提升。该模型在处理精确计算、符号操作和算法操作方面表现出色,尤其适合解决复杂的数学和算法推理任务', + contextWindowTokens: 4096, + description: + 'Qwen2.5-Math-72B 是阿里云发布的 Qwen2.5-Math 系列数学大语言模型之一。该模型支持使用思维链(CoT)和工具集成推理(TIR)方法解决中文和英文数学问题。相比前代 Qwen2-Math 系列,Qwen2.5-Math 系列在中英文数学基准测试中取得了显著的性能提升。该模型在处理精确计算、符号操作和算法操作方面表现出色,尤其适合解决复杂的数学和算法推理任务', displayName: 'Qwen2.5 Math 72B Instruct', id: 'Qwen/Qwen2.5-Math-72B-Instruct', pricing: { @@ -201,10 +219,11 @@ const SiliconCloud: ModelProviderCard = { input: 4.13, output: 4.13, }, - tokens: 4096, }, { - description: 'Qwen2-1.5B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 1.5B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型。与 Qwen1.5-1.8B-Chat 相比,Qwen2-1.5B-Instruct 在 MMLU、HumanEval、GSM8K、C-Eval 和 IFEval 等测试中均显示出显著的性能提升,尽管参数量略少', + contextWindowTokens: 32_768, + description: + 'Qwen2-1.5B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 1.5B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型。与 Qwen1.5-1.8B-Chat 相比,Qwen2-1.5B-Instruct 在 MMLU、HumanEval、GSM8K、C-Eval 和 IFEval 等测试中均显示出显著的性能提升,尽管参数量略少', displayName: 'Qwen2 1.5B Instruct (Free)', id: 'Qwen/Qwen2-1.5B-Instruct', pricing: { @@ -212,10 +231,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { - description: 'Qwen2-1.5B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 1.5B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型。与 Qwen1.5-1.8B-Chat 相比,Qwen2-1.5B-Instruct 在 MMLU、HumanEval、GSM8K、C-Eval 和 IFEval 等测试中均显示出显著的性能提升,尽管参数量略少', + contextWindowTokens: 32_768, + description: + 'Qwen2-1.5B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 1.5B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型。与 Qwen1.5-1.8B-Chat 相比,Qwen2-1.5B-Instruct 在 MMLU、HumanEval、GSM8K、C-Eval 和 IFEval 等测试中均显示出显著的性能提升,尽管参数量略少', displayName: 'Qwen2 1.5B Instruct (Pro)', id: 'Pro/Qwen/Qwen2-1.5B-Instruct', pricing: { @@ -223,10 +243,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.14, output: 0.14, }, - tokens: 32_768, }, { - description: 'Qwen2-7B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 7B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它能够处理大规模输入。该模型在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型,并在某些任务上展现出与专有模型相当的竞争力。Qwen2-7B-Instruct 在多项评测中均优于 Qwen1.5-7B-Chat,显示出显著的性能提升', + contextWindowTokens: 32_768, + description: + 'Qwen2-7B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 7B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它能够处理大规模输入。该模型在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型,并在某些任务上展现出与专有模型相当的竞争力。Qwen2-7B-Instruct 在多项评测中均优于 Qwen1.5-7B-Chat,显示出显著的性能提升', displayName: 'Qwen2 7B Instruct (Free)', id: 'Qwen/Qwen2-7B-Instruct', pricing: { @@ -234,10 +255,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { - description: 'Qwen2-7B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 7B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它能够处理大规模输入。该模型在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型,并在某些任务上展现出与专有模型相当的竞争力。Qwen2-7B-Instruct 在多项评测中均优于 Qwen1.5-7B-Chat,显示出显著的性能提升', + contextWindowTokens: 32_768, + description: + 'Qwen2-7B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 7B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它能够处理大规模输入。该模型在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型,并在某些任务上展现出与专有模型相当的竞争力。Qwen2-7B-Instruct 在多项评测中均优于 Qwen1.5-7B-Chat,显示出显著的性能提升', displayName: 'Qwen2 7B Instruct (Pro)', id: 'Pro/Qwen/Qwen2-7B-Instruct', pricing: { @@ -245,10 +267,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.35, output: 0.35, }, - tokens: 32_768, }, { - description: 'Qwen2-72B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 72B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它能够处理大规模输入。该模型在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型,并在某些任务上展现出与专有模型相当的竞争力', + contextWindowTokens: 32_768, + description: + 'Qwen2-72B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 72B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它能够处理大规模输入。该模型在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型,并在某些任务上展现出与专有模型相当的竞争力', displayName: 'Qwen2 72B Instruct', id: 'Qwen/Qwen2-7B-Instruct', pricing: { @@ -256,10 +279,11 @@ const SiliconCloud: ModelProviderCard = { input: 4.13, output: 4.13, }, - tokens: 32_768, }, { - description: 'Qwen2-72B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 72B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它能够处理大规模输入。该模型在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型,并在某些任务上展现出与专有模型相当的竞争力', + contextWindowTokens: 32_768, + description: + 'Qwen2-72B-Instruct 是 Qwen2 系列中的指令微调大语言模型,参数规模为 72B。该模型基于 Transformer 架构,采用了 SwiGLU 激活函数、注意力 QKV 偏置和组查询注意力等技术。它能够处理大规模输入。该模型在语言理解、生成、多语言能力、编码、数学和推理等多个基准测试中表现出色,超越了大多数开源模型,并在某些任务上展现出与专有模型相当的竞争力', displayName: 'Qwen2 72B Instruct (Vendor-A)', id: 'Vendor-A/Qwen/Qwen2-7B-Instruct', pricing: { @@ -267,10 +291,11 @@ const SiliconCloud: ModelProviderCard = { input: 1, output: 1, }, - tokens: 32_768, }, { - description: 'Qwen2-VL-7B-Instruct 是 Qwen-VL 模型的最新迭代版本,在视觉理解基准测试中达到了最先进的性能,包括 MathVista、DocVQA、RealWorldQA 和 MTVQA 等。Qwen2-VL 能够用于高质量的基于视频的问答、对话和内容创作,还具备复杂推理和决策能力,可以与移动设备、机器人等集成,基于视觉环境和文本指令进行自动操作。除了英语和中文,Qwen2-VL 现在还支持理解图像中不同语言的文本,包括大多数欧洲语言、日语、韩语、阿拉伯语和越南语等', + contextWindowTokens: 32_768, + description: + 'Qwen2-VL-7B-Instruct 是 Qwen-VL 模型的最新迭代版本,在视觉理解基准测试中达到了最先进的性能,包括 MathVista、DocVQA、RealWorldQA 和 MTVQA 等。Qwen2-VL 能够用于高质量的基于视频的问答、对话和内容创作,还具备复杂推理和决策能力,可以与移动设备、机器人等集成,基于视觉环境和文本指令进行自动操作。除了英语和中文,Qwen2-VL 现在还支持理解图像中不同语言的文本,包括大多数欧洲语言、日语、韩语、阿拉伯语和越南语等', displayName: 'Qwen2 VL 7B Instruct (Pro)', enabled: true, id: 'Pro/Qwen/Qwen2-VL-7B-Instruct', @@ -279,11 +304,12 @@ const SiliconCloud: ModelProviderCard = { input: 0.35, output: 0.35, }, - tokens: 32_768, vision: true, }, { - description: 'Qwen2-VL 是 Qwen-VL 模型的最新迭代版本,在视觉理解基准测试中达到了最先进的性能,包括 MathVista、DocVQA、RealWorldQA 和 MTVQA 等。Qwen2-VL 能够理解超过 20 分钟的视频,用于高质量的基于视频的问答、对话和内容创作。它还具备复杂推理和决策能力,可以与移动设备、机器人等集成,基于视觉环境和文本指令进行自动操作。除了英语和中文,Qwen2-VL 现在还支持理解图像中不同语言的文本,包括大多数欧洲语言、日语、韩语、阿拉伯语和越南语等', + contextWindowTokens: 32_768, + description: + 'Qwen2-VL 是 Qwen-VL 模型的最新迭代版本,在视觉理解基准测试中达到了最先进的性能,包括 MathVista、DocVQA、RealWorldQA 和 MTVQA 等。Qwen2-VL 能够理解超过 20 分钟的视频,用于高质量的基于视频的问答、对话和内容创作。它还具备复杂推理和决策能力,可以与移动设备、机器人等集成,基于视觉环境和文本指令进行自动操作。除了英语和中文,Qwen2-VL 现在还支持理解图像中不同语言的文本,包括大多数欧洲语言、日语、韩语、阿拉伯语和越南语等', displayName: 'Qwen2 VL 72B Instruct', enabled: true, id: 'Qwen/Qwen2-VL-72B-Instruct', @@ -292,11 +318,12 @@ const SiliconCloud: ModelProviderCard = { input: 4.13, output: 4.13, }, - tokens: 32_768, vision: true, }, { - description: 'InternLM2.5-7B-Chat 是一个开源的对话模型,基于 InternLM2 架构开发。该 7B 参数规模的模型专注于对话生成任务,支持中英双语交互。模型采用了最新的训练技术,旨在提供流畅、智能的对话体验。InternLM2.5-7B-Chat 适用于各种对话应用场景,包括但不限于智能客服、个人助手等领域', + contextWindowTokens: 32_768, + description: + 'InternLM2.5-7B-Chat 是一个开源的对话模型,基于 InternLM2 架构开发。该 7B 参数规模的模型专注于对话生成任务,支持中英双语交互。模型采用了最新的训练技术,旨在提供流畅、智能的对话体验。InternLM2.5-7B-Chat 适用于各种对话应用场景,包括但不限于智能客服、个人助手等领域', displayName: 'InternLM2.5 7B Chat (Free)', functionCall: true, id: 'internlm/internlm2_5-7b-chat', @@ -305,10 +332,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { - description: 'InternLM2.5-20B-Chat 是一个开源的大规模对话模型,基于 InternLM2 架构开发。该模型拥有 200 亿参数,在数学推理方面表现出色,超越了同量级的 Llama3 和 Gemma2-27B 模型。InternLM2.5-20B-Chat 在工具调用能力方面有显著提升,支持从上百个网页收集信息进行分析推理,并具备更强的指令理解、工具选择和结果反思能力。它适用于构建复杂智能体,可进行多轮工具调用以完成复杂任务', + contextWindowTokens: 32_768, + description: + 'InternLM2.5-20B-Chat 是一个开源的大规模对话模型,基于 InternLM2 架构开发。该模型拥有 200 亿参数,在数学推理方面表现出色,超越了同量级的 Llama3 和 Gemma2-27B 模型。InternLM2.5-20B-Chat 在工具调用能力方面有显著提升,支持从上百个网页收集信息进行分析推理,并具备更强的指令理解、工具选择和结果反思能力。它适用于构建复杂智能体,可进行多轮工具调用以完成复杂任务', displayName: 'InternLM2.5 20B Chat', functionCall: true, id: 'internlm/internlm2_5-20b-chat', @@ -317,10 +345,11 @@ const SiliconCloud: ModelProviderCard = { input: 1, output: 1, }, - tokens: 32_768, }, { - description: 'InternVL2-8B 是 InternVL 2.0 系列多模态大语言模型中的一员。该模型由 InternViT-300M-448px 视觉模型、MLP 投影层和 internlm2_5-7b-chat 语言模型组成。它在各种视觉语言任务上展现出了卓越的性能,包括文档和图表理解、场景文本理解、OCR、科学和数学问题解决等。InternVL2-8B 使用 8K 上下文窗口训练,能够处理长文本、多图像和视频输入,显著提升了模型在这些任务上的处理能力', + contextWindowTokens: 32_768, + description: + 'InternVL2-8B 是 InternVL 2.0 系列多模态大语言模型中的一员。该模型由 InternViT-300M-448px 视觉模型、MLP 投影层和 internlm2_5-7b-chat 语言模型组成。它在各种视觉语言任务上展现出了卓越的性能,包括文档和图表理解、场景文本理解、OCR、科学和数学问题解决等。InternVL2-8B 使用 8K 上下文窗口训练,能够处理长文本、多图像和视频输入,显著提升了模型在这些任务上的处理能力', displayName: 'InternVL2 8B (Pro)', id: 'Pro/OpenGVLab/InternVL2-8B', pricing: { @@ -328,11 +357,12 @@ const SiliconCloud: ModelProviderCard = { input: 0.35, output: 0.35, }, - tokens: 32_768, vision: true, }, { - description: 'InternVL2-26B 是 InternVL 2.0 系列多模态大语言模型中的一员。该模型由 InternViT-6B-448px-V1-5 视觉模型、MLP 投影层和 internlm2-chat-20b 语言模型组成。它在各种视觉语言任务上展现出了卓越的性能,包括文档和图表理解、场景文本理解、OCR、科学和数学问题解决等。InternVL2-26B 使用 8K 上下文窗口训练,能够处理长文本、多图像和视频输入,显著提升了模型在这些任务上的处理能力', + contextWindowTokens: 32_768, + description: + 'InternVL2-26B 是 InternVL 2.0 系列多模态大语言模型中的一员。该模型由 InternViT-6B-448px-V1-5 视觉模型、MLP 投影层和 internlm2-chat-20b 语言模型组成。它在各种视觉语言任务上展现出了卓越的性能,包括文档和图表理解、场景文本理解、OCR、科学和数学问题解决等。InternVL2-26B 使用 8K 上下文窗口训练,能够处理长文本、多图像和视频输入,显著提升了模型在这些任务上的处理能力', displayName: 'InternVL2 26B', id: 'OpenGVLab/InternVL2-26B', pricing: { @@ -340,11 +370,12 @@ const SiliconCloud: ModelProviderCard = { input: 1, output: 1, }, - tokens: 32_768, vision: true, }, { - description: 'InternVL2-Llama3-76B 是 InternVL 2.0 系列中的大规模多模态模型。它由 InternViT-6B-448px-V1-5 视觉模型、MLP 投影层和 Hermes-2-Theta-Llama-3-70B 语言模型组成。该模型在各种视觉语言任务上表现出色,包括文档和图表理解、信息图表问答、场景文本理解和 OCR 任务等。InternVL2-Llama3-76B 使用 8K 上下文窗口训练,能够处理长文本、多图像和视频输入,显著提升了模型在这些任务上的处理能力,在多项基准测试中达到或接近最先进的商业模型水平', + contextWindowTokens: 8192, + description: + 'InternVL2-Llama3-76B 是 InternVL 2.0 系列中的大规模多模态模型。它由 InternViT-6B-448px-V1-5 视觉模型、MLP 投影层和 Hermes-2-Theta-Llama-3-70B 语言模型组成。该模型在各种视觉语言任务上表现出色,包括文档和图表理解、信息图表问答、场景文本理解和 OCR 任务等。InternVL2-Llama3-76B 使用 8K 上下文窗口训练,能够处理长文本、多图像和视频输入,显著提升了模型在这些任务上的处理能力,在多项基准测试中达到或接近最先进的商业模型水平', displayName: 'InternVL2 Llama3 76B', id: 'OpenGVLab/InternVL2-Llama3-76B', pricing: { @@ -352,11 +383,12 @@ const SiliconCloud: ModelProviderCard = { input: 4.13, output: 4.13, }, - tokens: 8192, vision: true, }, { - description: 'GLM-4-9B-Chat 是智谱 AI 推出的 GLM-4 系列预训练模型中的开源版本。该模型在语义、数学、推理、代码和知识等多个方面表现出色。除了支持多轮对话外,GLM-4-9B-Chat 还具备网页浏览、代码执行、自定义工具调用(Function Call)和长文本推理等高级功能。模型支持 26 种语言,包括中文、英文、日语、韩语和德语等。在多项基准测试中,GLM-4-9B-Chat 展现了优秀的性能,如 AlignBench-v2、MT-Bench、MMLU 和 C-Eval 等。该模型支持最大 128K 的上下文长度,适用于学术研究和商业应用', + contextWindowTokens: 131_072, + description: + 'GLM-4-9B-Chat 是智谱 AI 推出的 GLM-4 系列预训练模型中的开源版本。该模型在语义、数学、推理、代码和知识等多个方面表现出色。除了支持多轮对话外,GLM-4-9B-Chat 还具备网页浏览、代码执行、自定义工具调用(Function Call)和长文本推理等高级功能。模型支持 26 种语言,包括中文、英文、日语、韩语和德语等。在多项基准测试中,GLM-4-9B-Chat 展现了优秀的性能,如 AlignBench-v2、MT-Bench、MMLU 和 C-Eval 等。该模型支持最大 128K 的上下文长度,适用于学术研究和商业应用', displayName: 'GLM-4 9B Chat (Free)', functionCall: true, id: 'THUDM/glm-4-9b-chat', @@ -365,10 +397,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 131_072, }, { - description: 'GLM-4-9B-Chat 是智谱 AI 推出的 GLM-4 系列预训练模型中的开源版本。该模型在语义、数学、推理、代码和知识等多个方面表现出色。除了支持多轮对话外,GLM-4-9B-Chat 还具备网页浏览、代码执行、自定义工具调用(Function Call)和长文本推理等高级功能。模型支持 26 种语言,包括中文、英文、日语、韩语和德语等。在多项基准测试中,GLM-4-9B-Chat 展现了优秀的性能,如 AlignBench-v2、MT-Bench、MMLU 和 C-Eval 等。该模型支持最大 128K 的上下文长度,适用于学术研究和商业应用', + contextWindowTokens: 131_072, + description: + 'GLM-4-9B-Chat 是智谱 AI 推出的 GLM-4 系列预训练模型中的开源版本。该模型在语义、数学、推理、代码和知识等多个方面表现出色。除了支持多轮对话外,GLM-4-9B-Chat 还具备网页浏览、代码执行、自定义工具调用(Function Call)和长文本推理等高级功能。模型支持 26 种语言,包括中文、英文、日语、韩语和德语等。在多项基准测试中,GLM-4-9B-Chat 展现了优秀的性能,如 AlignBench-v2、MT-Bench、MMLU 和 C-Eval 等。该模型支持最大 128K 的上下文长度,适用于学术研究和商业应用', displayName: 'GLM-4 9B Chat (Pro)', functionCall: true, id: 'Pro/THUDM/glm-4-9b-chat', @@ -377,10 +410,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.6, output: 0.6, }, - tokens: 131_072, }, { - description: 'ChatGLM3-6B 是 ChatGLM 系列的开源模型,由智谱 AI 开发。该模型保留了前代模型的优秀特性,如对话流畅和部署门槛低,同时引入了新的特性。它采用了更多样的训练数据、更充分的训练步数和更合理的训练策略,在 10B 以下的预训练模型中表现出色。ChatGLM3-6B 支持多轮对话、工具调用、代码执行和 Agent 任务等复杂场景。除对话模型外,还开源了基础模型 ChatGLM-6B-Base 和长文本对话模型 ChatGLM3-6B-32K。该模型对学术研究完全开放,在登记后也允许免费商业使用', + contextWindowTokens: 32_768, + description: + 'ChatGLM3-6B 是 ChatGLM 系列的开源模型,由智谱 AI 开发。该模型保留了前代模型的优秀特性,如对话流畅和部署门槛低,同时引入了新的特性。它采用了更多样的训练数据、更充分的训练步数和更合理的训练策略,在 10B 以下的预训练模型中表现出色。ChatGLM3-6B 支持多轮对话、工具调用、代码执行和 Agent 任务等复杂场景。除对话模型外,还开源了基础模型 ChatGLM-6B-Base 和长文本对话模型 ChatGLM3-6B-32K。该模型对学术研究完全开放,在登记后也允许免费商业使用', displayName: 'ChatGLM3 6B (Free)', id: 'THUDM/chatglm3-6b', pricing: { @@ -388,10 +422,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { - description: 'Yi-1.5-6B-Chat 是 Yi-1.5 系列的一个变体,属于开源聊天模型。Yi-1.5 是 Yi 的升级版本,在 500B 个高质量语料上进行了持续预训练,并在 3M 多样化的微调样本上进行了微调。相比于 Yi,Yi-1.5 在编码、数学、推理和指令遵循能力方面表现更强,同时保持了出色的语言理解、常识推理和阅读理解能力。该模型具有 4K、16K 和 32K 的上下文长度版本,预训练总量达到 3.6T 个 token', + contextWindowTokens: 4096, + description: + 'Yi-1.5-6B-Chat 是 Yi-1.5 系列的一个变体,属于开源聊天模型。Yi-1.5 是 Yi 的升级版本,在 500B 个高质量语料上进行了持续预训练,并在 3M 多样化的微调样本上进行了微调。相比于 Yi,Yi-1.5 在编码、数学、推理和指令遵循能力方面表现更强,同时保持了出色的语言理解、常识推理和阅读理解能力。该模型具有 4K、16K 和 32K 的上下文长度版本,预训练总量达到 3.6T 个 token', displayName: 'Yi-1.5 6B Chat (Free)', id: '01-ai/Yi-1.5-6B-Chat', pricing: { @@ -399,10 +434,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 4096, }, { - description: 'Yi-1.5-9B-Chat-16K 是 Yi-1.5 系列的一个变体,属于开源聊天模型。Yi-1.5 是 Yi 的升级版本,在 500B 个高质量语料上进行了持续预训练,并在 3M 多样化的微调样本上进行了微调。相比于 Yi,Yi-1.5 在编码、数学、推理和指令遵循能力方面表现更强,同时保持了出色的语言理解、常识推理和阅读理解能力。该模型在同等规模的开源模型中表现最佳', + contextWindowTokens: 16_384, + description: + 'Yi-1.5-9B-Chat-16K 是 Yi-1.5 系列的一个变体,属于开源聊天模型。Yi-1.5 是 Yi 的升级版本,在 500B 个高质量语料上进行了持续预训练,并在 3M 多样化的微调样本上进行了微调。相比于 Yi,Yi-1.5 在编码、数学、推理和指令遵循能力方面表现更强,同时保持了出色的语言理解、常识推理和阅读理解能力。该模型在同等规模的开源模型中表现最佳', displayName: 'Yi-1.5 9B Chat 16K (Free)', id: '01-ai/Yi-1.5-9B-Chat-16K', pricing: { @@ -410,10 +446,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 16_384, }, { - description: 'Yi-1.5-34B-Chat-16K 是 Yi-1.5 系列的一个变体,属于开源聊天模型。Yi-1.5 是 Yi 的升级版本,在 500B 个高质量语料上进行了持续预训练,并在 3M 多样化的微调样本上进行了微调。相比于 Yi,Yi-1.5 在编码、数学、推理和指令遵循能力方面表现更强,同时保持了出色的语言理解、常识推理和阅读理解能力。该模型在大多数基准测试中与更大的模型相当或表现更佳,具有 16K 的上下文长度', + contextWindowTokens: 16_384, + description: + 'Yi-1.5-34B-Chat-16K 是 Yi-1.5 系列的一个变体,属于开源聊天模型。Yi-1.5 是 Yi 的升级版本,在 500B 个高质量语料上进行了持续预训练,并在 3M 多样化的微调样本上进行了微调。相比于 Yi,Yi-1.5 在编码、数学、推理和指令遵循能力方面表现更强,同时保持了出色的语言理解、常识推理和阅读理解能力。该模型在大多数基准测试中与更大的模型相当或表现更佳,具有 16K 的上下文长度', displayName: 'Yi-1.5 34B Chat 16K', id: '01-ai/Yi-1.5-34B-Chat-16K', pricing: { @@ -421,10 +458,11 @@ const SiliconCloud: ModelProviderCard = { input: 1.26, output: 1.26, }, - tokens: 16_384, }, { - description: 'Gemma 是 Google 开发的轻量级、最先进的开放模型系列之一。它是一个仅解码器的大型语言模型,支持英语,提供开放权重、预训练变体和指令微调变体。Gemma 模型适用于各种文本生成任务,包括问答、摘要和推理。该 9B 模型是通过 8 万亿个 tokens 训练而成。其相对较小的规模使其可以在资源有限的环境中部署,如笔记本电脑、台式机或您自己的云基础设施,从而使更多人能够访问最先进的 AI 模型并促进创新', + contextWindowTokens: 8192, + description: + 'Gemma 是 Google 开发的轻量级、最先进的开放模型系列之一。它是一个仅解码器的大型语言模型,支持英语,提供开放权重、预训练变体和指令微调变体。Gemma 模型适用于各种文本生成任务,包括问答、摘要和推理。该 9B 模型是通过 8 万亿个 tokens 训练而成。其相对较小的规模使其可以在资源有限的环境中部署,如笔记本电脑、台式机或您自己的云基础设施,从而使更多人能够访问最先进的 AI 模型并促进创新', displayName: 'Gemma 2 9B (Free)', enabled: true, id: 'google/gemma-2-9b-it', @@ -433,10 +471,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 8192, }, { - description: 'Gemma 是 Google 开发的轻量级、最先进的开放模型系列之一。它是一个仅解码器的大型语言模型,支持英语,提供开放权重、预训练变体和指令微调变体。Gemma 模型适用于各种文本生成任务,包括问答、摘要和推理。该 9B 模型是通过 8 万亿个 tokens 训练而成。其相对较小的规模使其可以在资源有限的环境中部署,如笔记本电脑、台式机或您自己的云基础设施,从而使更多人能够访问最先进的 AI 模型并促进创新', + contextWindowTokens: 8192, + description: + 'Gemma 是 Google 开发的轻量级、最先进的开放模型系列之一。它是一个仅解码器的大型语言模型,支持英语,提供开放权重、预训练变体和指令微调变体。Gemma 模型适用于各种文本生成任务,包括问答、摘要和推理。该 9B 模型是通过 8 万亿个 tokens 训练而成。其相对较小的规模使其可以在资源有限的环境中部署,如笔记本电脑、台式机或您自己的云基础设施,从而使更多人能够访问最先进的 AI 模型并促进创新', displayName: 'Gemma 2 9B (Pro)', id: 'Pro/google/gemma-2-9b-it', pricing: { @@ -444,10 +483,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.6, output: 0.6, }, - tokens: 8192, }, { - description: 'Gemma 是由 Google 开发的轻量级、最先进的开放模型系列,采用与 Gemini 模型相同的研究和技术构建。这些模型是仅解码器的大型语言模型,支持英语,提供预训练和指令微调两种变体的开放权重。Gemma 模型适用于各种文本生成任务,包括问答、摘要和推理。其相对较小的规模使其能够部署在资源有限的环境中,如笔记本电脑、台式机或个人云基础设施,从而让所有人都能获得最先进的 AI 模型,促进创新', + contextWindowTokens: 8192, + description: + 'Gemma 是由 Google 开发的轻量级、最先进的开放模型系列,采用与 Gemini 模型相同的研究和技术构建。这些模型是仅解码器的大型语言模型,支持英语,提供预训练和指令微调两种变体的开放权重。Gemma 模型适用于各种文本生成任务,包括问答、摘要和推理。其相对较小的规模使其能够部署在资源有限的环境中,如笔记本电脑、台式机或个人云基础设施,从而让所有人都能获得最先进的 AI 模型,促进创新', displayName: 'Gemma 2 27B', enabled: true, id: 'google/gemma-2-27b-it', @@ -456,10 +496,11 @@ const SiliconCloud: ModelProviderCard = { input: 1.26, output: 1.26, }, - tokens: 8192, }, { - description: 'Meta Llama 3.1 是由 Meta 开发的多语言大型语言模型家族,包括 8B、70B 和 405B 三种参数规模的预训练和指令微调变体。该 8B 指令微调模型针对多语言对话场景进行了优化,在多项行业基准测试中表现优异。模型训练使用了超过 15 万亿个 tokens 的公开数据,并采用了监督微调和人类反馈强化学习等技术来提升模型的有用性和安全性。Llama 3.1 支持文本生成和代码生成,知识截止日期为 2023 年 12 月', + contextWindowTokens: 32_768, + description: + 'Meta Llama 3.1 是由 Meta 开发的多语言大型语言模型家族,包括 8B、70B 和 405B 三种参数规模的预训练和指令微调变体。该 8B 指令微调模型针对多语言对话场景进行了优化,在多项行业基准测试中表现优异。模型训练使用了超过 15 万亿个 tokens 的公开数据,并采用了监督微调和人类反馈强化学习等技术来提升模型的有用性和安全性。Llama 3.1 支持文本生成和代码生成,知识截止日期为 2023 年 12 月', displayName: 'Llama 3.1 8B Instruct (Free)', enabled: true, functionCall: true, @@ -469,10 +510,11 @@ const SiliconCloud: ModelProviderCard = { input: 0, output: 0, }, - tokens: 32_768, }, { - description: 'Meta Llama 3.1 是由 Meta 开发的多语言大型语言模型家族,包括 8B、70B 和 405B 三种参数规模的预训练和指令微调变体。该 8B 指令微调模型针对多语言对话场景进行了优化,在多项行业基准测试中表现优异。模型训练使用了超过 15 万亿个 tokens 的公开数据,并采用了监督微调和人类反馈强化学习等技术来提升模型的有用性和安全性。Llama 3.1 支持文本生成和代码生成,知识截止日期为 2023 年 12 月', + contextWindowTokens: 32_768, + description: + 'Meta Llama 3.1 是由 Meta 开发的多语言大型语言模型家族,包括 8B、70B 和 405B 三种参数规模的预训练和指令微调变体。该 8B 指令微调模型针对多语言对话场景进行了优化,在多项行业基准测试中表现优异。模型训练使用了超过 15 万亿个 tokens 的公开数据,并采用了监督微调和人类反馈强化学习等技术来提升模型的有用性和安全性。Llama 3.1 支持文本生成和代码生成,知识截止日期为 2023 年 12 月', displayName: 'Llama 3.1 8B Instruct (Pro)', id: 'Pro/meta-llama/Meta-Llama-3.1-8B-Instruct', pricing: { @@ -480,10 +522,11 @@ const SiliconCloud: ModelProviderCard = { input: 0.42, output: 0.42, }, - tokens: 32_768, }, { - description: 'Meta Llama 3.1 是由 Meta 开发的多语言大型语言模型家族,包括 8B、70B 和 405B 三种参数规模的预训练和指令微调变体。该 70B 指令微调模型针对多语言对话场景进行了优化,在多项行业基准测试中表现优异。模型训练使用了超过 15 万亿个 tokens 的公开数据,并采用了监督微调和人类反馈强化学习等技术来提升模型的有用性和安全性。Llama 3.1 支持文本生成和代码生成,知识截止日期为 2023 年 12 月', + contextWindowTokens: 32_768, + description: + 'Meta Llama 3.1 是由 Meta 开发的多语言大型语言模型家族,包括 8B、70B 和 405B 三种参数规模的预训练和指令微调变体。该 70B 指令微调模型针对多语言对话场景进行了优化,在多项行业基准测试中表现优异。模型训练使用了超过 15 万亿个 tokens 的公开数据,并采用了监督微调和人类反馈强化学习等技术来提升模型的有用性和安全性。Llama 3.1 支持文本生成和代码生成,知识截止日期为 2023 年 12 月', displayName: 'Llama 3.1 70B Instruct', enabled: true, functionCall: true, @@ -493,10 +536,11 @@ const SiliconCloud: ModelProviderCard = { input: 4.13, output: 4.13, }, - tokens: 32_768, }, { - description: 'Meta Llama 3.1 是由 Meta 开发的多语言大型语言模型家族,包括 8B、70B 和 405B 三种参数规模的预训练和指令微调变体。该 405B 指令微调模型针对多语言对话场景进行了优化,在多项行业基准测试中表现优异。模型训练使用了超过 15 万亿个 tokens 的公开数据,并采用了监督微调和人类反馈强化学习等技术来提升模型的有用性和安全性。Llama 3.1 支持文本生成和代码生成,知识截止日期为 2023 年 12 月', + contextWindowTokens: 32_768, + description: + 'Meta Llama 3.1 是由 Meta 开发的多语言大型语言模型家族,包括 8B、70B 和 405B 三种参数规模的预训练和指令微调变体。该 405B 指令微调模型针对多语言对话场景进行了优化,在多项行业基准测试中表现优异。模型训练使用了超过 15 万亿个 tokens 的公开数据,并采用了监督微调和人类反馈强化学习等技术来提升模型的有用性和安全性。Llama 3.1 支持文本生成和代码生成,知识截止日期为 2023 年 12 月', displayName: 'Llama 3.1 405B Instruct', enabled: true, id: 'meta-llama/Meta-Llama-3.1-405B-Instruct', @@ -505,10 +549,11 @@ const SiliconCloud: ModelProviderCard = { input: 21, output: 21, }, - tokens: 32_768, }, { - description: 'Llama-3.1-Nemotron-70B-Instruct 是由 NVIDIA 定制的大型语言模型,旨在提高 LLM 生成的响应对用户查询的帮助程度。该模型在 Arena Hard、AlpacaEval 2 LC 和 GPT-4-Turbo MT-Bench 等基准测试中表现出色,截至 2024 年 10 月 1 日,在所有三个自动对齐基准测试中排名第一。该模型使用 RLHF(特别是 REINFORCE)、Llama-3.1-Nemotron-70B-Reward 和 HelpSteer2-Preference 提示在 Llama-3.1-70B-Instruct 模型基础上进行训练', + contextWindowTokens: 32_768, + description: + 'Llama-3.1-Nemotron-70B-Instruct 是由 NVIDIA 定制的大型语言模型,旨在提高 LLM 生成的响应对用户查询的帮助程度。该模型在 Arena Hard、AlpacaEval 2 LC 和 GPT-4-Turbo MT-Bench 等基准测试中表现出色,截至 2024 年 10 月 1 日,在所有三个自动对齐基准测试中排名第一。该模型使用 RLHF(特别是 REINFORCE)、Llama-3.1-Nemotron-70B-Reward 和 HelpSteer2-Preference 提示在 Llama-3.1-70B-Instruct 模型基础上进行训练', displayName: 'Llama 3.1 Nemotron 70B Instruct', enabled: true, id: 'nvidia/Llama-3.1-Nemotron-70B-Instruct', @@ -517,10 +562,11 @@ const SiliconCloud: ModelProviderCard = { input: 4.13, output: 4.13, }, - tokens: 32_768, }, { - description: 'TeleChat2大模型是由中国电信从0到1自主研发的生成式语义大模型,支持百科问答、代码生成、长文生成等功能,为用户提供对话咨询服务,能够与用户进行对话互动,回答问题,协助创作,高效便捷地帮助用户获取信息、知识和灵感。模型在幻觉问题、长文生成、逻辑理解等方面均有较出色表现。', + contextWindowTokens: 8192, + description: + 'TeleChat2大模型是由中国电信从0到1自主研发的生成式语义大模型,支持百科问答、代码生成、长文生成等功能,为用户提供对话咨询服务,能够与用户进行对话互动,回答问题,协助创作,高效便捷地帮助用户获取信息、知识和灵感。模型在幻觉问题、长文生成、逻辑理解等方面均有较出色表现。', displayName: 'TeleChat2', id: 'TeleAI/TeleChat2', pricing: { @@ -528,10 +574,11 @@ const SiliconCloud: ModelProviderCard = { input: 1.33, output: 1.33, }, - tokens: 8192, }, { - description: 'TeleMM多模态大模型是由中国电信自主研发的多模态理解大模型,能够处理文本、图像等多种模态输入,支持图像理解、图表分析等功能,为用户提供跨模态的理解服务。模型能够与用户进行多模态交互,准确理解输入内容,回答问题、协助创作,并高效提供多模态信息和灵感支持。在细粒度感知,逻辑推理等多模态任务上有出色表现', + contextWindowTokens: 32_768, + description: + 'TeleMM多模态大模型是由中国电信自主研发的多模态理解大模型,能够处理文本、图像等多种模态输入,支持图像理解、图表分析等功能,为用户提供跨模态的理解服务。模型能够与用户进行多模态交互,准确理解输入内容,回答问题、协助创作,并高效提供多模态信息和灵感支持。在细粒度感知,逻辑推理等多模态任务上有出色表现', displayName: 'TeleMM', id: 'TeleAI/TeleMM', pricing: { @@ -539,7 +586,6 @@ const SiliconCloud: ModelProviderCard = { input: 1.33, output: 1.33, }, - tokens: 32_768, vision: true, }, ], diff --git a/src/config/modelProviders/spark.ts b/src/config/modelProviders/spark.ts index a3ad64c202d2..a03c8e853aa8 100644 --- a/src/config/modelProviders/spark.ts +++ b/src/config/modelProviders/spark.ts @@ -5,6 +5,7 @@ import { ModelProviderCard } from '@/types/llm'; const Spark: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 8192, description: 'Spark Lite 是一款轻量级大语言模型,具备极低的延迟与高效的处理能力,完全免费开放,支持实时在线搜索功能。其快速响应的特性使其在低算力设备上的推理应用和模型微调中表现出色,为用户带来出色的成本效益和智能体验,尤其在知识问答、内容生成及搜索场景下表现不俗。', displayName: 'Spark Lite', @@ -12,9 +13,9 @@ const Spark: ModelProviderCard = { functionCall: false, id: 'lite', maxOutput: 4096, - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Spark Pro 是一款为专业领域优化的高性能大语言模型,专注数学、编程、医疗、教育等多个领域,并支持联网搜索及内置天气、日期等插件。其优化后模型在复杂知识问答、语言理解及高层次文本创作中展现出色表现和高效性能,是适合专业应用场景的理想选择。', displayName: 'Spark Pro', @@ -22,9 +23,9 @@ const Spark: ModelProviderCard = { functionCall: false, id: 'generalv3', maxOutput: 8192, - tokens: 8192, }, { + contextWindowTokens: 131_072, description: 'Spark Pro 128K 配置了特大上下文处理能力,能够处理多达128K的上下文信息,特别适合需通篇分析和长期逻辑关联处理的长文内容,可在复杂文本沟通中提供流畅一致的逻辑与多样的引用支持。', displayName: 'Spark Pro 128K', @@ -32,9 +33,9 @@ const Spark: ModelProviderCard = { functionCall: false, id: 'pro-128k', maxOutput: 4096, - tokens: 131_072, }, { + contextWindowTokens: 8192, description: 'Spark Max 为功能最为全面的版本,支持联网搜索及众多内置插件。其全面优化的核心能力以及系统角色设定和函数调用功能,使其在各种复杂应用场景中的表现极为优异和出色。', displayName: 'Spark Max', @@ -42,9 +43,9 @@ const Spark: ModelProviderCard = { functionCall: false, id: 'generalv3.5', maxOutput: 8192, - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Spark Max 32K 配置了大上下文处理能力,更强的上下文理解和逻辑推理能力,支持32K tokens的文本输入,适用于长文档阅读、私有知识问答等场景', displayName: 'Spark Max 32K', @@ -52,9 +53,9 @@ const Spark: ModelProviderCard = { functionCall: false, id: 'max-32k', maxOutput: 8192, - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'Spark Ultra 是星火大模型系列中最为强大的版本,在升级联网搜索链路同时,提升对文本内容的理解和总结能力。它是用于提升办公生产力和准确响应需求的全方位解决方案,是引领行业的智能产品。', displayName: 'Spark 4.0 Ultra', @@ -62,7 +63,6 @@ const Spark: ModelProviderCard = { functionCall: false, id: '4.0Ultra', maxOutput: 8192, - tokens: 8192, }, ], checkModel: 'lite', diff --git a/src/config/modelProviders/stepfun.ts b/src/config/modelProviders/stepfun.ts index 4f775df8f1ed..a4eac4e5533f 100644 --- a/src/config/modelProviders/stepfun.ts +++ b/src/config/modelProviders/stepfun.ts @@ -5,6 +5,7 @@ import { ModelProviderCard } from '@/types/llm'; const Stepfun: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 8000, description: '高速模型,适合实时对话。', displayName: 'Step 1 Flash', enabled: true, @@ -15,9 +16,9 @@ const Stepfun: ModelProviderCard = { input: 1, output: 4, }, - tokens: 8000, }, { + contextWindowTokens: 8000, description: '小型模型,适合轻量级任务。', displayName: 'Step 1 8K', enabled: true, @@ -28,9 +29,9 @@ const Stepfun: ModelProviderCard = { input: 5, output: 20, }, - tokens: 8000, }, { + contextWindowTokens: 32_000, description: '支持中等长度的对话,适用于多种应用场景。', displayName: 'Step 1 32K', enabled: true, @@ -41,9 +42,9 @@ const Stepfun: ModelProviderCard = { input: 15, output: 70, }, - tokens: 32_000, }, { + contextWindowTokens: 128_000, description: '平衡性能与成本,适合一般场景。', displayName: 'Step 1 128K', enabled: true, @@ -54,9 +55,9 @@ const Stepfun: ModelProviderCard = { input: 40, output: 200, }, - tokens: 128_000, }, { + contextWindowTokens: 256_000, description: '具备超长上下文处理能力,尤其适合长文档分析。', displayName: 'Step 1 256K', functionCall: true, @@ -66,9 +67,9 @@ const Stepfun: ModelProviderCard = { input: 95, output: 300, }, - tokens: 256_000, }, { + contextWindowTokens: 16_000, description: '支持大规模上下文交互,适合复杂对话场景。', displayName: 'Step 2 16K', enabled: true, @@ -79,9 +80,9 @@ const Stepfun: ModelProviderCard = { input: 38, output: 120, }, - tokens: 16_000, }, { + contextWindowTokens: 8000, description: '小型视觉模型,适合基本的图文任务。', displayName: 'Step 1V 8K', enabled: true, @@ -92,10 +93,10 @@ const Stepfun: ModelProviderCard = { input: 5, output: 20, }, - tokens: 8000, vision: true, }, { + contextWindowTokens: 32_000, description: '支持视觉输入,增强多模态交互体验。', displayName: 'Step 1V 32K', enabled: true, @@ -106,10 +107,10 @@ const Stepfun: ModelProviderCard = { input: 15, output: 70, }, - tokens: 32_000, vision: true, }, { + contextWindowTokens: 32_000, description: '该模型拥有强大的视频理解能力。', displayName: 'Step 1.5V Mini', enabled: true, @@ -119,7 +120,6 @@ const Stepfun: ModelProviderCard = { input: 8, output: 35, }, - tokens: 32_000, vision: true, }, ], diff --git a/src/config/modelProviders/taichu.ts b/src/config/modelProviders/taichu.ts index 11e6fefd94ae..9592045c5332 100644 --- a/src/config/modelProviders/taichu.ts +++ b/src/config/modelProviders/taichu.ts @@ -4,13 +4,12 @@ import { ModelProviderCard } from '@/types/llm'; const Taichu: ModelProviderCard = { chatModels: [ { - description: - 'Taichu 2.0 基于海量高质数据训练,具有更强的文本理解、内容创作、对话问答等能力', + contextWindowTokens: 32_768, + description: 'Taichu 2.0 基于海量高质数据训练,具有更强的文本理解、内容创作、对话问答等能力', displayName: 'Taichu 2.0', enabled: true, functionCall: true, id: 'taichu_llm', - tokens: 32_768, }, /* // TODO: Not support for now diff --git a/src/config/modelProviders/togetherai.ts b/src/config/modelProviders/togetherai.ts index fbe34fad3bcd..c3b58849e1db 100644 --- a/src/config/modelProviders/togetherai.ts +++ b/src/config/modelProviders/togetherai.ts @@ -5,267 +5,276 @@ import { ModelProviderCard } from '@/types/llm'; const TogetherAI: ModelProviderCard = { chatModels: [ { - description: 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', + contextWindowTokens: 131_072, + description: + 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 3B Instruct Turbo', enabled: true, id: 'meta-llama/Llama-3.2-3B-Instruct-Turbo', - tokens: 131_072, }, { - description: 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', + contextWindowTokens: 131_072, + description: + 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 11B Vision Instruct Turbo (Free)', enabled: true, id: 'meta-llama/Llama-Vision-Free', - tokens: 131_072, vision: true, }, { - description: 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', + contextWindowTokens: 131_072, + description: + 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 11B Vision Instruct Turbo', id: 'meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo', - tokens: 131_072, vision: true, }, { - description: 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', + contextWindowTokens: 131_072, + description: + 'LLaMA 3.2 旨在处理结合视觉和文本数据的任务。它在图像描述和视觉问答等任务中表现出色,跨越了语言生成和视觉推理之间的鸿沟。', displayName: 'Llama 3.2 90B Vision Instruct Turbo', enabled: true, id: 'meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo', - tokens: 131_072, vision: true, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 8B 模型采用FP8量化,支持高达131,072个上下文标记,是开源模型中的佼佼者,适合复杂任务,表现优异于许多行业基准。', displayName: 'Llama 3.1 8B Instruct Turbo', enabled: true, functionCall: true, id: 'meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo', - tokens: 131_072, }, { + contextWindowTokens: 131_072, description: 'Llama 3.1 70B 模型经过精细调整,适用于高负载应用,量化至FP8提供更高效的计算能力和准确性,确保在复杂场景中的卓越表现。', displayName: 'Llama 3.1 70B Instruct Turbo', enabled: true, functionCall: true, id: 'meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo', - tokens: 131_072, }, { + contextWindowTokens: 130_815, description: '405B 的 Llama 3.1 Turbo 模型,为大数据处理提供超大容量的上下文支持,在超大规模的人工智能应用中表现突出。', displayName: 'Llama 3.1 405B Instruct Turbo', enabled: true, functionCall: true, id: 'meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo', - tokens: 130_815, }, { + contextWindowTokens: 32_768, description: 'Llama 3.1 Nemotron 70B 是由 NVIDIA 定制的大型语言模型,旨在提高 LLM 生成的响应对用户查询的帮助程度。该模型在 Arena Hard、AlpacaEval 2 LC 和 GPT-4-Turbo MT-Bench 等基准测试中表现出色,截至 2024 年 10 月 1 日,在所有三个自动对齐基准测试中排名第一。该模型使用 RLHF(特别是 REINFORCE)、Llama-3.1-Nemotron-70B-Reward 和 HelpSteer2-Preference 提示在 Llama-3.1-70B-Instruct 模型基础上进行训练', displayName: 'Llama 3.1 Nemotron 70B', enabled: true, id: 'nvidia/Llama-3.1-Nemotron-70B-Instruct-HF', - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'Llama 3 8B Instruct Turbo 是一款高效能的大语言模型,支持广泛的应用场景。', displayName: 'Llama 3 8B Instruct Turbo', id: 'meta-llama/Meta-Llama-3-8B-Instruct-Turbo', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 70B Instruct Turbo 提供卓越的语言理解和生成能力,适合最苛刻的计算任务。', displayName: 'Llama 3 70B Instruct Turbo', id: 'meta-llama/Meta-Llama-3-70B-Instruct-Turbo', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 8B Instruct Lite 适合资源受限的环境,提供出色的平衡性能。', displayName: 'Llama 3 8B Instruct Lite', id: 'meta-llama/Meta-Llama-3-8B-Instruct-Lite', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 70B Instruct Lite 适合需要高效能和低延迟的环境。', displayName: 'Llama 3 70B Instruct Lite', id: 'meta-llama/Meta-Llama-3-70B-Instruct-Lite', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 8B Instruct Reference 提供多语言支持,涵盖丰富的领域知识。', displayName: 'Llama 3 8B Instruct Reference', id: 'meta-llama/Llama-3-8b-chat-hf', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Llama 3 70B Instruct Reference 是功能强大的聊天模型,支持复杂的对话需求。', displayName: 'Llama 3 70B Instruct Reference', id: 'meta-llama/Llama-3-70b-chat-hf', - tokens: 8192, }, { + contextWindowTokens: 4096, description: 'LLaMA-2 Chat (13B) 提供优秀的语言处理能力和出色的交互体验。', displayName: 'LLaMA-2 Chat (13B)', id: 'meta-llama/Llama-2-13b-chat-hf', - tokens: 4096, }, { + contextWindowTokens: 4096, description: 'LLaMA-2 提供优秀的语言处理能力和出色的交互体验。', displayName: 'LLaMA-2 (70B)', id: 'meta-llama/Llama-2-70b-hf', - tokens: 4096, }, { - description: 'Code Llama 是一款专注于代码生成和讨论的 LLM,结合广泛的编程语言支持,适用于开发者环境。', + contextWindowTokens: 16_384, + description: + 'Code Llama 是一款专注于代码生成和讨论的 LLM,结合广泛的编程语言支持,适用于开发者环境。', displayName: 'CodeLlama 34B Instruct', id: 'codellama/CodeLlama-34b-Instruct-hf', - tokens: 16_384, }, { + contextWindowTokens: 8192, description: 'Gemma 2 9B 由Google开发,提供高效的指令响应和综合能力。', displayName: 'Gemma 2 9B', enabled: true, id: 'google/gemma-2-9b-it', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma 2 27B 是一款通用大语言模型,具有优异的性能和广泛的应用场景。', displayName: 'Gemma 2 27B', enabled: true, id: 'google/gemma-2-27b-it', - tokens: 8192, }, { + contextWindowTokens: 8192, description: 'Gemma Instruct (2B) 提供基本的指令处理能力,适合轻量级应用。', displayName: 'Gemma Instruct (2B)', id: 'google/gemma-2b-it', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Mistral (7B) Instruct v0.3 提供高效的计算能力和自然语言理解,适合广泛的应用。', displayName: 'Mistral (7B) Instruct v0.3', enabled: true, id: 'mistralai/Mistral-7B-Instruct-v0.3', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Mistral (7B) Instruct v0.2 提供改进的指令处理能力和更精确的结果。', displayName: 'Mistral (7B) Instruct v0.2', id: 'mistralai/Mistral-7B-Instruct-v0.2', - tokens: 32_768, }, { + contextWindowTokens: 8192, description: 'Mistral (7B) Instruct 以高性能著称,适用于多种语言任务。', displayName: 'Mistral (7B) Instruct', functionCall: true, id: 'mistralai/Mistral-7B-Instruct-v0.1', - tokens: 8192, }, { - description: 'Mistral 7B是一款紧凑但高性能的模型,擅长批量处理和简单任务,如分类和文本生成,具有良好的推理能力。', + contextWindowTokens: 8192, + description: + 'Mistral 7B是一款紧凑但高性能的模型,擅长批量处理和简单任务,如分类和文本生成,具有良好的推理能力。', displayName: 'Mistral (7B)', id: 'mistralai/Mistral-7B-v0.1', - tokens: 8192, }, { + contextWindowTokens: 32_768, description: 'Mixtral-8x7B Instruct (46.7B) 提供高容量的计算框架,适合大规模数据处理。', displayName: 'Mixtral-8x7B Instruct (46.7B)', enabled: true, functionCall: true, id: 'mistralai/Mixtral-8x7B-Instruct-v0.1', - tokens: 32_768, }, { - description: 'Mixtral 8x7B是一个稀疏专家模型,利用多个参数提高推理速度,适合处理多语言和代码生成任务。', + contextWindowTokens: 32_768, + description: + 'Mixtral 8x7B是一个稀疏专家模型,利用多个参数提高推理速度,适合处理多语言和代码生成任务。', displayName: 'Mixtral-8x7B (46.7B)', id: 'mistralai/Mixtral-8x7B-v0.1', - tokens: 32_768, }, { + contextWindowTokens: 65_536, description: 'Mixtral-8x22B Instruct (141B) 是一款超级大语言模型,支持极高的处理需求。', displayName: 'Mixtral-8x22B Instruct (141B)', enabled: true, id: 'mistralai/Mixtral-8x22B-Instruct-v0.1', - tokens: 65_536, }, { - description: 'WizardLM 2 是微软AI提供的语言模型,在复杂对话、多语言、推理和智能助手领域表现尤为出色。', + contextWindowTokens: 65_536, + description: + 'WizardLM 2 是微软AI提供的语言模型,在复杂对话、多语言、推理和智能助手领域表现尤为出色。', displayName: 'WizardLM-2 8x22B', id: 'microsoft/WizardLM-2-8x22B', - tokens: 65_536, }, { + contextWindowTokens: 4096, description: 'DeepSeek LLM Chat (67B) 是创新的 AI 模型 提供深度语言理解和互动能力。', displayName: 'DeepSeek LLM Chat (67B)', enabled: true, id: 'deepseek-ai/deepseek-llm-67b-chat', - tokens: 4096, }, { + contextWindowTokens: 32_768, description: 'QwQ模型是由 Qwen 团队开发的实验性研究模型,专注于增强 AI 推理能力。', displayName: 'QwQ 32B Preview', enabled: true, id: 'Qwen/QwQ-32B-Preview', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Qwen2.5 是全新的大型语言模型系列,旨在优化指令式任务的处理。', displayName: 'Qwen 2.5 7B Instruct Turbo', enabled: true, id: 'Qwen/Qwen2.5-7B-Instruct-Turbo', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Qwen2.5 是全新的大型语言模型系列,旨在优化指令式任务的处理。', displayName: 'Qwen 2.5 72B Instruct Turbo', enabled: true, id: 'Qwen/Qwen2.5-72B-Instruct-Turbo', - tokens: 32_768, }, { - description: 'Qwen2.5 Coder 32B Instruct 是阿里云发布的代码特定大语言模型系列的最新版本。该模型在 Qwen2.5 的基础上,通过 5.5 万亿个 tokens 的训练,显著提升了代码生成、推理和修复能力。它不仅增强了编码能力,还保持了数学和通用能力的优势。模型为代码智能体等实际应用提供了更全面的基础', + contextWindowTokens: 32_768, + description: + 'Qwen2.5 Coder 32B Instruct 是阿里云发布的代码特定大语言模型系列的最新版本。该模型在 Qwen2.5 的基础上,通过 5.5 万亿个 tokens 的训练,显著提升了代码生成、推理和修复能力。它不仅增强了编码能力,还保持了数学和通用能力的优势。模型为代码智能体等实际应用提供了更全面的基础', displayName: 'Qwen 2.5 Coder 32B Instruct', id: 'Qwen/Qwen2.5-Coder-32B-Instruct', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Qwen 2 Instruct (72B) 为企业级应用提供精准的指令理解和响应。', displayName: 'Qwen 2 Instruct (72B)', id: 'Qwen/Qwen2-72B-Instruct', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'DBRX Instruct 提供高可靠性的指令处理能力,支持多行业应用。', displayName: 'DBRX Instruct', id: 'databricks/dbrx-instruct', - tokens: 32_768, }, { + contextWindowTokens: 4096, description: 'Upstage SOLAR Instruct v1 (11B) 适用于精细化指令任务,提供出色的语言处理能力。', displayName: 'Upstage SOLAR Instruct v1 (11B)', id: 'upstage/SOLAR-10.7B-Instruct-v1.0', - tokens: 4096, }, { + contextWindowTokens: 32_768, description: 'Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B) 是高精度的指令模型,适用于复杂计算。', displayName: 'Nous Hermes 2 - Mixtral 8x7B-DPO (46.7B)', id: 'NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO', - tokens: 32_768, }, { + contextWindowTokens: 4096, description: 'MythoMax-L2 (13B) 是一种创新模型,适合多领域应用和复杂任务。', displayName: 'MythoMax-L2 (13B)', id: 'Gryphe/MythoMax-L2-13b', - tokens: 4096, }, { + contextWindowTokens: 32_768, description: 'StripedHyena Nous (7B) 通过高效的策略和模型架构,提供增强的计算能力。', displayName: 'StripedHyena Nous (7B)', id: 'togethercomputer/StripedHyena-Nous-7B', - tokens: 32_768, }, ], checkModel: 'meta-llama/Llama-Vision-Free', diff --git a/src/config/modelProviders/upstage.ts b/src/config/modelProviders/upstage.ts index 917d8c7f6df2..190abdc020ad 100644 --- a/src/config/modelProviders/upstage.ts +++ b/src/config/modelProviders/upstage.ts @@ -4,30 +4,30 @@ import { ModelProviderCard } from '@/types/llm'; const Upstage: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 32_768, description: 'Solar Mini 是一种紧凑型 LLM,性能优于 GPT-3.5,具备强大的多语言能力,支持英语和韩语,提供高效小巧的解决方案。', displayName: 'Solar Mini', enabled: true, functionCall: true, id: 'solar-1-mini-chat', - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: 'Solar Mini (Ja) 扩展了 Solar Mini 的能力,专注于日语,同时在英语和韩语的使用中保持高效和卓越性能。', displayName: 'Solar Mini (Ja)', functionCall: false, id: 'solar-1-mini-chat-ja', - tokens: 32_768, }, { + contextWindowTokens: 4096, description: 'Solar Pro 是 Upstage 推出的一款高智能LLM,专注于单GPU的指令跟随能力,IFEval得分80以上。目前支持英语,正式版本计划于2024年11月推出,将扩展语言支持和上下文长度。', displayName: 'Solar Pro', enabled: true, functionCall: false, id: 'solar-pro', - tokens: 4096, }, ], checkModel: 'solar-1-mini-chat', diff --git a/src/config/modelProviders/wenxin.ts b/src/config/modelProviders/wenxin.ts index b8330f88e725..3368d2d1a17e 100644 --- a/src/config/modelProviders/wenxin.ts +++ b/src/config/modelProviders/wenxin.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const BaiduWenxin: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 8192, description: '百度自研的旗舰级大规模⼤语⾔模型,覆盖海量中英文语料,具有强大的通用能力,可满足绝大部分对话问答、创作生成、插件应用场景要求;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 3.5 8K', @@ -14,9 +15,9 @@ const BaiduWenxin: ModelProviderCard = { input: 0.8, output: 2, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级大规模⼤语⾔模型,覆盖海量中英文语料,具有强大的通用能力,可满足绝大部分对话问答、创作生成、插件应用场景要求;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 3.5 8K Preview', @@ -26,9 +27,9 @@ const BaiduWenxin: ModelProviderCard = { input: 0.8, output: 2, }, - tokens: 8192, }, { + contextWindowTokens: 128_000, description: '百度自研的旗舰级大规模⼤语⾔模型,覆盖海量中英文语料,具有强大的通用能力,可满足绝大部分对话问答、创作生成、插件应用场景要求;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 3.5 128K', @@ -39,9 +40,9 @@ const BaiduWenxin: ModelProviderCard = { input: 0.8, output: 2, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级超大规模⼤语⾔模型,相较ERNIE 3.5实现了模型能力全面升级,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 4.0 8K', @@ -52,9 +53,9 @@ const BaiduWenxin: ModelProviderCard = { input: 30, output: 90, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级超大规模⼤语⾔模型,相较ERNIE 3.5实现了模型能力全面升级,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。', displayName: 'ERNIE 4.0 8K Preview', @@ -64,9 +65,9 @@ const BaiduWenxin: ModelProviderCard = { input: 30, output: 90, }, - tokens: 8192, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级超大规模⼤语⾔模型,综合效果表现出色,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。相较于ERNIE 4.0在性能表现上更优秀', displayName: 'ERNIE 4.0 Turbo 8K', @@ -77,9 +78,9 @@ const BaiduWenxin: ModelProviderCard = { input: 20, output: 60, }, - tokens: 8192, }, { + contextWindowTokens: 128_000, description: '百度自研的旗舰级超大规模⼤语⾔模型,综合效果表现出色,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。相较于ERNIE 4.0在性能表现上更优秀', displayName: 'ERNIE 4.0 Turbo 128K', @@ -90,9 +91,9 @@ const BaiduWenxin: ModelProviderCard = { input: 20, output: 60, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: '百度自研的旗舰级超大规模⼤语⾔模型,综合效果表现出色,广泛适用于各领域复杂任务场景;支持自动对接百度搜索插件,保障问答信息时效。相较于ERNIE 4.0在性能表现上更优秀', displayName: 'ERNIE 4.0 Turbo 8K Preview', @@ -102,9 +103,9 @@ const BaiduWenxin: ModelProviderCard = { input: 20, output: 60, }, - tokens: 8192, }, { + contextWindowTokens: 128_000, description: '百度自研的轻量级大语言模型,兼顾优异的模型效果与推理性能,效果比ERNIE Lite更优,适合低算力AI加速卡推理使用。', displayName: 'ERNIE Lite Pro 128K', @@ -115,9 +116,9 @@ const BaiduWenxin: ModelProviderCard = { input: 0.2, output: 0.4, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: '百度2024年最新发布的自研高性能大语言模型,通用能力优异,效果比ERNIE Speed更优,适合作为基座模型进行精调,更好地处理特定场景问题,同时具备极佳的推理性能。', displayName: 'ERNIE Speed Pro 128K', @@ -128,9 +129,9 @@ const BaiduWenxin: ModelProviderCard = { input: 0.3, output: 0.6, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: '百度2024年最新发布的自研高性能大语言模型,通用能力优异,适合作为基座模型进行精调,更好地处理特定场景问题,同时具备极佳的推理性能。', displayName: 'ERNIE Speed 128K', @@ -140,9 +141,9 @@ const BaiduWenxin: ModelProviderCard = { input: 0, output: 0, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: '百度自研的垂直场景大语言模型,适合游戏NPC、客服对话、对话角色扮演等应用场景,人设风格更为鲜明、一致,指令遵循能力更强,推理性能更优。', displayName: 'ERNIE Character 8K', @@ -152,7 +153,6 @@ const BaiduWenxin: ModelProviderCard = { input: 4, output: 8, }, - tokens: 8192, }, ], checkModel: 'ERNIE-Speed-128K', diff --git a/src/config/modelProviders/xai.ts b/src/config/modelProviders/xai.ts index 82661c0f4949..a5272a88a15f 100644 --- a/src/config/modelProviders/xai.ts +++ b/src/config/modelProviders/xai.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const XAI: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 131_072, description: '拥有与 Grok 2 相当的性能,但具有更高的效率、速度和功能。', displayName: 'Grok Beta', enabled: true, @@ -13,9 +14,9 @@ const XAI: ModelProviderCard = { input: 5, output: 15, }, - tokens: 131_072, }, { + contextWindowTokens: 8192, description: '最新的图像理解模型,可以处理各种各样的视觉信息,包括文档、图表、截图和照片等。', displayName: 'Grok Vision Beta', enabled: true, @@ -25,10 +26,10 @@ const XAI: ModelProviderCard = { input: 5, output: 15, }, - tokens: 8192, vision: true, }, { + contextWindowTokens: 131_072, description: '该模型在准确性、指令遵循和多语言能力方面有所改进。', displayName: 'Grok 2 1212', enabled: true, @@ -39,9 +40,9 @@ const XAI: ModelProviderCard = { output: 10, }, releasedAt: '2024-12-12', - tokens: 131_072, }, { + contextWindowTokens: 32_768, description: '该模型在准确性、指令遵循和多语言能力方面有所改进。', displayName: 'Grok 2 Vision 1212', enabled: true, @@ -52,7 +53,6 @@ const XAI: ModelProviderCard = { output: 10, }, releasedAt: '2024-12-12', - tokens: 32_768, vision: true, }, ], diff --git a/src/config/modelProviders/zeroone.ts b/src/config/modelProviders/zeroone.ts index 706a67966f8a..e0c3ac5feab2 100644 --- a/src/config/modelProviders/zeroone.ts +++ b/src/config/modelProviders/zeroone.ts @@ -4,6 +4,7 @@ import { ModelProviderCard } from '@/types/llm'; const ZeroOne: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 16_384, description: '最新高性能模型,保证高质量输出同时,推理速度大幅提升。', displayName: 'Yi Lightning', enabled: true, @@ -13,9 +14,9 @@ const ZeroOne: ModelProviderCard = { input: 0.99, output: 0.99, }, - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: '小而精悍,轻量极速模型。提供强化数学运算和代码编写能力。', displayName: 'Yi Spark', enabled: true, @@ -25,9 +26,9 @@ const ZeroOne: ModelProviderCard = { input: 1, output: 1, }, - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: '中型尺寸模型升级微调,能力均衡,性价比高。深度优化指令遵循能力。', displayName: 'Yi Medium', enabled: true, @@ -37,9 +38,9 @@ const ZeroOne: ModelProviderCard = { input: 2.5, output: 2.5, }, - tokens: 16_384, }, { + contextWindowTokens: 200_000, description: '200K 超长上下文窗口,提供长文本深度理解和生成能力。', displayName: 'Yi Medium 200K', enabled: true, @@ -49,9 +50,9 @@ const ZeroOne: ModelProviderCard = { input: 12, output: 12, }, - tokens: 200_000, }, { + contextWindowTokens: 16_384, description: '超高性价比、卓越性能。根据性能和推理速度、成本,进行平衡性高精度调优。', displayName: 'Yi Large Turbo', enabled: true, @@ -61,9 +62,9 @@ const ZeroOne: ModelProviderCard = { input: 12, output: 12, }, - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: '基于 yi-large 超强模型的高阶服务,结合检索与生成技术提供精准答案,实时全网检索信息服务。', displayName: 'Yi Large RAG', @@ -74,9 +75,9 @@ const ZeroOne: ModelProviderCard = { input: 25, output: 25, }, - tokens: 16_384, }, { + contextWindowTokens: 32_768, description: '在 yi-large 模型的基础上支持并强化了工具调用的能力,适用于各种需要搭建 agent 或 workflow 的业务场景。', displayName: 'Yi Large FC', @@ -88,9 +89,9 @@ const ZeroOne: ModelProviderCard = { input: 20, output: 20, }, - tokens: 32_768, }, { + contextWindowTokens: 32_768, description: '全新千亿参数模型,提供超强问答及文本生成能力。', displayName: 'Yi Large', id: 'yi-large', @@ -99,9 +100,9 @@ const ZeroOne: ModelProviderCard = { input: 20, output: 20, }, - tokens: 32_768, }, { + contextWindowTokens: 16_384, description: '复杂视觉任务模型,提供高性能图片理解、分析能力。', displayName: 'Yi Vision', enabled: true, @@ -111,10 +112,10 @@ const ZeroOne: ModelProviderCard = { input: 6, output: 6, }, - tokens: 16_384, vision: true, }, { + contextWindowTokens: 16_384, description: '初期版本,推荐使用 yi-large(新版本)。', displayName: 'Yi Large Preview', id: 'yi-large-preview', @@ -123,9 +124,9 @@ const ZeroOne: ModelProviderCard = { input: 20, output: 20, }, - tokens: 16_384, }, { + contextWindowTokens: 16_384, description: '轻量化版本,推荐使用 yi-lightning。', displayName: 'Yi Lightning Lite', id: 'yi-lightning-lite', @@ -134,7 +135,6 @@ const ZeroOne: ModelProviderCard = { input: 0.99, output: 0.99, }, - tokens: 16_384, }, ], checkModel: 'yi-lightning', diff --git a/src/config/modelProviders/zhipu.ts b/src/config/modelProviders/zhipu.ts index 26c15179f381..2e0003c046d6 100644 --- a/src/config/modelProviders/zhipu.ts +++ b/src/config/modelProviders/zhipu.ts @@ -6,6 +6,7 @@ import { ModelProviderCard } from '@/types/llm'; const ZhiPu: ModelProviderCard = { chatModels: [ { + contextWindowTokens: 128_000, description: 'GLM-4-Flash 是处理简单任务的理想选择,速度最快且免费。', displayName: 'GLM-4-Flash', enabled: true, @@ -16,9 +17,9 @@ const ZhiPu: ModelProviderCard = { input: 0, output: 0, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4-FlashX 是Flash的增强版本,超快推理速度。', displayName: 'GLM-4-FlashX', enabled: true, @@ -29,9 +30,9 @@ const ZhiPu: ModelProviderCard = { input: 0.1, output: 0.1, }, - tokens: 128_000, }, { + contextWindowTokens: 1_024_000, description: 'GLM-4-Long 支持超长文本输入,适合记忆型任务与大规模文档处理。', displayName: 'GLM-4-Long', functionCall: true, @@ -41,9 +42,9 @@ const ZhiPu: ModelProviderCard = { input: 1, output: 1, }, - tokens: 1_024_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4-Air 是性价比高的版本,性能接近GLM-4,提供快速度和实惠的价格。', displayName: 'GLM-4-Air', enabled: true, @@ -54,9 +55,9 @@ const ZhiPu: ModelProviderCard = { input: 1, output: 1, }, - tokens: 128_000, }, { + contextWindowTokens: 8192, description: 'GLM-4-AirX 提供 GLM-4-Air 的高效版本,推理速度可达其2.6倍。', displayName: 'GLM-4-AirX', enabled: true, @@ -67,9 +68,9 @@ const ZhiPu: ModelProviderCard = { input: 10, output: 10, }, - tokens: 8192, }, { + contextWindowTokens: 128_000, description: 'GLM-4-AllTools 是一个多功能智能体模型,优化以支持复杂指令规划与工具调用,如网络浏览、代码解释和文本生成,适用于多任务执行。', displayName: 'GLM-4-AllTools', @@ -80,9 +81,9 @@ const ZhiPu: ModelProviderCard = { input: 100, output: 100, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4-Plus 作为高智能旗舰,具备强大的处理长文本和复杂任务的能力,性能全面提升。', displayName: 'GLM-4-Plus', @@ -94,9 +95,9 @@ const ZhiPu: ModelProviderCard = { input: 50, output: 50, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4-0520 是最新模型版本,专为高度复杂和多样化任务设计,表现卓越。', displayName: 'GLM-4-0520', functionCall: true, @@ -106,9 +107,9 @@ const ZhiPu: ModelProviderCard = { input: 100, output: 100, }, - tokens: 128_000, }, { + contextWindowTokens: 128_000, description: 'GLM-4 是发布于2024年1月的旧旗舰版本,目前已被更强的 GLM-4-0520 取代。', displayName: 'GLM-4', functionCall: true, @@ -118,10 +119,11 @@ const ZhiPu: ModelProviderCard = { input: 100, output: 100, }, - tokens: 128_000, }, { - description: 'GLM-4V-Flash 专注于高效的单一图像理解,适用于快速图像解析的场景,例如实时图像分析或批量图像处理。', + contextWindowTokens: 8192, + description: + 'GLM-4V-Flash 专注于高效的单一图像理解,适用于快速图像解析的场景,例如实时图像分析或批量图像处理。', displayName: 'GLM-4V-Flash', enabled: true, id: 'glm-4v-flash', @@ -131,10 +133,10 @@ const ZhiPu: ModelProviderCard = { output: 0, }, releasedAt: '2024-12-09', - tokens: 8192, vision: true, }, { + contextWindowTokens: 8192, description: 'GLM-4V-Plus 具备对视频内容及多图片的理解能力,适合多模态任务。', displayName: 'GLM-4V-Plus', enabled: true, @@ -144,10 +146,10 @@ const ZhiPu: ModelProviderCard = { input: 10, output: 10, }, - tokens: 8192, vision: true, }, { + contextWindowTokens: 2048, description: 'GLM-4V 提供强大的图像理解与推理能力,支持多种视觉任务。', displayName: 'GLM-4V', id: 'glm-4v', @@ -156,10 +158,10 @@ const ZhiPu: ModelProviderCard = { input: 50, output: 50, }, - tokens: 2048, vision: true, }, { + contextWindowTokens: 128_000, description: 'CodeGeeX-4 是强大的AI编程助手,支持多种编程语言的智能问答与代码补全,提升开发效率。', displayName: 'CodeGeeX-4', @@ -169,9 +171,9 @@ const ZhiPu: ModelProviderCard = { input: 0.1, output: 0.1, }, - tokens: 128_000, }, { + contextWindowTokens: 4096, description: 'CharGLM-3 专为角色扮演与情感陪伴设计,支持超长多轮记忆与个性化对话,应用广泛。', displayName: 'CharGLM-3', id: 'charglm-3', @@ -180,9 +182,9 @@ const ZhiPu: ModelProviderCard = { input: 15, output: 15, }, - tokens: 4096, }, { + contextWindowTokens: 8192, description: 'Emohaa 是心理模型,具备专业咨询能力,帮助用户理解情感问题。', displayName: 'Emohaa', id: 'emohaa', @@ -191,7 +193,6 @@ const ZhiPu: ModelProviderCard = { input: 15, output: 15, }, - tokens: 8192, }, ], checkModel: 'glm-4-flash', diff --git a/src/database/_deprecated/core/model.ts b/src/database/_deprecated/core/model.ts index caefefc817af..fe5587617a82 100644 --- a/src/database/_deprecated/core/model.ts +++ b/src/database/_deprecated/core/model.ts @@ -102,7 +102,7 @@ export class BaseModel { const offset = current * pageSize; - return this.db - .select({ - createdAt: topics.createdAt, - favorite: topics.favorite, - historySummary: topics.historySummary, - id: topics.id, - metadata: topics.metadata, - title: topics.title, - updatedAt: topics.updatedAt, - }) - .from(topics) - .where(and(eq(topics.userId, this.userId), this.matchSession(sessionId))) - // In boolean sorting, false is considered "smaller" than true. - // So here we use desc to ensure that topics with favorite as true are in front. - .orderBy(desc(topics.favorite), desc(topics.updatedAt)) - .limit(pageSize) - .offset(offset); + return ( + this.db + .select({ + createdAt: topics.createdAt, + favorite: topics.favorite, + historySummary: topics.historySummary, + id: topics.id, + metadata: topics.metadata, + title: topics.title, + updatedAt: topics.updatedAt, + }) + .from(topics) + .where(and(eq(topics.userId, this.userId), this.matchSession(sessionId))) + // In boolean sorting, false is considered "smaller" than true. + // So here we use desc to ensure that topics with favorite as true are in front. + .orderBy(desc(topics.favorite), desc(topics.updatedAt)) + .limit(pageSize) + .offset(offset) + ); }; findById = async (id: string) => { diff --git a/src/features/ChangelogModal/index.tsx b/src/features/ChangelogModal/index.tsx index fac8bf59d85a..cc527bdafaa0 100644 --- a/src/features/ChangelogModal/index.tsx +++ b/src/features/ChangelogModal/index.tsx @@ -7,11 +7,17 @@ import { memo } from 'react'; import { useGlobalStore } from '@/store/global'; const ChangelogModal = memo<{ currentId?: string }>(({ currentId }) => { - const latestChangelogId = useGlobalStore((s) => s.status.latestChangelogId); + const [latestChangelogId, updateSystemStatus] = useGlobalStore((s) => [ + s.status.latestChangelogId, + s.updateSystemStatus, + ]); const router = useRouter(); useTimeout(() => { - if (latestChangelogId !== currentId) { + if (!currentId) return; + if (!latestChangelogId) { + updateSystemStatus({ latestChangelogId: currentId }); + } else if (latestChangelogId !== currentId) { router.push('/changelog/modal'); } }, 1000); diff --git a/src/features/PluginStore/index.tsx b/src/features/PluginStore/index.tsx index b69372bec372..2440033169a8 100644 --- a/src/features/PluginStore/index.tsx +++ b/src/features/PluginStore/index.tsx @@ -31,8 +31,8 @@ export const PluginStore = memo(({ setOpen, open }) => { title={t('store.title')} width={800} > - diff --git a/src/layout/GlobalProvider/AntdV5MonkeyPatch.tsx b/src/layout/GlobalProvider/AntdV5MonkeyPatch.tsx index deb80e0a1671..71f10c73888e 100644 --- a/src/layout/GlobalProvider/AntdV5MonkeyPatch.tsx +++ b/src/layout/GlobalProvider/AntdV5MonkeyPatch.tsx @@ -1,8 +1,8 @@ -"use client"; +'use client'; -import { useEffect } from "react"; -import { unstableSetRender } from "antd"; -import { createRoot } from "react-dom/client"; +import { unstableSetRender } from 'antd'; +import { useEffect } from 'react'; +import { createRoot } from 'react-dom/client'; const AntdV5MonkeyPatch = () => { useEffect(() => { diff --git a/src/libs/agent-runtime/google/index.ts b/src/libs/agent-runtime/google/index.ts index 5a8c26cfce75..67b226d2ca9b 100644 --- a/src/libs/agent-runtime/google/index.ts +++ b/src/libs/agent-runtime/google/index.ts @@ -289,9 +289,10 @@ export class LobeGoogleAI implements LobeRuntimeAI { const functionDeclaration = tool.function; const parameters = functionDeclaration.parameters; // refs: https://github.com/lobehub/lobe-chat/pull/5002 - const properties = parameters?.properties && Object.keys(parameters.properties).length > 0 - ? parameters.properties - : { dummy: { type: 'string' } }; // dummy property to avoid empty object + const properties = + parameters?.properties && Object.keys(parameters.properties).length > 0 + ? parameters.properties + : { dummy: { type: 'string' } }; // dummy property to avoid empty object return { description: functionDeclaration.description, diff --git a/src/libs/agent-runtime/higress/index.ts b/src/libs/agent-runtime/higress/index.ts index 932204d9701d..6b591ffe8357 100644 --- a/src/libs/agent-runtime/higress/index.ts +++ b/src/libs/agent-runtime/higress/index.ts @@ -23,6 +23,7 @@ export const LobeHigressAI = LobeOpenAICompatibleFactory({ const model = m as any; return { + contextWindowTokens: model.context_length, description: model.description, displayName: model.name, enabled: LOBE_DEFAULT_MODEL_LIST.find((m) => model.id.endsWith(m.id))?.enabled || false, @@ -33,7 +34,6 @@ export const LobeHigressAI = LobeOpenAICompatibleFactory({ typeof model.top_provider.max_completion_tokens === 'number' ? model.top_provider.max_completion_tokens : undefined, - tokens: model.context_length, vision: model.description.includes('vision') || model.description.includes('multimodal') || diff --git a/src/libs/agent-runtime/huggingface/index.ts b/src/libs/agent-runtime/huggingface/index.ts index 552652f9470a..0739830f82fd 100644 --- a/src/libs/agent-runtime/huggingface/index.ts +++ b/src/libs/agent-runtime/huggingface/index.ts @@ -22,11 +22,9 @@ export const LobeHuggingFaceAI = LobeOpenAICompatibleFactory({ }, customClient: { createChatCompletionStream: (client: HfInference, payload, instance) => { - const { max_tokens = 4096} = payload; + const { max_tokens = 4096 } = payload; const hfRes = client.chatCompletionStream({ - endpointUrl: instance.baseURL - ? urlJoin(instance.baseURL, payload.model) - : instance.baseURL, + endpointUrl: instance.baseURL ? urlJoin(instance.baseURL, payload.model) : instance.baseURL, max_tokens: max_tokens, messages: payload.messages, model: payload.model, diff --git a/src/libs/agent-runtime/minimax/index.ts b/src/libs/agent-runtime/minimax/index.ts index 07b0ff5435cb..1e7b51e0c3aa 100644 --- a/src/libs/agent-runtime/minimax/index.ts +++ b/src/libs/agent-runtime/minimax/index.ts @@ -129,8 +129,8 @@ export class LobeMinimaxAI implements LobeRuntimeAI { switch (model) { case 'abab6.5t-chat': case 'abab6.5g-chat': - case 'abab5.5s-chat': - case 'abab5.5-chat':{ + case 'abab5.5s-chat': + case 'abab5.5-chat': { return 4096; } case 'abab6.5s-chat': { @@ -145,16 +145,11 @@ export class LobeMinimaxAI implements LobeRuntimeAI { return { ...params, frequency_penalty: undefined, - max_tokens: - payload.max_tokens !== undefined - ? payload.max_tokens - : this.getMaxTokens(payload.model), + max_tokens: + payload.max_tokens !== undefined ? payload.max_tokens : this.getMaxTokens(payload.model), presence_penalty: undefined, stream: true, - temperature: - temperature === undefined || temperature <= 0 - ? undefined - : temperature / 2, + temperature: temperature === undefined || temperature <= 0 ? undefined : temperature / 2, tools: params.tools?.map((tool) => ({ function: { diff --git a/src/libs/agent-runtime/mistral/index.ts b/src/libs/agent-runtime/mistral/index.ts index 055616d0235b..73fb9d0461c1 100644 --- a/src/libs/agent-runtime/mistral/index.ts +++ b/src/libs/agent-runtime/mistral/index.ts @@ -5,15 +5,12 @@ export const LobeMistralAI = LobeOpenAICompatibleFactory({ baseURL: 'https://api.mistral.ai/v1', chatCompletion: { handlePayload: (payload) => ({ - ...payload.max_tokens !== undefined && { max_tokens: payload.max_tokens }, + ...(payload.max_tokens !== undefined && { max_tokens: payload.max_tokens }), messages: payload.messages as any, model: payload.model, stream: true, - temperature: - payload.temperature !== undefined - ? payload.temperature / 2 - : undefined, - ...payload.tools && { tools: payload.tools }, + temperature: payload.temperature !== undefined ? payload.temperature / 2 : undefined, + ...(payload.tools && { tools: payload.tools }), top_p: payload.top_p, }), noUserId: true, diff --git a/src/libs/agent-runtime/moonshot/index.ts b/src/libs/agent-runtime/moonshot/index.ts index 89e7a061b987..28c976d9dcf5 100644 --- a/src/libs/agent-runtime/moonshot/index.ts +++ b/src/libs/agent-runtime/moonshot/index.ts @@ -9,12 +9,9 @@ export const LobeMoonshotAI = LobeOpenAICompatibleFactory({ handlePayload: (payload: ChatStreamPayload) => { const { temperature, ...rest } = payload; - return { - ...rest, - temperature: - temperature !== undefined - ? temperature / 2 - : undefined, + return { + ...rest, + temperature: temperature !== undefined ? temperature / 2 : undefined, } as OpenAI.ChatCompletionCreateParamsStreaming; }, }, diff --git a/src/libs/agent-runtime/novita/__snapshots__/index.test.ts.snap b/src/libs/agent-runtime/novita/__snapshots__/index.test.ts.snap index 9fddae086328..b765e49432c8 100644 --- a/src/libs/agent-runtime/novita/__snapshots__/index.test.ts.snap +++ b/src/libs/agent-runtime/novita/__snapshots__/index.test.ts.snap @@ -3,149 +3,149 @@ exports[`NovitaAI > models > should get models 1`] = ` [ { + "contextWindowTokens": 8192, "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations.", "displayName": "meta-llama/llama-3-8b-instruct", "enabled": true, "functionCall": false, "id": "meta-llama/llama-3-8b-instruct", - "tokens": 8192, }, { + "contextWindowTokens": 8192, "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations.", "displayName": "meta-llama/llama-3-70b-instruct", "enabled": true, "functionCall": false, "id": "meta-llama/llama-3-70b-instruct", - "tokens": 8192, }, { + "contextWindowTokens": 8192, "description": "Meta's latest class of models, Llama 3.1, launched with a variety of sizes and configurations. The 8B instruct-tuned version is particularly fast and efficient. It has demonstrated strong performance in human evaluations, outperforming several leading closed-source models.", "displayName": "meta-llama/llama-3.1-8b-instruct", "enabled": true, "functionCall": false, "id": "meta-llama/llama-3.1-8b-instruct", - "tokens": 8192, }, { + "contextWindowTokens": 8192, "description": "Meta's latest class of models, Llama 3.1, has launched with a variety of sizes and configurations. The 70B instruct-tuned version is optimized for high-quality dialogue use cases. It has demonstrated strong performance in human evaluations compared to leading closed-source models.", "displayName": "meta-llama/llama-3.1-70b-instruct", "enabled": true, "functionCall": false, "id": "meta-llama/llama-3.1-70b-instruct", - "tokens": 8192, }, { + "contextWindowTokens": 32768, "description": "Meta's latest class of models, Llama 3.1, launched with a variety of sizes and configurations. This 405B instruct-tuned version is optimized for high-quality dialogue use cases. It has demonstrated strong performance compared to leading closed-source models, including GPT-4o and Claude 3.5 Sonnet, in evaluations.", "displayName": "meta-llama/llama-3.1-405b-instruct", "enabled": true, "functionCall": false, "id": "meta-llama/llama-3.1-405b-instruct", - "tokens": 32768, }, { + "contextWindowTokens": 8192, "description": "Gemma 2 9B by Google is an advanced, open-source language model that sets a new standard for efficiency and performance in its size class. Designed for a wide variety of tasks, it empowers developers and researchers to build innovative applications, while maintaining accessibility, safety, and cost-effectiveness.", "displayName": "google/gemma-2-9b-it", "enabled": true, "functionCall": false, "id": "google/gemma-2-9b-it", - "tokens": 8192, }, { + "contextWindowTokens": 4096, "description": "This is a fine-tuned Llama-2 model designed to support longer and more detailed writing prompts, as well as next-chapter generation. It also includes an experimental role-playing instruction set with multi-round dialogues, character interactions, and varying numbers of participants", "displayName": "jondurbin/airoboros-l2-70b", "enabled": true, "functionCall": false, "id": "jondurbin/airoboros-l2-70b", - "tokens": 4096, }, { + "contextWindowTokens": 8192, "description": "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.", "displayName": "nousresearch/hermes-2-pro-llama-3-8b", "enabled": true, "functionCall": true, "id": "nousresearch/hermes-2-pro-llama-3-8b", - "tokens": 8192, }, { + "contextWindowTokens": 32768, "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length.", "displayName": "mistralai/mistral-7b-instruct", "enabled": true, "functionCall": false, "id": "mistralai/mistral-7b-instruct", - "tokens": 32768, }, { + "contextWindowTokens": 16000, "description": "Dolphin 2.9 is designed for instruction following, conversational, and coding. This model is a finetune of Mixtral 8x22B Instruct. It features a 64k context length and was fine-tuned with a 16k sequence length using ChatML templates.The model is uncensored and is stripped of alignment and bias. It requires an external alignment layer for ethical use.", "displayName": "cognitivecomputations/dolphin-mixtral-8x22b", "enabled": true, "functionCall": false, "id": "cognitivecomputations/dolphin-mixtral-8x22b", - "tokens": 16000, }, { + "contextWindowTokens": 16000, "description": "The uncensored llama3 model is a powerhouse of creativity, excelling in both roleplay and story writing. It offers a liberating experience during roleplays, free from any restrictions. This model stands out for its immense creativity, boasting a vast array of unique ideas and plots, truly a treasure trove for those seeking originality. Its unrestricted nature during roleplays allows for the full breadth of imagination to unfold, akin to an enhanced, big-brained version of Stheno. Perfect for creative minds seeking a boundless platform for their imaginative expressions, the uncensored llama3 model is an ideal choice", "displayName": "sao10k/l3-70b-euryale-v2.1", "enabled": true, "functionCall": false, "id": "sao10k/l3-70b-euryale-v2.1", - "tokens": 16000, }, { + "contextWindowTokens": 4096, "description": "A merge with a complex family tree, this model was crafted for roleplaying and storytelling. Midnight Rose is a successor to Rogue Rose and Aurora Nights and improves upon them both. It wants to produce lengthy output by default and is the best creative writing merge produced so far by sophosympatheia.", "displayName": "sophosympatheia/midnight-rose-70b", "enabled": true, "functionCall": false, "id": "sophosympatheia/midnight-rose-70b", - "tokens": 4096, }, { + "contextWindowTokens": 4096, "description": "The idea behind this merge is that each layer is composed of several tensors, which are in turn responsible for specific functions. Using MythoLogic-L2's robust understanding as its input and Huginn's extensive writing capability as its output seems to have resulted in a model that exceeds at both, confirming my theory. (More details to be released at a later time).", "displayName": "gryphe/mythomax-l2-13b", "enabled": true, "functionCall": false, "id": "gryphe/mythomax-l2-13b", - "tokens": 4096, }, { + "contextWindowTokens": 4096, "description": "Nous-Hermes-Llama2-13b is a state-of-the-art language model fine-tuned on over 300,000 instructions. This model was fine-tuned by Nous Research, with Teknium and Emozilla leading the fine tuning process and dataset curation, Redmond AI sponsoring the compute, and several other contributors.", "displayName": "nousresearch/nous-hermes-llama2-13b", "enabled": true, "functionCall": false, "id": "nousresearch/nous-hermes-llama2-13b", - "tokens": 4096, }, { + "contextWindowTokens": 32768, "description": "Nous Hermes 2 Mixtral 8x7B DPO is the new flagship Nous Research model trained over the Mixtral 8x7B MoE LLM. The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.", "displayName": "Nous-Hermes-2-Mixtral-8x7B-DPO", "enabled": true, "functionCall": false, "id": "Nous-Hermes-2-Mixtral-8x7B-DPO", - "tokens": 32768, }, { + "contextWindowTokens": 4096, "description": "A Mythomax/MLewd_13B-style merge of selected 70B models. A multi-model merge of several LLaMA2 70B finetunes for roleplaying and creative work. The goal was to create a model that combines creativity with intelligence for an enhanced experience.", "displayName": "lzlv_70b", "enabled": true, "functionCall": false, "id": "lzlv_70b", - "tokens": 4096, }, { + "contextWindowTokens": 4096, "description": "OpenHermes 2.5 Mistral 7B is a state of the art Mistral Fine-tune, a continuation of OpenHermes 2 model, which trained on additional code datasets.", "displayName": "teknium/openhermes-2.5-mistral-7b", "enabled": true, "functionCall": false, "id": "teknium/openhermes-2.5-mistral-7b", - "tokens": 4096, }, { + "contextWindowTokens": 65535, "description": "WizardLM-2 8x22B is Microsoft AI's most advanced Wizard model. It demonstrates highly competitive performance compared to leading proprietary models, and it consistently outperforms all existing state-of-the-art opensource models.", "displayName": "microsoft/wizardlm-2-8x22b", "enabled": true, "functionCall": false, "id": "microsoft/wizardlm-2-8x22b", - "tokens": 65535, }, ] `; diff --git a/src/libs/agent-runtime/novita/index.ts b/src/libs/agent-runtime/novita/index.ts index 1ac79a565996..aa05831909b8 100644 --- a/src/libs/agent-runtime/novita/index.ts +++ b/src/libs/agent-runtime/novita/index.ts @@ -17,12 +17,12 @@ export const LobeNovitaAI = LobeOpenAICompatibleFactory({ const model = m as unknown as NovitaModelCard; return { + contextWindowTokens: model.context_size, description: model.description, displayName: model.title, enabled: model.status === 1, functionCall: model.description.toLowerCase().includes('function calling'), id: model.id, - tokens: model.context_size, }; }, }, diff --git a/src/libs/agent-runtime/openai/__snapshots__/index.test.ts.snap b/src/libs/agent-runtime/openai/__snapshots__/index.test.ts.snap index 7d2cf1a95bba..95eb7fa4fd26 100644 --- a/src/libs/agent-runtime/openai/__snapshots__/index.test.ts.snap +++ b/src/libs/agent-runtime/openai/__snapshots__/index.test.ts.snap @@ -3,6 +3,7 @@ exports[`LobeOpenAI > models > should get models 1`] = ` [ { + "contextWindowTokens": 16385, "description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125", "displayName": "GPT-3.5 Turbo", "functionCall": true, @@ -11,7 +12,6 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 0.5, "output": 1.5, }, - "tokens": 16385, }, { "id": "gpt-3.5-turbo-16k", @@ -26,6 +26,7 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "id": "gpt-3.5-turbo-instruct-0914", }, { + "contextWindowTokens": 128000, "description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。", "displayName": "GPT-4 Turbo Preview 0125", "functionCall": true, @@ -34,9 +35,9 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 10, "output": 30, }, - "tokens": 128000, }, { + "contextWindowTokens": 128000, "description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。", "displayName": "GPT-4 Turbo Preview", "functionCall": true, @@ -45,9 +46,9 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 10, "output": 30, }, - "tokens": 128000, }, { + "contextWindowTokens": 4096, "description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125", "displayName": "GPT-3.5 Turbo Instruct", "id": "gpt-3.5-turbo-instruct", @@ -55,7 +56,6 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 1.5, "output": 2, }, - "tokens": 4096, }, { "id": "gpt-3.5-turbo-0301", @@ -64,6 +64,7 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "id": "gpt-3.5-turbo-0613", }, { + "contextWindowTokens": 16385, "description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125", "displayName": "GPT-3.5 Turbo 1106", "functionCall": true, @@ -72,9 +73,9 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 1, "output": 2, }, - "tokens": 16385, }, { + "contextWindowTokens": 128000, "description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。", "displayName": "GPT-4 Turbo Preview 1106", "functionCall": true, @@ -83,17 +84,17 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 10, "output": 30, }, - "tokens": 128000, }, { + "contextWindowTokens": 128000, "deploymentName": "gpt-4-vision", "description": "GPT-4 视觉预览版,专为图像分析和处理任务设计。", "displayName": "GPT 4 Turbo with Vision Preview", "id": "gpt-4-vision-preview", - "tokens": 128000, "vision": true, }, { + "contextWindowTokens": 8192, "description": "GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。", "displayName": "GPT-4", "functionCall": true, @@ -102,9 +103,9 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 30, "output": 60, }, - "tokens": 8192, }, { + "contextWindowTokens": 16385, "description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125", "displayName": "GPT-3.5 Turbo 0125", "functionCall": true, @@ -113,9 +114,9 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 0.5, "output": 1.5, }, - "tokens": 16385, }, { + "contextWindowTokens": 8192, "description": "GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。", "displayName": "GPT-4 0613", "functionCall": true, @@ -124,7 +125,6 @@ exports[`LobeOpenAI > models > should get models 1`] = ` "input": 30, "output": 60, }, - "tokens": 8192, }, ] `; diff --git a/src/libs/agent-runtime/openrouter/__snapshots__/index.test.ts.snap b/src/libs/agent-runtime/openrouter/__snapshots__/index.test.ts.snap index 28233af5f021..4d1997bb70ca 100644 --- a/src/libs/agent-runtime/openrouter/__snapshots__/index.test.ts.snap +++ b/src/libs/agent-runtime/openrouter/__snapshots__/index.test.ts.snap @@ -3,6 +3,7 @@ exports[`LobeOpenRouterAI > models > should get models 1`] = ` [ { + "contextWindowTokens": 128000, "description": "Command-R is a 35B parameter model that performs conversational language tasks at a higher quality, more reliably, and with a longer context than previous models. It can be used for complex workflows like code generation, retrieval augmented generation (RAG), tool use, and agents. Read the launch post [here](https://txt.cohere.com/command-r/). @@ -13,10 +14,10 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co "functionCall": false, "id": "cohere/command-r-03-2024", "maxTokens": 4000, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Command R+ is a new, 104B-parameter LLM from Cohere. It's useful for roleplay, general consumer usecases, and Retrieval Augmented Generation (RAG). It offers multilingual support for ten key languages to facilitate global business operations. See benchmarks and the launch post [here](https://txt.cohere.com/command-r-plus-microsoft-azure/). @@ -27,10 +28,10 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co "functionCall": false, "id": "cohere/command-r-plus-04-2024", "maxTokens": 4000, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "command-r-plus-08-2024 is an update of the [Command R+](/models/cohere/command-r-plus) with roughly 50% higher throughput and 25% lower latencies as compared to the previous Command R+ version, while keeping the hardware footprint the same. Read the launch post [here](https://docs.cohere.com/changelog/command-gets-refreshed). @@ -41,10 +42,10 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co "functionCall": false, "id": "cohere/command-r-plus-08-2024", "maxTokens": 4000, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "command-r-08-2024 is an update of the [Command R](/models/cohere/command-r) with improved performance for multilingual retrieval-augmented generation (RAG) and tool use. More broadly, it is better at math, code and reasoning and is competitive with the previous version of the larger Command R+ model. Read the launch post [here](https://docs.cohere.com/changelog/command-gets-refreshed). @@ -55,10 +56,10 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co "functionCall": false, "id": "cohere/command-r-08-2024", "maxTokens": 4000, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 4000000, "description": "Gemini 1.5 Flash 8B Experimental is an experimental, 8B parameter version of the [Gemini 1.5 Flash](/models/google/gemini-flash-1.5) model. Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.dev/terms). @@ -71,10 +72,10 @@ Note: This model is experimental and not suited for production use-cases. It may "functionCall": false, "id": "google/gemini-flash-8b-1.5-exp", "maxTokens": 32768, - "tokens": 4000000, "vision": true, }, { + "contextWindowTokens": 4000000, "description": "Gemini 1.5 Flash Experimental is an experimental version of the [Gemini 1.5 Flash](/models/google/gemini-flash-1.5) model. Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.dev/terms). @@ -87,20 +88,20 @@ Note: This model is experimental and not suited for production use-cases. It may "functionCall": false, "id": "google/gemini-flash-1.5-exp", "maxTokens": 32768, - "tokens": 4000000, "vision": true, }, { + "contextWindowTokens": 8192, "description": "Euryale L3.1 70B v2.2 is a model focused on creative roleplay from [Sao10k](https://ko-fi.com/sao10k). It is the successor of [Euryale L3 70B v2.1](/models/sao10k/l3-euryale-70b).", "displayName": "Llama 3.1 Euryale 70B v2.2", "enabled": false, "functionCall": false, "id": "sao10k/l3.1-euryale-70b", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 256000, "description": "Jamba 1.5 Large is part of AI21's new family of open models, offering superior speed, efficiency, and quality. It features a 256K effective context window, the longest among open models, enabling improved performance on tasks like document summarization and analysis. @@ -113,10 +114,10 @@ Read their [announcement](https://www.ai21.com/blog/announcing-jamba-model-famil "functionCall": false, "id": "ai21/jamba-1-5-large", "maxTokens": 4096, - "tokens": 256000, "vision": false, }, { + "contextWindowTokens": 256000, "description": "Jamba 1.5 Mini is the world's first production-grade Mamba-based model, combining SSM and Transformer architectures for a 256K context window and high efficiency. It works with 9 languages and can handle various writing and analysis tasks as well as or better than similar small models. @@ -129,10 +130,10 @@ Read their [announcement](https://www.ai21.com/blog/announcing-jamba-model-famil "functionCall": false, "id": "ai21/jamba-1-5-mini", "maxTokens": 4096, - "tokens": 256000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Phi-3.5 models are lightweight, state-of-the-art open models. These models were trained with Phi-3 datasets that include both synthetic data and the filtered, publicly available websites data, with a focus on high quality and reasoning-dense properties. Phi-3.5 Mini uses 3.8B parameters, and is a dense decoder-only transformer model using the same tokenizer as [Phi-3 Mini](/models/microsoft/phi-3-mini-128k-instruct). The models underwent a rigorous enhancement process, incorporating both supervised fine-tuning, proximal policy optimization, and direct preference optimization to ensure precise instruction adherence and robust safety measures. When assessed against benchmarks that test common sense, language understanding, math, code, long context and logical reasoning, Phi-3.5 models showcased robust and state-of-the-art performance among models with less than 13 billion parameters.", @@ -141,10 +142,10 @@ The models underwent a rigorous enhancement process, incorporating both supervis "functionCall": false, "id": "microsoft/phi-3.5-mini-128k-instruct", "maxTokens": undefined, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 131072, "description": "Hermes 3 is a generalist language model with many improvements over [Hermes 2](/models/nousresearch/nous-hermes-2-mistral-7b-dpo), including advanced agentic capabilities, much better roleplaying, reasoning, multi-turn conversation, long context coherence, and improvements across the board. Hermes 3 70B is a competitive, if not superior finetune of the [Llama-3.1 70B foundation model](/models/meta-llama/llama-3.1-70b-instruct), focused on aligning LLMs to the user, with powerful steering capabilities and control given to the end user. @@ -155,10 +156,10 @@ The Hermes 3 series builds and expands on the Hermes 2 set of capabilities, incl "functionCall": true, "id": "nousresearch/hermes-3-llama-3.1-70b", "maxTokens": undefined, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 131072, "description": "Hermes 3 is a generalist language model with many improvements over Hermes 2, including advanced agentic capabilities, much better roleplaying, reasoning, multi-turn conversation, long context coherence, and improvements across the board. Hermes 3 405B is a frontier-level, full-parameter finetune of the Llama-3.1 405B foundation model, focused on aligning LLMs to the user, with powerful steering capabilities and control given to the end user. @@ -171,10 +172,10 @@ Hermes 3 is competitive, if not superior, to Llama-3.1 Instruct models at genera "functionCall": true, "id": "nousresearch/hermes-3-llama-3.1-405b", "maxTokens": undefined, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Hermes 3 is a generalist language model with many improvements over Hermes 2, including advanced agentic capabilities, much better roleplaying, reasoning, multi-turn conversation, long context coherence, and improvements across the board. Hermes 3 405B is a frontier-level, full-parameter finetune of the Llama-3.1 405B foundation model, focused on aligning LLMs to the user, with powerful steering capabilities and control given to the end user. @@ -189,20 +190,20 @@ _These are extended-context endpoints for [Hermes 3 405B Instruct](/models/nousr "functionCall": true, "id": "nousresearch/hermes-3-llama-3.1-405b:extended", "maxTokens": undefined, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 127072, "description": "Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. The model is built upon the Llama 3.1 405B and has internet access.", "displayName": "Perplexity: Llama 3.1 Sonar 405B Online", "enabled": true, "functionCall": false, "id": "perplexity/llama-3.1-sonar-huge-128k-online", "maxTokens": undefined, - "tokens": 127072, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Dynamic model continuously updated to the current version of [GPT-4o](/models/openai/gpt-4o) in ChatGPT. Intended for research and evaluation. Note: This model is experimental and not suited for production use-cases. It may be removed or redirected to another model in the future.", @@ -211,10 +212,10 @@ Note: This model is experimental and not suited for production use-cases. It may "functionCall": false, "id": "openai/chatgpt-4o-latest", "maxTokens": 16384, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Lunaris 8B is a versatile generalist and roleplaying model based on Llama 3. It's a strategic merge of multiple models, designed to balance creativity with improved logic and general knowledge. Created by [Sao10k](https://huggingface.co/Sao10k), this model aims to offer an improved experience over Stheno v3.2, with enhanced creativity and logical reasoning. @@ -225,10 +226,10 @@ For best results, use with Llama 3 Instruct context template, temperature 1.4, a "functionCall": false, "id": "sao10k/l3-lunaris-8b", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 12000, "description": "Starcannon 12B is a creative roleplay and story writing model, using [nothingiisreal/mn-celeste-12b](https://openrouter.ai/models/nothingiisreal/mn-celeste-12b) as a base and [intervitens/mini-magnum-12b-v1.1](https://huggingface.co/intervitens/mini-magnum-12b-v1.1) merged in using the [TIES](https://arxiv.org/abs/2306.01708) method. Although more similar to Magnum overall, the model remains very creative, with a pleasant writing style. It is recommended for people wanting more variety than Magnum, and yet more verbose prose than Celeste.", @@ -237,10 +238,10 @@ Although more similar to Magnum overall, the model remains very creative, with a "functionCall": false, "id": "aetherwiing/mn-starcannon-12b", "maxTokens": undefined, - "tokens": 12000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "The 2024-08-06 version of GPT-4o offers improved performance in structured outputs, with the ability to supply a JSON schema in the respone_format. Read more [here](https://openai.com/index/introducing-structured-outputs-in-the-api/). GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities. @@ -251,10 +252,10 @@ For benchmarking against other models, it was briefly called ["im-also-a-good-gp "functionCall": false, "id": "openai/gpt-4o-2024-08-06", "maxTokens": 16384, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 131072, "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This is the base 405B pre-trained version. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -265,10 +266,10 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta "functionCall": false, "id": "meta-llama/llama-3.1-405b", "maxTokens": undefined, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 32000, "description": "A specialized story writing and roleplaying model based on Mistral's NeMo 12B Instruct. Fine-tuned on curated datasets including Reddit Writing Prompts and Opus Instruct 25K. This model excels at creative writing, offering improved NSFW capabilities, with smarter and more active narration. It demonstrates remarkable versatility in both SFW and NSFW scenarios, with strong Out of Character (OOC) steering capabilities, allowing fine-tuned control over narrative direction and character behavior. @@ -279,10 +280,10 @@ Check out the model's [HuggingFace page](https://huggingface.co/nothingiisreal/M "functionCall": false, "id": "nothingiisreal/mn-celeste-12b", "maxTokens": undefined, - "tokens": 32000, "vision": false, }, { + "contextWindowTokens": 4000000, "description": "Gemini 1.5 Pro (0827) is an experimental version of the [Gemini 1.5 Pro](/models/google/gemini-pro-1.5) model. Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.dev/terms). @@ -295,10 +296,10 @@ Note: This model is experimental and not suited for production use-cases. It may "functionCall": false, "id": "google/gemini-pro-1.5-exp", "maxTokens": 32768, - "tokens": 4000000, "vision": true, }, { + "contextWindowTokens": 127072, "description": "Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is the online version of the [offline chat model](/models/perplexity/llama-3.1-sonar-large-128k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online", @@ -307,10 +308,10 @@ This is the online version of the [offline chat model](/models/perplexity/llama- "functionCall": false, "id": "perplexity/llama-3.1-sonar-large-128k-online", "maxTokens": undefined, - "tokens": 127072, "vision": false, }, { + "contextWindowTokens": 131072, "description": "Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is a normal offline LLM, but the [online version](/models/perplexity/llama-3.1-sonar-large-128k-online) of this model has Internet access.", @@ -319,10 +320,10 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama- "functionCall": false, "id": "perplexity/llama-3.1-sonar-large-128k-chat", "maxTokens": undefined, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 127072, "description": "Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is the online version of the [offline chat model](/models/perplexity/llama-3.1-sonar-small-128k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online", @@ -331,10 +332,10 @@ This is the online version of the [offline chat model](/models/perplexity/llama- "functionCall": false, "id": "perplexity/llama-3.1-sonar-small-128k-online", "maxTokens": undefined, - "tokens": 127072, "vision": false, }, { + "contextWindowTokens": 131072, "description": "Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is a normal offline LLM, but the [online version](/models/perplexity/llama-3.1-sonar-small-128k-online) of this model has Internet access.", @@ -343,10 +344,10 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama- "functionCall": false, "id": "perplexity/llama-3.1-sonar-small-128k-chat", "maxTokens": undefined, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 131072, "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -357,10 +358,10 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta "functionCall": false, "id": "meta-llama/llama-3.1-70b-instruct", "maxTokens": undefined, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 131072, "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 8B instruct-tuned version is fast and efficient. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -373,10 +374,10 @@ _These are free, rate-limited endpoints for [Llama 3.1 8B Instruct](/models/meta "functionCall": false, "id": "meta-llama/llama-3.1-8b-instruct:free", "maxTokens": 4096, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 131072, "description": "Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 8B instruct-tuned version is fast and efficient. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -387,10 +388,10 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta "functionCall": false, "id": "meta-llama/llama-3.1-8b-instruct", "maxTokens": undefined, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 131072, "description": "The highly anticipated 400B class of Llama3 is here! Clocking in at 128k context with impressive eval scores, the Meta AI team continues to push the frontier of open-source LLMs. Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 405B instruct-tuned version is optimized for high quality dialogue usecases. @@ -403,10 +404,10 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta "functionCall": false, "id": "meta-llama/llama-3.1-405b-instruct", "maxTokens": undefined, - "tokens": 131072, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Dolphin 2.9 is designed for instruction following, conversational, and coding. This model is a fine-tune of [Llama 3 70B](/models/meta-llama/llama-3-70b-instruct). It demonstrates improvements in instruction, conversation, coding, and function calling abilities, when compared to the original. Uncensored and is stripped of alignment and bias, it requires an external alignment layer for ethical use. Users are cautioned to use this highly compliant model responsibly, as detailed in a blog post about uncensored models at [erichartford.com/uncensored-models](https://erichartford.com/uncensored-models). @@ -417,10 +418,10 @@ Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.m "functionCall": true, "id": "cognitivecomputations/dolphin-llama-3-70b", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 256000, "description": "A 7.3B parameter Mamba-based model designed for code and reasoning tasks. - Linear time inference, allowing for theoretically infinite sequence lengths @@ -433,10 +434,10 @@ Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.m "functionCall": false, "id": "mistralai/codestral-mamba", "maxTokens": undefined, - "tokens": 256000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA. The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi. @@ -447,10 +448,10 @@ It supports function calling and is released under the Apache 2.0 license.", "functionCall": true, "id": "mistralai/mistral-nemo", "maxTokens": undefined, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs. As their most advanced small model, it is many multiples more affordable than other recent frontier models, and more than 60% cheaper than [GPT-3.5 Turbo](/models/openai/gpt-3.5-turbo). It maintains SOTA intelligence, while being significantly more cost-effective. @@ -463,10 +464,10 @@ Check out the [launch announcement](https://openai.com/index/gpt-4o-mini-advanci "functionCall": false, "id": "openai/gpt-4o-mini-2024-07-18", "maxTokens": 16384, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs. As their most advanced small model, it is many multiples more affordable than other recent frontier models, and more than 60% cheaper than [GPT-3.5 Turbo](/models/openai/gpt-3.5-turbo). It maintains SOTA intelligence, while being significantly more cost-effective. @@ -479,10 +480,10 @@ Check out the [launch announcement](https://openai.com/index/gpt-4o-mini-advanci "functionCall": false, "id": "openai/gpt-4o-mini", "maxTokens": 16384, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 32768, "description": "Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning. It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization. @@ -497,10 +498,10 @@ _These are free, rate-limited endpoints for [Qwen 2 7B Instruct](/models/qwen/qw "functionCall": false, "id": "qwen/qwen-2-7b-instruct:free", "maxTokens": 4096, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "Qwen2 7B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning. It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization. @@ -513,10 +514,10 @@ Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://hug "functionCall": false, "id": "qwen/qwen-2-7b-instruct", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Gemma 2 27B by Google is an open model built from the same research and technology used to create the [Gemini models](/models?q=gemini). Gemma models are well-suited for a variety of text generation tasks, including question answering, summarization, and reasoning. @@ -527,10 +528,10 @@ See the [launch announcement](https://blog.google/technology/developers/google-g "functionCall": false, "id": "google/gemma-2-27b-it", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 16384, "description": "From the maker of [Goliath](https://openrouter.ai/models/alpindale/goliath-120b), Magnum 72B is the first in a new family of models designed to achieve the prose quality of the Claude 3 models, notably Opus & Sonnet. The model is based on [Qwen2 72B](https://openrouter.ai/models/qwen/qwen-2-72b-instruct) and trained with 55 million tokens of highly curated roleplay (RP) data.", @@ -539,10 +540,10 @@ The model is based on [Qwen2 72B](https://openrouter.ai/models/qwen/qwen-2-72b-i "functionCall": false, "id": "alpindale/magnum-72b", "maxTokens": 1024, - "tokens": 16384, "vision": false, }, { + "contextWindowTokens": 16384, "description": "An experimental merge model based on Llama 3, exhibiting a very distinctive style of writing. It combines the the best of [Meta's Llama 3 8B](https://openrouter.ai/models/meta-llama/llama-3-8b-instruct) and Nous Research's [Hermes 2 Pro](https://openrouter.ai/models/nousresearch/hermes-2-pro-llama-3-8b). Hermes-2 Θ (theta) was specifically designed with a few capabilities in mind: executing function calls, generating JSON output, and most remarkably, demonstrating metacognitive abilities (contemplating the nature of thought and recognizing the diversity of cognitive processes among individuals).", @@ -551,10 +552,10 @@ Hermes-2 Θ (theta) was specifically designed with a few capabilities in mind: e "functionCall": false, "id": "nousresearch/hermes-2-theta-llama-3-8b", "maxTokens": 2048, - "tokens": 16384, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Gemma 2 9B by Google is an advanced, open-source language model that sets a new standard for efficiency and performance in its size class. Designed for a wide variety of tasks, it empowers developers and researchers to build innovative applications, while maintaining accessibility, safety, and cost-effectiveness. @@ -567,10 +568,10 @@ _These are free, rate-limited endpoints for [Gemma 2 9B](/models/google/gemma-2- "functionCall": false, "id": "google/gemma-2-9b-it:free", "maxTokens": 2048, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Gemma 2 9B by Google is an advanced, open-source language model that sets a new standard for efficiency and performance in its size class. Designed for a wide variety of tasks, it empowers developers and researchers to build innovative applications, while maintaining accessibility, safety, and cost-effectiveness. @@ -581,10 +582,10 @@ See the [launch announcement](https://blog.google/technology/developers/google-g "functionCall": false, "id": "google/gemma-2-9b-it", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 32000, "description": "Stheno 8B 32K is a creative writing/roleplay model from [Sao10k](https://ko-fi.com/sao10k). It was trained at 8K context, then expanded to 32K context. Compared to older Stheno version, this model is trained on: @@ -596,10 +597,10 @@ Compared to older Stheno version, this model is trained on: "functionCall": false, "id": "sao10k/l3-stheno-8b", "maxTokens": undefined, - "tokens": 32000, "vision": false, }, { + "contextWindowTokens": 256000, "description": "The Jamba-Instruct model, introduced by AI21 Labs, is an instruction-tuned variant of their hybrid SSM-Transformer Jamba model, specifically optimized for enterprise applications. - 256K Context Window: It can process extensive information, equivalent to a 400-page novel, which is beneficial for tasks involving large documents such as financial reports or legal documents @@ -613,10 +614,10 @@ Jamba has a knowledge cutoff of February 2024.", "functionCall": false, "id": "ai21/jamba-instruct", "maxTokens": 4096, - "tokens": 256000, "vision": false, }, { + "contextWindowTokens": 200000, "description": "Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at: - Coding: Autonomously writes, edits, and runs code with reasoning and troubleshooting @@ -630,10 +631,10 @@ Jamba has a knowledge cutoff of February 2024.", "functionCall": true, "id": "anthropic/claude-3.5-sonnet", "maxTokens": 8192, - "tokens": 200000, "vision": true, }, { + "contextWindowTokens": 200000, "description": "Claude 3.5 Sonnet delivers better-than-Opus capabilities, faster-than-Sonnet speeds, at the same Sonnet prices. Sonnet is particularly good at: - Coding: Autonomously writes, edits, and runs code with reasoning and troubleshooting @@ -649,10 +650,10 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": true, "id": "anthropic/claude-3.5-sonnet:beta", "maxTokens": 8192, - "tokens": 200000, "vision": true, }, { + "contextWindowTokens": 8192, "description": "Euryale 70B v2.1 is a model focused on creative roleplay from [Sao10k](https://ko-fi.com/sao10k). - Better prompt adherence. @@ -665,10 +666,10 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": false, "id": "sao10k/l3-euryale-70b", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 4000, "description": "Phi-3 4K Medium is a powerful 14-billion parameter model designed for advanced language understanding, reasoning, and instruction following. Optimized through supervised fine-tuning and preference adjustments, it excels in tasks involving common sense, mathematics, logical reasoning, and code processing. At time of release, Phi-3 Medium demonstrated state-of-the-art performance among lightweight models. In the MMLU-Pro eval, the model even comes close to a Llama3 70B level of performance. @@ -679,10 +680,10 @@ For 128k context length, try [Phi-3 Medium 128K](/models/microsoft/phi-3-medium- "functionCall": false, "id": "microsoft/phi-3-medium-4k-instruct", "maxTokens": undefined, - "tokens": 4000, "vision": false, }, { + "contextWindowTokens": 65536, "description": "Dolphin 2.9 is designed for instruction following, conversational, and coding. This model is a finetune of [Mixtral 8x22B Instruct](/models/mistralai/mixtral-8x22b-instruct). It features a 64k context length and was fine-tuned with a 16k sequence length using ChatML templates. This model is a successor to [Dolphin Mixtral 8x7B](/models/cognitivecomputations/dolphin-mixtral-8x7b). @@ -695,10 +696,10 @@ The model is uncensored and is stripped of alignment and bias. It requires an ex "functionCall": false, "id": "cognitivecomputations/dolphin-mixtral-8x22b", "maxTokens": undefined, - "tokens": 65536, "vision": false, }, { + "contextWindowTokens": 32768, "description": "Qwen2 72B is a transformer-based model that excels in language understanding, multilingual capabilities, coding, mathematics, and reasoning. It features SwiGLU activation, attention QKV bias, and group query attention. It is pretrained on extensive data with supervised finetuning and direct preference optimization. @@ -711,10 +712,10 @@ Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://hug "functionCall": false, "id": "qwen/qwen-2-72b-instruct", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 8192, "description": "OpenChat 8B is a library of open-source language models, fine-tuned with "C-RLFT (Conditioned Reinforcement Learning Fine-Tuning)" - a strategy inspired by offline reinforcement learning. It has been trained on mixed-quality data without preference labels. It outperforms many similarly sized models including [Llama 3 8B Instruct](/models/meta-llama/llama-3-8b-instruct) and various fine-tuned models. It excels in general conversation, coding assistance, and mathematical reasoning. @@ -728,20 +729,20 @@ It outperforms many similarly sized models including [Llama 3 8B Instruct](/mode "functionCall": false, "id": "openchat/openchat-8b", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Hermes 2 Pro is an upgraded, retrained version of Nous Hermes 2, consisting of an updated and cleaned version of the OpenHermes 2.5 Dataset, as well as a newly introduced Function Calling and JSON Mode dataset developed in-house.", "displayName": "NousResearch: Hermes 2 Pro - Llama-3 8B", "enabled": false, "functionCall": false, "id": "nousresearch/hermes-2-pro-llama-3-8b", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 32768, "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. An improved version of [Mistral 7B Instruct v0.2](/models/mistralai/mistral-7b-instruct-v0.2), with the following changes: @@ -756,10 +757,10 @@ NOTE: Support for function calling depends on the provider.", "functionCall": true, "id": "mistralai/mistral-7b-instruct-v0.3", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. *Mistral 7B Instruct has multiple version variants, and this is intended to be the latest version.* @@ -770,10 +771,10 @@ _These are free, rate-limited endpoints for [Mistral 7B Instruct](/models/mistra "functionCall": false, "id": "mistralai/mistral-7b-instruct:free", "maxTokens": 4096, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. *Mistral 7B Instruct has multiple version variants, and this is intended to be the latest version.*", @@ -782,10 +783,10 @@ _These are free, rate-limited endpoints for [Mistral 7B Instruct](/models/mistra "functionCall": false, "id": "mistralai/mistral-7b-instruct", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. *Mistral 7B Instruct has multiple version variants, and this is intended to be the latest version.* @@ -796,10 +797,10 @@ _These are higher-throughput endpoints for [Mistral 7B Instruct](/models/mistral "functionCall": false, "id": "mistralai/mistral-7b-instruct:nitro", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Phi-3 Mini is a powerful 3.8B parameter model designed for advanced language understanding, reasoning, and instruction following. Optimized through supervised fine-tuning and preference adjustments, it excels in tasks involving common sense, mathematics, logical reasoning, and code processing. At time of release, Phi-3 Medium demonstrated state-of-the-art performance among lightweight models. This model is static, trained on an offline dataset with an October 2023 cutoff date. @@ -810,10 +811,10 @@ _These are free, rate-limited endpoints for [Phi-3 Mini 128K Instruct](/models/m "functionCall": false, "id": "microsoft/phi-3-mini-128k-instruct:free", "maxTokens": 4096, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Phi-3 Mini is a powerful 3.8B parameter model designed for advanced language understanding, reasoning, and instruction following. Optimized through supervised fine-tuning and preference adjustments, it excels in tasks involving common sense, mathematics, logical reasoning, and code processing. At time of release, Phi-3 Medium demonstrated state-of-the-art performance among lightweight models. This model is static, trained on an offline dataset with an October 2023 cutoff date.", @@ -822,10 +823,10 @@ At time of release, Phi-3 Medium demonstrated state-of-the-art performance among "functionCall": false, "id": "microsoft/phi-3-mini-128k-instruct", "maxTokens": undefined, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Phi-3 128K Medium is a powerful 14-billion parameter model designed for advanced language understanding, reasoning, and instruction following. Optimized through supervised fine-tuning and preference adjustments, it excels in tasks involving common sense, mathematics, logical reasoning, and code processing. At time of release, Phi-3 Medium demonstrated state-of-the-art performance among lightweight models. In the MMLU-Pro eval, the model even comes close to a Llama3 70B level of performance. @@ -838,10 +839,10 @@ _These are free, rate-limited endpoints for [Phi-3 Medium 128K Instruct](/models "functionCall": false, "id": "microsoft/phi-3-medium-128k-instruct:free", "maxTokens": 4096, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Phi-3 128K Medium is a powerful 14-billion parameter model designed for advanced language understanding, reasoning, and instruction following. Optimized through supervised fine-tuning and preference adjustments, it excels in tasks involving common sense, mathematics, logical reasoning, and code processing. At time of release, Phi-3 Medium demonstrated state-of-the-art performance among lightweight models. In the MMLU-Pro eval, the model even comes close to a Llama3 70B level of performance. @@ -852,10 +853,10 @@ For 4k context length, try [Phi-3 Medium 4K](/models/microsoft/phi-3-medium-4k-i "functionCall": false, "id": "microsoft/phi-3-medium-128k-instruct", "maxTokens": undefined, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 8192, "description": "The NeverSleep team is back, with a Llama 3 70B finetune trained on their curated roleplay data. Striking a balance between eRP and RP, Lumimaid was designed to be serious, yet uncensored when necessary. To enhance it's overall intelligence and chat capability, roughly 40% of the training data was not roleplay. This provides a breadth of knowledge to access, while still keeping roleplay as the primary strength. @@ -866,10 +867,10 @@ Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.m "functionCall": false, "id": "neversleep/llama-3-lumimaid-70b", "maxTokens": 2048, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 4000000, "description": "Gemini 1.5 Flash is a foundation model that performs well at a variety of multimodal tasks such as visual understanding, classification, summarization, and creating content from image, audio and video. It's adept at processing visual and text inputs such as photographs, documents, infographics, and screenshots. Gemini 1.5 Flash is designed for high-volume, high-frequency tasks where cost and latency matter. On most common tasks, Flash achieves comparable quality to other Gemini Pro models at a significantly reduced cost. Flash is well-suited for applications like chat assistants and on-demand content generation where speed and scale matter. @@ -882,10 +883,10 @@ Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.d "functionCall": false, "id": "google/gemini-flash-1.5", "maxTokens": 32768, - "tokens": 4000000, "vision": true, }, { + "contextWindowTokens": 128000, "description": "DeepSeek-Coder-V2, an open-source Mixture-of-Experts (MoE) code language model. It is further pre-trained from an intermediate checkpoint of DeepSeek-V2 with additional 6 trillion tokens. The original V1 model was trained from scratch on 2T tokens, with a composition of 87% code and 13% natural language in both English and Chinese. It was pre-trained on project-level code corpus by employing a extra fill-in-the-blank task.", @@ -894,10 +895,10 @@ The original V1 model was trained from scratch on 2T tokens, with a composition "functionCall": false, "id": "deepseek/deepseek-coder", "maxTokens": 4096, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "DeepSeek-V2 Chat is a conversational finetune of DeepSeek-V2, a Mixture-of-Experts (MoE) language model. It comprises 236B total parameters, of which 21B are activated for each token. Compared with DeepSeek 67B, DeepSeek-V2 achieves stronger performance, and meanwhile saves 42.5% of training costs, reduces the KV cache by 93.3%, and boosts the maximum generation throughput to 5.76 times. @@ -908,10 +909,10 @@ DeepSeek-V2 achieves remarkable performance on both standard benchmarks and open "functionCall": false, "id": "deepseek/deepseek-chat", "maxTokens": 4096, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 28000, "description": "Llama3 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is the online version of the [offline chat model](/models/perplexity/llama-3-sonar-large-32k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online", @@ -920,10 +921,10 @@ This is the online version of the [offline chat model](/models/perplexity/llama- "functionCall": false, "id": "perplexity/llama-3-sonar-large-32k-online", "maxTokens": undefined, - "tokens": 28000, "vision": false, }, { + "contextWindowTokens": 32768, "description": "Llama3 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is a normal offline LLM, but the [online version](/models/perplexity/llama-3-sonar-large-32k-online) of this model has Internet access.", @@ -932,10 +933,10 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama- "functionCall": false, "id": "perplexity/llama-3-sonar-large-32k-chat", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 28000, "description": "Llama3 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is the online version of the [offline chat model](/models/perplexity/llama-3-sonar-small-32k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online", @@ -944,10 +945,10 @@ This is the online version of the [offline chat model](/models/perplexity/llama- "functionCall": false, "id": "perplexity/llama-3-sonar-small-32k-online", "maxTokens": undefined, - "tokens": 28000, "vision": false, }, { + "contextWindowTokens": 32768, "description": "Llama3 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is a normal offline LLM, but the [online version](/models/perplexity/llama-3-sonar-small-32k-online) of this model has Internet access.", @@ -956,10 +957,10 @@ This is a normal offline LLM, but the [online version](/models/perplexity/llama- "functionCall": false, "id": "perplexity/llama-3-sonar-small-32k-chat", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 8192, "description": "This safeguard model has 8B parameters and is based on the Llama 3 family. Just like is predecessor, [LlamaGuard 1](https://huggingface.co/meta-llama/LlamaGuard-7b), it can do both prompt and response classification. LlamaGuard 2 acts as a normal LLM would, generating text that indicates whether the given input/output is safe/unsafe. If deemed unsafe, it will also share the content categories violated. @@ -974,10 +975,10 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta "functionCall": false, "id": "meta-llama/llama-guard-2-8b", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 128000, "description": "GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities. For benchmarking against other models, it was briefly called ["im-also-a-good-gpt2-chatbot"](https://twitter.com/LiamFedus/status/1790064963966370209)", @@ -986,10 +987,10 @@ For benchmarking against other models, it was briefly called ["im-also-a-good-gp "functionCall": false, "id": "openai/gpt-4o-2024-05-13", "maxTokens": 4096, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "GPT-4o ("o" for "omni") is OpenAI's latest AI model, supporting both text and image inputs with text outputs. It maintains the intelligence level of [GPT-4 Turbo](/models/openai/gpt-4-turbo) while being twice as fast and 50% more cost-effective. GPT-4o also offers improved performance in processing non-English languages and enhanced visual capabilities. For benchmarking against other models, it was briefly called ["im-also-a-good-gpt2-chatbot"](https://twitter.com/LiamFedus/status/1790064963966370209)", @@ -998,10 +999,10 @@ For benchmarking against other models, it was briefly called ["im-also-a-good-gp "functionCall": false, "id": "openai/gpt-4o", "maxTokens": 4096, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "GPT-4o Extended is an experimental variant of GPT-4o with an extended max output tokens. This model supports only text input to text output. _These are extended-context endpoints for [GPT-4o](/models/openai/gpt-4o). They may have higher prices._", @@ -1010,10 +1011,10 @@ _These are extended-context endpoints for [GPT-4o](/models/openai/gpt-4o). They "functionCall": false, "id": "openai/gpt-4o:extended", "maxTokens": 64000, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 32768, "description": "Qwen1.5 72B is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. In comparison with the previous released Qwen, the improvements include: - Significant performance improvement in human preference for chat models @@ -1028,10 +1029,10 @@ Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://hug "functionCall": false, "id": "qwen/qwen-72b-chat", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "Qwen1.5 110B is the beta version of Qwen2, a transformer-based decoder-only language model pretrained on a large amount of data. In comparison with the previous released Qwen, the improvements include: - Significant performance improvement in human preference for chat models @@ -1046,10 +1047,10 @@ Usage of this model is subject to [Tongyi Qianwen LICENSE AGREEMENT](https://hug "functionCall": false, "id": "qwen/qwen-110b-chat", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 24576, "description": "The NeverSleep team is back, with a Llama 3 8B finetune trained on their curated roleplay data. Striking a balance between eRP and RP, Lumimaid was designed to be serious, yet uncensored when necessary. To enhance it's overall intelligence and chat capability, roughly 40% of the training data was not roleplay. This provides a breadth of knowledge to access, while still keeping roleplay as the primary strength. @@ -1060,10 +1061,10 @@ Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.m "functionCall": false, "id": "neversleep/llama-3-lumimaid-8b", "maxTokens": undefined, - "tokens": 24576, "vision": false, }, { + "contextWindowTokens": 24576, "description": "The NeverSleep team is back, with a Llama 3 8B finetune trained on their curated roleplay data. Striking a balance between eRP and RP, Lumimaid was designed to be serious, yet uncensored when necessary. To enhance it's overall intelligence and chat capability, roughly 40% of the training data was not roleplay. This provides a breadth of knowledge to access, while still keeping roleplay as the primary strength. @@ -1076,10 +1077,10 @@ _These are extended-context endpoints for [Llama 3 Lumimaid 8B](/models/neversle "functionCall": false, "id": "neversleep/llama-3-lumimaid-8b:extended", "maxTokens": 2048, - "tokens": 24576, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Creative writing model, routed with permission. It's fast, it keeps the conversation going, and it stays in character. If you submit a raw prompt, you can use Alpaca or Vicuna formats.", @@ -1088,10 +1089,10 @@ If you submit a raw prompt, you can use Alpaca or Vicuna formats.", "functionCall": false, "id": "sao10k/fimbulvetr-11b-v2", "maxTokens": 2048, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -1102,10 +1103,10 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta "functionCall": false, "id": "meta-llama/llama-3-70b-instruct", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 70B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -1118,10 +1119,10 @@ _These are higher-throughput endpoints for [Llama 3 70B Instruct](/models/meta-l "functionCall": false, "id": "meta-llama/llama-3-70b-instruct:nitro", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -1134,10 +1135,10 @@ _These are free, rate-limited endpoints for [Llama 3 8B Instruct](/models/meta-l "functionCall": false, "id": "meta-llama/llama-3-8b-instruct:free", "maxTokens": 4096, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -1148,10 +1149,10 @@ To read more about the model release, [click here](https://ai.meta.com/blog/meta "functionCall": false, "id": "meta-llama/llama-3-8b-instruct", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -1164,10 +1165,10 @@ _These are higher-throughput endpoints for [Llama 3 8B Instruct](/models/meta-ll "functionCall": false, "id": "meta-llama/llama-3-8b-instruct:nitro", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 16384, "description": "Meta's latest class of model (Llama 3) launched with a variety of sizes & flavors. This 8B instruct-tuned version was optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. @@ -1180,10 +1181,10 @@ _These are extended-context endpoints for [Llama 3 8B Instruct](/models/meta-lla "functionCall": false, "id": "meta-llama/llama-3-8b-instruct:extended", "maxTokens": 2048, - "tokens": 16384, "vision": false, }, { + "contextWindowTokens": 65536, "description": "Mistral's official instruct fine-tuned version of [Mixtral 8x22B](/models/mistralai/mixtral-8x22b). It uses 39B active parameters out of 141B, offering unparalleled cost efficiency for its size. Its strengths include: - strong math, coding, and reasoning - large context length (64k) @@ -1196,10 +1197,10 @@ See benchmarks on the launch announcement [here](https://mistral.ai/news/mixtral "functionCall": false, "id": "mistralai/mixtral-8x22b-instruct", "maxTokens": undefined, - "tokens": 65536, "vision": false, }, { + "contextWindowTokens": 32000, "description": "WizardLM-2 7B is the smaller variant of Microsoft AI's latest Wizard model. It is the fastest and achieves comparable performance with existing 10x larger opensource leading models It is a finetune of [Mistral 7B Instruct](/models/mistralai/mistral-7b-instruct), using the same technique as [WizardLM-2 8x22B](/models/microsoft/wizardlm-2-8x22b). @@ -1212,10 +1213,10 @@ To read more about the model release, [click here](https://wizardlm.github.io/Wi "functionCall": false, "id": "microsoft/wizardlm-2-7b", "maxTokens": undefined, - "tokens": 32000, "vision": false, }, { + "contextWindowTokens": 65536, "description": "WizardLM-2 8x22B is Microsoft AI's most advanced Wizard model. It demonstrates highly competitive performance compared to leading proprietary models, and it consistently outperforms all existing state-of-the-art opensource models. It is an instruct finetune of [Mixtral 8x22B](/models/mistralai/mixtral-8x22b). @@ -1228,10 +1229,10 @@ To read more about the model release, [click here](https://wizardlm.github.io/Wi "functionCall": false, "id": "microsoft/wizardlm-2-8x22b", "maxTokens": undefined, - "tokens": 65536, "vision": false, }, { + "contextWindowTokens": 4000000, "description": "Google's latest multimodal model, supporting image and video in text or chat prompts. Optimized for language tasks including: @@ -1253,10 +1254,10 @@ Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.d "functionCall": false, "id": "google/gemini-pro-1.5", "maxTokens": 32768, - "tokens": 4000000, "vision": true, }, { + "contextWindowTokens": 128000, "description": "The latest GPT-4 Turbo model with vision capabilities. Vision requests can now use JSON mode and function calling. Training data: up to December 2023.", @@ -1265,10 +1266,10 @@ Training data: up to December 2023.", "functionCall": true, "id": "openai/gpt-4-turbo", "maxTokens": 4096, - "tokens": 128000, "vision": true, }, { + "contextWindowTokens": 128000, "description": "Command R+ is a new, 104B-parameter LLM from Cohere. It's useful for roleplay, general consumer usecases, and Retrieval Augmented Generation (RAG). It offers multilingual support for ten key languages to facilitate global business operations. See benchmarks and the launch post [here](https://txt.cohere.com/command-r-plus-microsoft-azure/). @@ -1279,10 +1280,10 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co "functionCall": false, "id": "cohere/command-r-plus", "maxTokens": 4000, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 32768, "description": "DBRX is a new open source large language model developed by Databricks. At 132B, it outperforms existing open source LLMs like Llama 2 70B and [Mixtral-8x7b](/models/mistralai/mixtral-8x7b) on standard industry benchmarks for language understanding, programming, math, and logic. It uses a fine-grained mixture-of-experts (MoE) architecture. 36B parameters are active on any input. It was pre-trained on 12T tokens of text and code data. Compared to other open MoE models like Mixtral-8x7B and Grok-1, DBRX is fine-grained, meaning it uses a larger number of smaller experts. @@ -1295,10 +1296,10 @@ See the launch announcement and benchmark results [here](https://www.databricks. "functionCall": false, "id": "databricks/dbrx-instruct", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A merge with a complex family tree, this model was crafted for roleplaying and storytelling. Midnight Rose is a successor to Rogue Rose and Aurora Nights and improves upon them both. It wants to produce lengthy output by default and is the best creative writing merge produced so far by sophosympatheia. Descending from earlier versions of Midnight Rose and [Wizard Tulu Dolphin 70B](https://huggingface.co/sophosympatheia/Wizard-Tulu-Dolphin-70B-v1.0), it inherits the best qualities of each.", @@ -1307,10 +1308,10 @@ Descending from earlier versions of Midnight Rose and [Wizard Tulu Dolphin 70B]( "functionCall": false, "id": "sophosympatheia/midnight-rose-70b", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Command-R is a 35B parameter model that performs conversational language tasks at a higher quality, more reliably, and with a longer context than previous models. It can be used for complex workflows like code generation, retrieval augmented generation (RAG), tool use, and agents. Read the launch post [here](https://txt.cohere.com/command-r/). @@ -1321,10 +1322,10 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co "functionCall": false, "id": "cohere/command-r", "maxTokens": 4000, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 4096, "description": "Command is an instruction-following conversational model that performs language tasks with high quality, more reliably and with a longer context than our base generative models. Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.cohere.com/docs/c4ai-acceptable-use-policy).", @@ -1333,10 +1334,10 @@ Use of this model is subject to Cohere's [Acceptable Use Policy](https://docs.co "functionCall": false, "id": "cohere/command", "maxTokens": 4000, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 200000, "description": "Claude 3 Haiku is Anthropic's fastest and most compact model for near-instant responsiveness. Quick and accurate targeted performance. @@ -1348,10 +1349,10 @@ See the launch announcement and benchmark results [here](https://www.anthropic.c "functionCall": false, "id": "anthropic/claude-3-haiku", "maxTokens": 4096, - "tokens": 200000, "vision": true, }, { + "contextWindowTokens": 200000, "description": "Claude 3 Haiku is Anthropic's fastest and most compact model for near-instant responsiveness. Quick and accurate targeted performance. @@ -1365,10 +1366,10 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": false, "id": "anthropic/claude-3-haiku:beta", "maxTokens": 4096, - "tokens": 200000, "vision": true, }, { + "contextWindowTokens": 200000, "description": "Claude 3 Sonnet is an ideal balance of intelligence and speed for enterprise workloads. Maximum utility at a lower price, dependable, balanced for scaled deployments. See the launch announcement and benchmark results [here](https://www.anthropic.com/news/claude-3-family) @@ -1379,10 +1380,10 @@ See the launch announcement and benchmark results [here](https://www.anthropic.c "functionCall": false, "id": "anthropic/claude-3-sonnet", "maxTokens": 4096, - "tokens": 200000, "vision": true, }, { + "contextWindowTokens": 200000, "description": "Claude 3 Sonnet is an ideal balance of intelligence and speed for enterprise workloads. Maximum utility at a lower price, dependable, balanced for scaled deployments. See the launch announcement and benchmark results [here](https://www.anthropic.com/news/claude-3-family) @@ -1395,10 +1396,10 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": false, "id": "anthropic/claude-3-sonnet:beta", "maxTokens": 4096, - "tokens": 200000, "vision": true, }, { + "contextWindowTokens": 200000, "description": "Claude 3 Opus is Anthropic's most powerful model for highly complex tasks. It boasts top-level performance, intelligence, fluency, and understanding. See the launch announcement and benchmark results [here](https://www.anthropic.com/news/claude-3-family) @@ -1409,10 +1410,10 @@ See the launch announcement and benchmark results [here](https://www.anthropic.c "functionCall": false, "id": "anthropic/claude-3-opus", "maxTokens": 4096, - "tokens": 200000, "vision": true, }, { + "contextWindowTokens": 200000, "description": "Claude 3 Opus is Anthropic's most powerful model for highly complex tasks. It boasts top-level performance, intelligence, fluency, and understanding. See the launch announcement and benchmark results [here](https://www.anthropic.com/news/claude-3-family) @@ -1425,10 +1426,10 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": false, "id": "anthropic/claude-3-opus:beta", "maxTokens": 4096, - "tokens": 200000, "vision": true, }, { + "contextWindowTokens": 128000, "description": "This is Mistral AI's flagship model, Mistral Large 2 (version \`mistral-large-2407\`). It's a proprietary weights-available model and excels at reasoning, code, JSON, chat, and more. Read the launch announcement [here](https://mistral.ai/news/mistral-large-2407/). It is fluent in English, French, Spanish, German, and Italian, with high grammatical accuracy, and its long context window allows precise information recall from large documents.", @@ -1437,10 +1438,10 @@ It is fluent in English, French, Spanish, German, and Italian, with high grammat "functionCall": false, "id": "mistralai/mistral-large", "maxTokens": undefined, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "The preview GPT-4 model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Dec 2023. **Note:** heavily rate limited by OpenAI while in preview.", @@ -1449,10 +1450,10 @@ It is fluent in English, French, Spanish, German, and Italian, with high grammat "functionCall": true, "id": "openai/gpt-4-turbo-preview", "maxTokens": 4096, - "tokens": 128000, "vision": false, }, { + "contextWindowTokens": 4095, "description": "GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks. Training data up to Sep 2021.", @@ -1461,10 +1462,10 @@ Training data up to Sep 2021.", "functionCall": false, "id": "openai/gpt-3.5-turbo-0613", "maxTokens": 4096, - "tokens": 4095, "vision": false, }, { + "contextWindowTokens": 32768, "description": "Nous Hermes 2 Mixtral 8x7B DPO is the new flagship Nous Research model trained over the [Mixtral 8x7B MoE LLM](/models/mistralai/mixtral-8x7b). The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/openai/gpt-4) generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks. @@ -1475,20 +1476,20 @@ The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/ope "functionCall": false, "id": "nousresearch/nous-hermes-2-mixtral-8x7b-dpo", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32000, "description": "This is Mistral AI's closed-source, medium-sided model. It's powered by a closed-source prototype and excels at reasoning, code, JSON, chat, and more. In benchmarks, it compares with many of the flagship models of other companies.", "displayName": "Mistral Medium", "enabled": false, "functionCall": false, "id": "mistralai/mistral-medium", "maxTokens": undefined, - "tokens": 32000, "vision": false, }, { + "contextWindowTokens": 32000, "description": "This model is currently powered by Mixtral-8X7B-v0.1, a sparse mixture of experts model with 12B active parameters. It has better reasoning, exhibits more capabilities, can produce and reason about code, and is multiligual, supporting English, French, German, Italian, and Spanish. #moe", "displayName": "Mistral Small", @@ -1496,30 +1497,30 @@ The model was trained on over 1,000,000 entries of primarily [GPT-4](/models/ope "functionCall": false, "id": "mistralai/mistral-small", "maxTokens": undefined, - "tokens": 32000, "vision": false, }, { + "contextWindowTokens": 32000, "description": "This model is currently powered by Mistral-7B-v0.2, and incorporates a "better" fine-tuning than [Mistral 7B](/models/mistralai/mistral-7b-instruct-v0.1), inspired by community work. It's best used for large batch processing tasks where cost is a significant factor but reasoning capabilities are not crucial.", "displayName": "Mistral Tiny", "enabled": false, "functionCall": false, "id": "mistralai/mistral-tiny", "maxTokens": undefined, - "tokens": 32000, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A 75/25 merge of [Chronos 13b v2](https://huggingface.co/elinas/chronos-13b-v2) and [Nous Hermes Llama2 13b](/models/nousresearch/nous-hermes-llama2-13b). This offers the imaginative writing style of Chronos while retaining coherency. Outputs are long and use exceptional prose. #merge", "displayName": "Chronos Hermes 13B v2", "enabled": false, "functionCall": false, "id": "austism/chronos-hermes-13b", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 4096, "description": "Nous Hermes 2 Yi 34B was trained on 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape. Nous-Hermes 2 on Yi 34B outperforms all Nous-Hermes & Open-Hermes models of the past, achieving new heights in all benchmarks for a Nous Research LLM as well as surpassing many popular finetunes.", @@ -1528,10 +1529,10 @@ Nous-Hermes 2 on Yi 34B outperforms all Nous-Hermes & Open-Hermes models of the "functionCall": false, "id": "nousresearch/nous-hermes-yi-34b", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 32768, "description": "A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. An improved version of [Mistral 7B Instruct](/modelsmistralai/mistral-7b-instruct-v0.1), with the following changes: @@ -1544,10 +1545,10 @@ An improved version of [Mistral 7B Instruct](/modelsmistralai/mistral-7b-instruc "functionCall": false, "id": "mistralai/mistral-7b-instruct-v0.2", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "This is a 16k context fine-tune of [Mixtral-8x7b](/models/mistralai/mixtral-8x7b). It excels in coding tasks due to extensive training with coding data and is known for its obedience, although it lacks DPO tuning. The model is uncensored and is stripped of alignment and bias. It requires an external alignment layer for ethical use. Users are cautioned to use this highly compliant model responsibly, as detailed in a blog post about uncensored models at [erichartford.com/uncensored-models](https://erichartford.com/uncensored-models). @@ -1558,10 +1559,10 @@ The model is uncensored and is stripped of alignment and bias. It requires an ex "functionCall": false, "id": "cognitivecomputations/dolphin-mixtral-8x7b", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 65536, "description": "Google's flagship multimodal model, supporting image and video in text or chat prompts for a text or code response. See the benchmarks and prompting guidelines from [Deepmind](https://deepmind.google/technologies/gemini/). @@ -1574,10 +1575,10 @@ Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.d "functionCall": false, "id": "google/gemini-pro-vision", "maxTokens": 8192, - "tokens": 65536, "vision": true, }, { + "contextWindowTokens": 131040, "description": "Google's flagship text generation model. Designed to handle natural language tasks, multiturn text and code chat, and code generation. See the benchmarks and prompting guidelines from [Deepmind](https://deepmind.google/technologies/gemini/). @@ -1588,10 +1589,10 @@ Usage of Gemini is subject to Google's [Gemini Terms of Use](https://ai.google.d "functionCall": false, "id": "google/gemini-pro", "maxTokens": 32768, - "tokens": 131040, "vision": false, }, { + "contextWindowTokens": 32768, "description": "A pretrained generative Sparse Mixture of Experts, by Mistral AI, for chat and instruction use. Incorporates 8 experts (feed-forward networks) for a total of 47 billion parameters. Instruct model fine-tuned by Mistral. #moe", @@ -1600,10 +1601,10 @@ Instruct model fine-tuned by Mistral. #moe", "functionCall": false, "id": "mistralai/mixtral-8x7b-instruct", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "A pretrained generative Sparse Mixture of Experts, by Mistral AI, for chat and instruction use. Incorporates 8 experts (feed-forward networks) for a total of 47 billion parameters. Instruct model fine-tuned by Mistral. #moe @@ -1614,10 +1615,10 @@ _These are higher-throughput endpoints for [Mixtral 8x7B Instruct](/models/mistr "functionCall": false, "id": "mistralai/mixtral-8x7b-instruct:nitro", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "A pretrained generative Sparse Mixture of Experts, by Mistral AI. Incorporates 8 experts (feed-forward networks) for a total of 47B parameters. Base model (not fine-tuned for instructions) - see [Mixtral 8x7B Instruct](/models/mistralai/mixtral-8x7b-instruct) for an instruct-tuned model. #moe", @@ -1626,10 +1627,10 @@ _These are higher-throughput endpoints for [Mixtral 8x7B Instruct](/models/mistr "functionCall": false, "id": "mistralai/mixtral-8x7b", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "This is the chat model variant of the [StripedHyena series](/models?q=stripedhyena) developed by Together in collaboration with Nous Research. StripedHyena uses a new architecture that competes with traditional Transformers, particularly in long-context data processing. It combines attention mechanisms with gated convolutions for improved speed, efficiency, and scaling. This model marks a significant advancement in AI architecture for sequence modeling tasks.", @@ -1638,10 +1639,10 @@ StripedHyena uses a new architecture that competes with traditional Transformers "functionCall": false, "id": "togethercomputer/stripedhyena-nous-7b", "maxTokens": undefined, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "From the creator of [MythoMax](/models/gryphe/mythomax-l2-13b), merges a suite of models to reduce word anticipation, ministrations, and other undesirable words in ChatGPT roleplaying data. It combines [Neural Chat 7B](/models/intel/neural-chat-7b), Airoboros 7b, [Toppy M 7B](/models/undi95/toppy-m-7b), [Zepher 7b beta](/models/huggingfaceh4/zephyr-7b-beta), [Nous Capybara 34B](/models/nousresearch/nous-capybara-34b), [OpenHeremes 2.5](/models/teknium/openhermes-2.5-mistral-7b), and many others. @@ -1654,10 +1655,10 @@ _These are free, rate-limited endpoints for [MythoMist 7B](/models/gryphe/mythom "functionCall": false, "id": "gryphe/mythomist-7b:free", "maxTokens": 4096, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 32768, "description": "From the creator of [MythoMax](/models/gryphe/mythomax-l2-13b), merges a suite of models to reduce word anticipation, ministrations, and other undesirable words in ChatGPT roleplaying data. It combines [Neural Chat 7B](/models/intel/neural-chat-7b), Airoboros 7b, [Toppy M 7B](/models/undi95/toppy-m-7b), [Zepher 7b beta](/models/huggingfaceh4/zephyr-7b-beta), [Nous Capybara 34B](/models/nousresearch/nous-capybara-34b), [OpenHeremes 2.5](/models/teknium/openhermes-2.5-mistral-7b), and many others. @@ -1668,10 +1669,10 @@ It combines [Neural Chat 7B](/models/intel/neural-chat-7b), Airoboros 7b, [Toppy "functionCall": false, "id": "gryphe/mythomist-7b", "maxTokens": 2048, - "tokens": 32768, "vision": false, }, { + "contextWindowTokens": 8192, "description": "OpenChat 7B is a library of open-source language models, fine-tuned with "C-RLFT (Conditioned Reinforcement Learning Fine-Tuning)" - a strategy inspired by offline reinforcement learning. It has been trained on mixed-quality data without preference labels. - For OpenChat fine-tuned on Mistral 7B, check out [OpenChat 7B](/models/openchat/openchat-7b). @@ -1685,10 +1686,10 @@ _These are free, rate-limited endpoints for [OpenChat 3.5 7B](/models/openchat/o "functionCall": false, "id": "openchat/openchat-7b:free", "maxTokens": 4096, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "OpenChat 7B is a library of open-source language models, fine-tuned with "C-RLFT (Conditioned Reinforcement Learning Fine-Tuning)" - a strategy inspired by offline reinforcement learning. It has been trained on mixed-quality data without preference labels. - For OpenChat fine-tuned on Mistral 7B, check out [OpenChat 7B](/models/openchat/openchat-7b). @@ -1700,10 +1701,10 @@ _These are free, rate-limited endpoints for [OpenChat 3.5 7B](/models/openchat/o "functionCall": false, "id": "openchat/openchat-7b", "maxTokens": undefined, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 8192, "description": "A collab between IkariDev and Undi. This merge is suitable for RP, ERP, and general knowledge. #merge #uncensored", @@ -1712,30 +1713,30 @@ _These are free, rate-limited endpoints for [OpenChat 3.5 7B](/models/openchat/o "functionCall": false, "id": "neversleep/noromaid-20b", "maxTokens": 2048, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 100000, "description": "Anthropic's model for low-latency, high throughput text generation. Supports hundreds of pages of text.", "displayName": "Anthropic: Claude Instant v1.1", "enabled": false, "functionCall": false, "id": "anthropic/claude-instant-1.1", "maxTokens": 2048, - "tokens": 100000, "vision": false, }, { + "contextWindowTokens": 200000, "description": "Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use.", "displayName": "Anthropic: Claude v2.1", "enabled": false, "functionCall": false, "id": "anthropic/claude-2.1", "maxTokens": 4096, - "tokens": 200000, "vision": false, }, { + "contextWindowTokens": 200000, "description": "Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use. _This is a faster endpoint, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the provider's side instead of OpenRouter's. For requests that pass moderation, it's identical to the [Standard](/models/anthropic/claude-2.1) variant._", @@ -1744,20 +1745,20 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": false, "id": "anthropic/claude-2.1:beta", "maxTokens": 4096, - "tokens": 200000, "vision": false, }, { + "contextWindowTokens": 200000, "description": "Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use.", "displayName": "Anthropic: Claude v2", "enabled": false, "functionCall": false, "id": "anthropic/claude-2", "maxTokens": 4096, - "tokens": 200000, "vision": false, }, { + "contextWindowTokens": 200000, "description": "Claude 2 delivers advancements in key capabilities for enterprises—including an industry-leading 200K token context window, significant reductions in rates of model hallucination, system prompts and a new beta feature: tool use. _This is a faster endpoint, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the provider's side instead of OpenRouter's. For requests that pass moderation, it's identical to the [Standard](/models/anthropic/claude-2) variant._", @@ -1766,10 +1767,10 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": false, "id": "anthropic/claude-2:beta", "maxTokens": 4096, - "tokens": 200000, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A continuation of [OpenHermes 2 model](/models/teknium/openhermes-2-mistral-7b), trained on additional code datasets. Potentially the most interesting finding from training on a good ratio (est. of around 7-14% of the total dataset) of code instruction was that it has boosted several non-code benchmarks, including TruthfulQA, AGIEval, and GPT4All suite. It did however reduce BigBench benchmark score, but the net gain overall is significant.", "displayName": "OpenHermes 2.5 Mistral 7B", @@ -1777,10 +1778,10 @@ Potentially the most interesting finding from training on a good ratio (est. of "functionCall": false, "id": "teknium/openhermes-2.5-mistral-7b", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 128000, "description": "Ability to understand images, in addition to all other [GPT-4 Turbo capabilties](/models/openai/gpt-4-turbo). Training data: up to Apr 2023. **Note:** heavily rate limited by OpenAI while in preview. @@ -1791,10 +1792,10 @@ Potentially the most interesting finding from training on a good ratio (est. of "functionCall": false, "id": "openai/gpt-4-vision-preview", "maxTokens": 4096, - "tokens": 128000, "vision": true, }, { + "contextWindowTokens": 4096, "description": "A Mythomax/MLewd_13B-style merge of selected 70B models. A multi-model merge of several LLaMA2 70B finetunes for roleplaying and creative work. The goal was to create a model that combines creativity with intelligence for an enhanced experience. @@ -1804,10 +1805,10 @@ A multi-model merge of several LLaMA2 70B finetunes for roleplaying and creative "functionCall": false, "id": "lizpreciatior/lzlv-70b-fp16-hf", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 6144, "description": "A large LLM created by combining two fine-tuned Llama 70B models into one 120B model. Combines Xwin and Euryale. Credits to @@ -1820,10 +1821,10 @@ Credits to "functionCall": false, "id": "alpindale/goliath-120b", "maxTokens": 400, - "tokens": 6144, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit. List of merged models: - NousResearch/Nous-Capybara-7B-V1.9 @@ -1840,10 +1841,10 @@ _These are free, rate-limited endpoints for [Toppy M 7B](/models/undi95/toppy-m- "functionCall": false, "id": "undi95/toppy-m-7b:free", "maxTokens": 2048, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit. List of merged models: - NousResearch/Nous-Capybara-7B-V1.9 @@ -1858,10 +1859,10 @@ List of merged models: "functionCall": false, "id": "undi95/toppy-m-7b", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A wild 7B parameter model that merges several models using the new task_arithmetic merge method from mergekit. List of merged models: - NousResearch/Nous-Capybara-7B-V1.9 @@ -1878,10 +1879,10 @@ _These are higher-throughput endpoints for [Toppy M 7B](/models/undi95/toppy-m-7 "functionCall": false, "id": "undi95/toppy-m-7b:nitro", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 200000, "description": "Depending on their size, subject, and complexity, your prompts will be sent to [Llama 3 70B Instruct](/models/meta-llama/llama-3-70b-instruct), [Claude 3.5 Sonnet (self-moderated)](/models/anthropic/claude-3.5-sonnet:beta) or [GPT-4o](/models/openai/gpt-4o). To see which model was used, visit [Activity](/activity). A major redesign of this router is coming soon. Stay tuned on [Discord](https://discord.gg/fVyRaUDgxW) for updates.", @@ -1890,10 +1891,10 @@ A major redesign of this router is coming soon. Stay tuned on [Discord](https:// "functionCall": false, "id": "openrouter/auto", "maxTokens": undefined, - "tokens": 200000, "vision": false, }, { + "contextWindowTokens": 128000, "description": "The latest GPT-4 Turbo model with vision capabilities. Vision requests can now use JSON mode and function calling. Training data: up to April 2023.", @@ -1902,40 +1903,40 @@ Training data: up to April 2023.", "functionCall": true, "id": "openai/gpt-4-1106-preview", "maxTokens": 4096, - "tokens": 128000, "vision": true, }, { + "contextWindowTokens": 16385, "description": "An older GPT-3.5 Turbo model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Sep 2021.", "displayName": "OpenAI: GPT-3.5 Turbo 16k (older v1106)", "enabled": false, "functionCall": true, "id": "openai/gpt-3.5-turbo-1106", "maxTokens": 4096, - "tokens": 16385, "vision": false, }, { + "contextWindowTokens": 131040, "description": "PaLM 2 fine-tuned for chatbot conversations that help with code-related questions.", "displayName": "Google: PaLM 2 Code Chat 32k", "enabled": false, "functionCall": false, "id": "google/palm-2-codechat-bison-32k", "maxTokens": 32768, - "tokens": 131040, "vision": false, }, { + "contextWindowTokens": 131040, "description": "PaLM 2 is a language model by Google with improved multilingual, reasoning and coding capabilities.", "displayName": "Google: PaLM 2 Chat 32k", "enabled": false, "functionCall": false, "id": "google/palm-2-chat-bison-32k", "maxTokens": 32768, - "tokens": 131040, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A Llama 2 70B fine-tune using synthetic data (the Airoboros dataset). Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin/airoboros-l2-70b-2.2.1), but might get updated in the future.", @@ -1944,90 +1945,90 @@ Currently based on [jondurbin/airoboros-l2-70b](https://huggingface.co/jondurbin "functionCall": false, "id": "jondurbin/airoboros-l2-70b", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 8192, "description": "Xwin-LM aims to develop and open-source alignment tech for LLMs. Our first release, built-upon on the [Llama2](/models/\${Model.Llama_2_13B_Chat}) base models, ranked TOP-1 on AlpacaEval. Notably, it's the first to surpass [GPT-4](/models/\${Model.GPT_4}) on this benchmark. The project will be continuously updated.", "displayName": "Xwin 70B", "enabled": false, "functionCall": false, "id": "xwin-lm/xwin-lm-70b", "maxTokens": 400, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length.", "displayName": "Mistral: Mistral 7B Instruct v0.1", "enabled": false, "functionCall": false, "id": "mistralai/mistral-7b-instruct-v0.1", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 4095, "description": "This model is a variant of GPT-3.5 Turbo tuned for instructional prompts and omitting chat-related optimizations. Training data: up to Sep 2021.", "displayName": "OpenAI: GPT-3.5 Turbo Instruct", "enabled": false, "functionCall": false, "id": "openai/gpt-3.5-turbo-instruct", "maxTokens": 4096, - "tokens": 4095, "vision": false, }, { + "contextWindowTokens": 8192, "description": "A blend of the new Pygmalion-13b and MythoMax. #merge", "displayName": "Pygmalion: Mythalion 13B", "enabled": false, "functionCall": false, "id": "pygmalionai/mythalion-13b", "maxTokens": 400, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 32767, "description": "GPT-4-32k is an extended version of GPT-4, with the same capabilities but quadrupled context length, allowing for processing up to 40 pages of text in a single pass. This is particularly beneficial for handling longer content like interacting with PDFs without an external vector database. Training data: up to Sep 2021.", "displayName": "OpenAI: GPT-4 32k (older v0314)", "enabled": false, "functionCall": false, "id": "openai/gpt-4-32k-0314", "maxTokens": 4096, - "tokens": 32767, "vision": false, }, { + "contextWindowTokens": 32767, "description": "GPT-4-32k is an extended version of GPT-4, with the same capabilities but quadrupled context length, allowing for processing up to 40 pages of text in a single pass. This is particularly beneficial for handling longer content like interacting with PDFs without an external vector database. Training data: up to Sep 2021.", "displayName": "OpenAI: GPT-4 32k", "enabled": false, "functionCall": false, "id": "openai/gpt-4-32k", "maxTokens": 4096, - "tokens": 32767, "vision": false, }, { + "contextWindowTokens": 16385, "description": "This model offers four times the context length of gpt-3.5-turbo, allowing it to support approximately 20 pages of text in a single request at a higher cost. Training data: up to Sep 2021.", "displayName": "OpenAI: GPT-3.5 Turbo 16k", "enabled": false, "functionCall": false, "id": "openai/gpt-3.5-turbo-16k", "maxTokens": 4096, - "tokens": 16385, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A state-of-the-art language model fine-tuned on over 300k instructions by Nous Research, with Teknium and Emozilla leading the fine tuning process.", "displayName": "Nous: Hermes 13B", "enabled": false, "functionCall": false, "id": "nousresearch/nous-hermes-llama2-13b", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 4096, "description": "Zephyr is a series of language models that are trained to act as helpful assistants. Zephyr-7B-β is the second model in the series, and is a fine-tuned version of [mistralai/Mistral-7B-v0.1](/models/mistralai/mistral-7b-instruct-v0.1) that was trained on a mix of publicly available, synthetic datasets using Direct Preference Optimization (DPO). _These are free, rate-limited endpoints for [Zephyr 7B](/models/huggingfaceh4/zephyr-7b-beta). Outputs may be cached. Read about rate limits [here](/docs/limits)._", @@ -2036,60 +2037,60 @@ _These are free, rate-limited endpoints for [Zephyr 7B](/models/huggingfaceh4/ze "functionCall": false, "id": "huggingfaceh4/zephyr-7b-beta:free", "maxTokens": 2048, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 8000, "description": "An attempt to recreate Claude-style verbosity, but don't expect the same level of coherence or memory. Meant for use in roleplay/narrative situations.", "displayName": "Mancer: Weaver (alpha)", "enabled": false, "functionCall": false, "id": "mancer/weaver", "maxTokens": 1000, - "tokens": 8000, "vision": false, }, { + "contextWindowTokens": 100000, "description": "Anthropic's model for low-latency, high throughput text generation. Supports hundreds of pages of text.", "displayName": "Anthropic: Claude Instant v1.0", "enabled": false, "functionCall": false, "id": "anthropic/claude-instant-1.0", "maxTokens": 4096, - "tokens": 100000, "vision": false, }, { + "contextWindowTokens": 100000, "description": "Anthropic's model for low-latency, high throughput text generation. Supports hundreds of pages of text.", "displayName": "Anthropic: Claude v1.2", "enabled": false, "functionCall": false, "id": "anthropic/claude-1.2", "maxTokens": 4096, - "tokens": 100000, "vision": false, }, { + "contextWindowTokens": 100000, "description": "Anthropic's model for low-latency, high throughput text generation. Supports hundreds of pages of text.", "displayName": "Anthropic: Claude v1", "enabled": false, "functionCall": false, "id": "anthropic/claude-1", "maxTokens": 4096, - "tokens": 100000, "vision": false, }, { + "contextWindowTokens": 100000, "description": "Anthropic's model for low-latency, high throughput text generation. Supports hundreds of pages of text.", "displayName": "Anthropic: Claude Instant v1", "enabled": false, "functionCall": false, "id": "anthropic/claude-instant-1", "maxTokens": 4096, - "tokens": 100000, "vision": false, }, { + "contextWindowTokens": 100000, "description": "Anthropic's model for low-latency, high throughput text generation. Supports hundreds of pages of text. _This is a faster endpoint, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the provider's side instead of OpenRouter's. For requests that pass moderation, it's identical to the [Standard](/models/anthropic/claude-instant-1) variant._", @@ -2098,20 +2099,20 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": false, "id": "anthropic/claude-instant-1:beta", "maxTokens": 4096, - "tokens": 100000, "vision": false, }, { + "contextWindowTokens": 100000, "description": "Anthropic's flagship model. Superior performance on tasks that require complex reasoning. Supports hundreds of pages of text.", "displayName": "Anthropic: Claude v2.0", "enabled": false, "functionCall": false, "id": "anthropic/claude-2.0", "maxTokens": 4096, - "tokens": 100000, "vision": false, }, { + "contextWindowTokens": 100000, "description": "Anthropic's flagship model. Superior performance on tasks that require complex reasoning. Supports hundreds of pages of text. _This is a faster endpoint, made available in collaboration with Anthropic, that is self-moderated: response moderation happens on the provider's side instead of OpenRouter's. For requests that pass moderation, it's identical to the [Standard](/models/anthropic/claude-2.0) variant._", @@ -2120,20 +2121,20 @@ _This is a faster endpoint, made available in collaboration with Anthropic, that "functionCall": false, "id": "anthropic/claude-2.0:beta", "maxTokens": 4096, - "tokens": 100000, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A recreation trial of the original MythoMax-L2-B13 but with updated models. #merge", "displayName": "ReMM SLERP 13B", "enabled": false, "functionCall": false, "id": "undi95/remm-slerp-l2-13b", "maxTokens": 400, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 6144, "description": "A recreation trial of the original MythoMax-L2-B13 but with updated models. #merge _These are extended-context endpoints for [ReMM SLERP 13B](/models/undi95/remm-slerp-l2-13b). They may have higher prices._", @@ -2142,40 +2143,40 @@ _These are extended-context endpoints for [ReMM SLERP 13B](/models/undi95/remm-s "functionCall": false, "id": "undi95/remm-slerp-l2-13b:extended", "maxTokens": 400, - "tokens": 6144, "vision": false, }, { + "contextWindowTokens": 28672, "description": "PaLM 2 fine-tuned for chatbot conversations that help with code-related questions.", "displayName": "Google: PaLM 2 Code Chat", "enabled": false, "functionCall": false, "id": "google/palm-2-codechat-bison", "maxTokens": 4096, - "tokens": 28672, "vision": false, }, { + "contextWindowTokens": 36864, "description": "PaLM 2 is a language model by Google with improved multilingual, reasoning and coding capabilities.", "displayName": "Google: PaLM 2 Chat", "enabled": false, "functionCall": false, "id": "google/palm-2-chat-bison", "maxTokens": 4096, - "tokens": 36864, "vision": false, }, { + "contextWindowTokens": 4096, "description": "One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge", "displayName": "MythoMax 13B", "enabled": false, "functionCall": false, "id": "gryphe/mythomax-l2-13b", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 4096, "description": "One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge _These are higher-throughput endpoints for [MythoMax 13B](/models/gryphe/mythomax-l2-13b). They may have higher prices._", @@ -2184,10 +2185,10 @@ _These are higher-throughput endpoints for [MythoMax 13B](/models/gryphe/mythoma "functionCall": false, "id": "gryphe/mythomax-l2-13b:nitro", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 8192, "description": "One of the highest performing and most popular fine-tunes of Llama 2 13B, with rich descriptions and roleplay. #merge _These are extended-context endpoints for [MythoMax 13B](/models/gryphe/mythomax-l2-13b). They may have higher prices._", @@ -2196,40 +2197,40 @@ _These are extended-context endpoints for [MythoMax 13B](/models/gryphe/mythomax "functionCall": false, "id": "gryphe/mythomax-l2-13b:extended", "maxTokens": 400, - "tokens": 8192, "vision": false, }, { + "contextWindowTokens": 4096, "description": "A 13 billion parameter language model from Meta, fine tuned for chat completions", "displayName": "Meta: Llama v2 13B Chat", "enabled": false, "functionCall": false, "id": "meta-llama/llama-2-13b-chat", "maxTokens": undefined, - "tokens": 4096, "vision": false, }, { + "contextWindowTokens": 8191, "description": "GPT-4-0314 is the first version of GPT-4 released, with a context length of 8,192 tokens, and was supported until June 14. Training data: up to Sep 2021.", "displayName": "OpenAI: GPT-4 (older v0314)", "enabled": false, "functionCall": false, "id": "openai/gpt-4-0314", "maxTokens": 4096, - "tokens": 8191, "vision": false, }, { + "contextWindowTokens": 8191, "description": "OpenAI's flagship model, GPT-4 is a large-scale multimodal language model capable of solving difficult problems with greater accuracy than previous models due to its broader general knowledge and advanced reasoning capabilities. Training data: up to Sep 2021.", "displayName": "OpenAI: GPT-4", "enabled": false, "functionCall": false, "id": "openai/gpt-4", "maxTokens": 4096, - "tokens": 8191, "vision": true, }, { + "contextWindowTokens": 4095, "description": "GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks. Training data up to Sep 2021.", @@ -2238,10 +2239,10 @@ Training data up to Sep 2021.", "functionCall": false, "id": "openai/gpt-3.5-turbo-0301", "maxTokens": 4096, - "tokens": 4095, "vision": false, }, { + "contextWindowTokens": 16385, "description": "The latest GPT-3.5 Turbo model with improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Training data: up to Sep 2021. This version has a higher accuracy at responding in requested formats and a fix for a bug which caused a text encoding issue for non-English language function calls.", @@ -2250,10 +2251,10 @@ This version has a higher accuracy at responding in requested formats and a fix "functionCall": true, "id": "openai/gpt-3.5-turbo-0125", "maxTokens": 4096, - "tokens": 16385, "vision": false, }, { + "contextWindowTokens": 16385, "description": "GPT-3.5 Turbo is OpenAI's fastest model. It can understand and generate natural language or code, and is optimized for chat and traditional completion tasks. Training data up to Sep 2021.", @@ -2262,7 +2263,6 @@ Training data up to Sep 2021.", "functionCall": false, "id": "openai/gpt-3.5-turbo", "maxTokens": 4096, - "tokens": 16385, "vision": false, }, ] diff --git a/src/libs/agent-runtime/openrouter/index.ts b/src/libs/agent-runtime/openrouter/index.ts index 109e17bfbbee..02d935d45d3b 100644 --- a/src/libs/agent-runtime/openrouter/index.ts +++ b/src/libs/agent-runtime/openrouter/index.ts @@ -20,6 +20,7 @@ export const LobeOpenRouterAI = LobeOpenAICompatibleFactory({ const model = m as unknown as OpenRouterModelCard; return { + contextWindowTokens: model.context_length, description: model.description, displayName: model.name, enabled: LOBE_DEFAULT_MODEL_LIST.find((m) => model.id.endsWith(m.id))?.enabled || false, @@ -30,7 +31,6 @@ export const LobeOpenRouterAI = LobeOpenAICompatibleFactory({ typeof model.top_provider.max_completion_tokens === 'number' ? model.top_provider.max_completion_tokens : undefined, - tokens: model.context_length, vision: model.description.includes('vision') || model.description.includes('multimodal') || diff --git a/src/libs/agent-runtime/sensenova/index.ts b/src/libs/agent-runtime/sensenova/index.ts index e95a22f05d5f..f551847a693c 100644 --- a/src/libs/agent-runtime/sensenova/index.ts +++ b/src/libs/agent-runtime/sensenova/index.ts @@ -9,10 +9,16 @@ export const LobeSenseNovaAI = LobeOpenAICompatibleFactory({ return { ...rest, - frequency_penalty: (frequency_penalty !== undefined && frequency_penalty > 0 && frequency_penalty <= 2) ? frequency_penalty : undefined, + frequency_penalty: + frequency_penalty !== undefined && frequency_penalty > 0 && frequency_penalty <= 2 + ? frequency_penalty + : undefined, stream: true, - temperature: (temperature !== undefined && temperature > 0 && temperature <= 2) ? temperature : undefined, - top_p: (top_p !== undefined && top_p > 0 && top_p < 1) ? top_p : undefined, + temperature: + temperature !== undefined && temperature > 0 && temperature <= 2 + ? temperature + : undefined, + top_p: top_p !== undefined && top_p > 0 && top_p < 1 ? top_p : undefined, } as any; }, }, diff --git a/src/libs/agent-runtime/taichu/index.ts b/src/libs/agent-runtime/taichu/index.ts index 14c2c2293734..23beae0d1f4e 100644 --- a/src/libs/agent-runtime/taichu/index.ts +++ b/src/libs/agent-runtime/taichu/index.ts @@ -9,16 +9,10 @@ export const LobeTaichuAI = LobeOpenAICompatibleFactory({ handlePayload: (payload: ChatStreamPayload) => { const { temperature, top_p, ...rest } = payload; - return { - ...rest, - temperature: - temperature !== undefined - ? Math.max(temperature / 2, 0.01) - : undefined, - top_p: - top_p !== undefined - ? Math.min(9.9, Math.max(top_p / 2, 0.1)) - : undefined + return { + ...rest, + temperature: temperature !== undefined ? Math.max(temperature / 2, 0.01) : undefined, + top_p: top_p !== undefined ? Math.min(9.9, Math.max(top_p / 2, 0.1)) : undefined, } as OpenAI.ChatCompletionCreateParamsStreaming; }, }, diff --git a/src/libs/agent-runtime/utils/streams/minimax.test.ts b/src/libs/agent-runtime/utils/streams/minimax.test.ts index 307aa126745b..f18cd6cd479d 100644 --- a/src/libs/agent-runtime/utils/streams/minimax.test.ts +++ b/src/libs/agent-runtime/utils/streams/minimax.test.ts @@ -1,5 +1,8 @@ -import { describe, it, expect } from 'vitest'; -import { processDoubleData } from './minimax'; // 假设文件名为 minimax.ts +import { describe, expect, it } from 'vitest'; + +import { processDoubleData } from './minimax'; + +// 假设文件名为 minimax.ts describe('processDoubleData', () => { it('should remove the second "data: {"id": and everything after it when matchCount is 2', () => { diff --git a/src/libs/agent-runtime/utils/streams/minimax.ts b/src/libs/agent-runtime/utils/streams/minimax.ts index e71ce6c5c3cc..ebd594e86f67 100644 --- a/src/libs/agent-runtime/utils/streams/minimax.ts +++ b/src/libs/agent-runtime/utils/streams/minimax.ts @@ -9,7 +9,10 @@ export const processDoubleData = (chunkValue: string): string => { const matchCount = (chunkValue.match(dataPattern) || []).length; let modifiedChunkValue = chunkValue; if (matchCount === 2) { - const secondDataIdIndex = chunkValue.indexOf('data: {"id":', chunkValue.indexOf('data: {"id":') + 1); + const secondDataIdIndex = chunkValue.indexOf( + 'data: {"id":', + chunkValue.indexOf('data: {"id":') + 1, + ); if (secondDataIdIndex !== -1) { modifiedChunkValue = chunkValue.slice(0, secondDataIdIndex).trim(); } diff --git a/src/libs/agent-runtime/zhipu/index.ts b/src/libs/agent-runtime/zhipu/index.ts index bea46fd33129..b49060e0166f 100644 --- a/src/libs/agent-runtime/zhipu/index.ts +++ b/src/libs/agent-runtime/zhipu/index.ts @@ -11,19 +11,18 @@ export const LobeZhipuAI = LobeOpenAICompatibleFactory({ ...payload, model, stream: true, - ...(model === "glm-4-alltools" ? { - temperature: temperature !== undefined - ? Math.max(0.01, Math.min(0.99, temperature / 2)) - : undefined, - top_p: top_p !== undefined - ? Math.max(0.01, Math.min(0.99, top_p)) - : undefined, - } : { - temperature: temperature !== undefined - ? temperature / 2 - : undefined, - top_p, - }), + ...(model === 'glm-4-alltools' + ? { + temperature: + temperature !== undefined + ? Math.max(0.01, Math.min(0.99, temperature / 2)) + : undefined, + top_p: top_p !== undefined ? Math.max(0.01, Math.min(0.99, top_p)) : undefined, + } + : { + temperature: temperature !== undefined ? temperature / 2 : undefined, + top_p, + }), }) as OpenAI.ChatCompletionCreateParamsStreaming, }, debug: { diff --git a/src/libs/langchain/loaders/index.ts b/src/libs/langchain/loaders/index.ts index e6626a8d5357..835392f32105 100644 --- a/src/libs/langchain/loaders/index.ts +++ b/src/libs/langchain/loaders/index.ts @@ -7,13 +7,13 @@ import { LANGCHAIN_SUPPORT_TEXT_LIST } from '@/libs/langchain/file'; import { LangChainLoaderType } from '@/libs/langchain/types'; import { CodeLoader } from './code'; +import { CsVLoader } from './csv'; import { DocxLoader } from './docx'; import { LatexLoader } from './latex'; import { MarkdownLoader } from './markdown'; import { PdfLoader } from './pdf'; import { PPTXLoader } from './pptx'; import { TextLoader } from './txt'; -import { CsVLoader} from './csv'; class LangChainError extends Error { constructor(message: string) { @@ -96,7 +96,7 @@ export class ChunkingLoader { return 'markdown'; } - if (filename.endsWith('csv') ) { + if (filename.endsWith('csv')) { return 'csv'; } diff --git a/src/libs/langchain/types.ts b/src/libs/langchain/types.ts index b403b29da6c9..0bcc746bcd9c 100644 --- a/src/libs/langchain/types.ts +++ b/src/libs/langchain/types.ts @@ -1 +1,9 @@ -export type LangChainLoaderType = 'code' | 'ppt' | 'pdf' | 'markdown' | 'doc' | 'text' | 'latex' | 'csv'; +export type LangChainLoaderType = + | 'code' + | 'ppt' + | 'pdf' + | 'markdown' + | 'doc' + | 'text' + | 'latex' + | 'csv'; diff --git a/src/locales/default/modelProvider.ts b/src/locales/default/modelProvider.ts index 523782471916..76e0b7ef459f 100644 --- a/src/locales/default/modelProvider.ts +++ b/src/locales/default/modelProvider.ts @@ -62,7 +62,7 @@ export default { desc: '填入 Cloudflare 账户 ID 或 自定义 API 地址', placeholder: 'Cloudflare Account ID / custom API URL', title: 'Cloudflare 账户 ID / API 地址', - } + }, }, github: { personalAccessToken: { diff --git a/src/migrations/FromV3ToV4/fixtures/ollama-output-v4.json b/src/migrations/FromV3ToV4/fixtures/ollama-output-v4.json index 5b9a9eb2afeb..05b742d78f7b 100644 --- a/src/migrations/FromV3ToV4/fixtures/ollama-output-v4.json +++ b/src/migrations/FromV3ToV4/fixtures/ollama-output-v4.json @@ -49,7 +49,7 @@ "displayName": "LLaVA 7B", "enabled": true, "id": "llava", - "tokens": 4096, + "contextWindowTokens": 4096, "vision": true } ] diff --git a/src/migrations/FromV6ToV7/types/v7.ts b/src/migrations/FromV6ToV7/types/v7.ts index d92567f103cd..c3fc69899014 100644 --- a/src/migrations/FromV6ToV7/types/v7.ts +++ b/src/migrations/FromV6ToV7/types/v7.ts @@ -1,5 +1,3 @@ - - interface OpenAICompatibleKeyVault { apiKey?: string; baseURL?: string; diff --git a/src/server/globalConfig/genServerLLMConfig.test.ts b/src/server/globalConfig/genServerLLMConfig.test.ts index a40ecceb12a6..141742cf959b 100644 --- a/src/server/globalConfig/genServerLLMConfig.test.ts +++ b/src/server/globalConfig/genServerLLMConfig.test.ts @@ -1,16 +1,16 @@ import { describe, expect, it, vi } from 'vitest'; -import { genServerLLMConfig } from './genServerLLMConfig'; - import { getLLMConfig } from '@/config/llm'; +import { genServerLLMConfig } from './genServerLLMConfig'; + // Mock ModelProvider enum vi.mock('@/libs/agent-runtime', () => ({ ModelProvider: { Azure: 'azure', Bedrock: 'bedrock', Ollama: 'ollama', - } + }, })); // Mock ProviderCards @@ -86,7 +86,7 @@ describe('genServerLLMConfig', () => { expect(config.ollama).toEqual({ enabled: true, - enabledModels: ['ollamaModels'], + enabledModels: ['ollamaModels'], fetchOnClient: true, serverModelCards: ['ollamaModel1', 'ollamaModel2'], }); diff --git a/src/server/globalConfig/genServerLLMConfig.ts b/src/server/globalConfig/genServerLLMConfig.ts index ec7b82432d2f..2f3b1ac2da4d 100644 --- a/src/server/globalConfig/genServerLLMConfig.ts +++ b/src/server/globalConfig/genServerLLMConfig.ts @@ -1,35 +1,40 @@ import { getLLMConfig } from '@/config/llm'; import * as ProviderCards from '@/config/modelProviders'; - import { ModelProvider } from '@/libs/agent-runtime'; - -import { extractEnabledModels, transformToChatModelCards } from '@/utils/parseModels'; - import { ModelProviderCard } from '@/types/llm'; +import { extractEnabledModels, transformToChatModelCards } from '@/utils/parseModels'; export const genServerLLMConfig = (specificConfig: Record) => { const llmConfig = getLLMConfig() as Record; - return Object.values(ModelProvider).reduce((config, provider) => { - const providerUpperCase = provider.toUpperCase(); - const providerCard = ProviderCards[`${provider}ProviderCard` as keyof typeof ProviderCards] as ModelProviderCard; - const providerConfig = specificConfig[provider as keyof typeof specificConfig] || {}; - const providerModelList = process.env[providerConfig.modelListKey ?? `${providerUpperCase}_MODEL_LIST`]; + return Object.values(ModelProvider).reduce( + (config, provider) => { + const providerUpperCase = provider.toUpperCase(); + const providerCard = ProviderCards[ + `${provider}ProviderCard` as keyof typeof ProviderCards + ] as ModelProviderCard; + const providerConfig = specificConfig[provider as keyof typeof specificConfig] || {}; + const providerModelList = + process.env[providerConfig.modelListKey ?? `${providerUpperCase}_MODEL_LIST`]; - config[provider] = { - enabled: llmConfig[providerConfig.enabledKey || `ENABLED_${providerUpperCase}`], - enabledModels: extractEnabledModels( - providerModelList, - providerConfig.withDeploymentName || false, - ), - serverModelCards: transformToChatModelCards({ - defaultChatModels: (providerCard as ModelProviderCard)?.chatModels || [], - modelString: providerModelList, - withDeploymentName: providerConfig.withDeploymentName || false, - }), - ...(providerConfig.fetchOnClient !== undefined && { fetchOnClient: providerConfig.fetchOnClient }), - }; + config[provider] = { + enabled: llmConfig[providerConfig.enabledKey || `ENABLED_${providerUpperCase}`], + enabledModels: extractEnabledModels( + providerModelList, + providerConfig.withDeploymentName || false, + ), + serverModelCards: transformToChatModelCards({ + defaultChatModels: (providerCard as ModelProviderCard)?.chatModels || [], + modelString: providerModelList, + withDeploymentName: providerConfig.withDeploymentName || false, + }), + ...(providerConfig.fetchOnClient !== undefined && { + fetchOnClient: providerConfig.fetchOnClient, + }), + }; - return config; - }, {} as Record); + return config; + }, + {} as Record, + ); }; diff --git a/src/server/globalConfig/index.ts b/src/server/globalConfig/index.ts index 089157fa548c..bd46080f18b3 100644 --- a/src/server/globalConfig/index.ts +++ b/src/server/globalConfig/index.ts @@ -6,10 +6,9 @@ import { enableNextAuth } from '@/const/auth'; import { parseSystemAgent } from '@/server/globalConfig/parseSystemAgent'; import { GlobalServerConfig } from '@/types/serverConfig'; +import { genServerLLMConfig } from './genServerLLMConfig'; import { parseAgentConfig } from './parseDefaultAgent'; -import { genServerLLMConfig } from './genServerLLMConfig' - export const getServerGlobalConfig = () => { const { ACCESS_CODES, DEFAULT_AGENT_CONFIG } = getAppConfig(); diff --git a/src/server/routers/edge/config/__snapshots__/index.test.ts.snap b/src/server/routers/edge/config/__snapshots__/index.test.ts.snap index dd9f562a93d9..dbebf0066bf3 100644 --- a/src/server/routers/edge/config/__snapshots__/index.test.ts.snap +++ b/src/server/routers/edge/config/__snapshots__/index.test.ts.snap @@ -20,6 +20,7 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST "id": "claude-2", }, { + "contextWindowTokens": 128000, "description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。", "displayName": "gpt-4-32k", "enabled": true, @@ -29,7 +30,6 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST "input": 10, "output": 30, }, - "tokens": 128000, }, ], } @@ -38,6 +38,7 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST > should work correct with gpt-4 1`] = ` [ { + "contextWindowTokens": 16385, "description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125", "displayName": "GPT-3.5 Turbo 1106", "enabled": true, @@ -47,9 +48,9 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST "input": 1, "output": 2, }, - "tokens": 16385, }, { + "contextWindowTokens": 16385, "description": "GPT 3.5 Turbo,适用于各种文本生成和理解任务,Currently points to gpt-3.5-turbo-0125", "displayName": "GPT-3.5 Turbo", "enabled": true, @@ -59,9 +60,9 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST "input": 0.5, "output": 1.5, }, - "tokens": 16385, }, { + "contextWindowTokens": 8192, "description": "GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。", "displayName": "GPT-4", "enabled": true, @@ -71,9 +72,9 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST "input": 30, "output": 60, }, - "tokens": 8192, }, { + "contextWindowTokens": 32768, "description": "GPT-4 提供了一个更大的上下文窗口,能够处理更长的文本输入,适用于需要广泛信息整合和数据分析的场景。", "displayName": "GPT-4 32K", "enabled": true, @@ -83,9 +84,9 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST "input": 60, "output": 120, }, - "tokens": 32768, }, { + "contextWindowTokens": 128000, "description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。", "displayName": "GPT-4 Turbo Preview 1106", "enabled": true, @@ -95,7 +96,6 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST "input": 10, "output": 30, }, - "tokens": 128000, }, { "displayName": "gpt-4-vision", @@ -107,6 +107,7 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST > show the hidden model 1`] = ` { + "contextWindowTokens": 128000, "description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。", "displayName": "GPT-4 Turbo Preview 1106", "enabled": true, @@ -116,7 +117,6 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENAI_MODEL_LIST "input": 10, "output": 30, }, - "tokens": 128000, } `; @@ -129,18 +129,18 @@ exports[`configRouter > getGlobalConfig > Model Provider env > OPENROUTER_MODEL_ ], "serverModelCards": [ { + "contextWindowTokens": 32768, "description": "LLaMA 3.1 提供多语言支持,是业界领先的生成模型之一。", "displayName": "Llama 3.1 8B (Free)", "enabled": true, "id": "meta-llama/llama-3.1-8b-instruct:free", - "tokens": 32768, }, { + "contextWindowTokens": 8192, "description": "Gemma 2 是Google轻量化的开源文本模型系列。", "displayName": "Gemma 2 9B (Free)", "enabled": true, "id": "google/gemma-2-9b-it:free", - "tokens": 8192, }, ], } diff --git a/src/server/routers/lambda/_template.ts b/src/server/routers/lambda/_template.ts index 9a3b96b7b2bf..38d509825ef4 100644 --- a/src/server/routers/lambda/_template.ts +++ b/src/server/routers/lambda/_template.ts @@ -1,8 +1,8 @@ import { z } from 'zod'; +import { insertSessionGroupSchema } from '@/database/schemas'; import { serverDB } from '@/database/server'; import { SessionGroupModel } from '@/database/server/models/sessionGroup'; -import { insertSessionGroupSchema } from '@/database/schemas'; import { authedProcedure, router } from '@/libs/trpc'; import { SessionGroupItem } from '@/types/session'; diff --git a/src/server/routers/lambda/knowledgeBase.ts b/src/server/routers/lambda/knowledgeBase.ts index ac0e8456689b..27904c73e039 100644 --- a/src/server/routers/lambda/knowledgeBase.ts +++ b/src/server/routers/lambda/knowledgeBase.ts @@ -1,8 +1,8 @@ import { z } from 'zod'; +import { insertKnowledgeBasesSchema } from '@/database/schemas'; import { serverDB } from '@/database/server'; import { KnowledgeBaseModel } from '@/database/server/models/knowledgeBase'; -import { insertKnowledgeBasesSchema } from '@/database/schemas'; import { authedProcedure, router } from '@/libs/trpc'; import { KnowledgeBaseItem } from '@/types/knowledgeBase'; diff --git a/src/server/routers/lambda/session.ts b/src/server/routers/lambda/session.ts index 65d94538e66e..d2419bd1e69c 100644 --- a/src/server/routers/lambda/session.ts +++ b/src/server/routers/lambda/session.ts @@ -1,9 +1,9 @@ import { z } from 'zod'; +import { insertAgentSchema, insertSessionSchema } from '@/database/schemas'; import { serverDB } from '@/database/server'; import { SessionModel } from '@/database/server/models/session'; import { SessionGroupModel } from '@/database/server/models/sessionGroup'; -import { insertAgentSchema, insertSessionSchema } from '@/database/schemas'; import { authedProcedure, publicProcedure, router } from '@/libs/trpc'; import { AgentChatConfigSchema } from '@/types/agent'; import { LobeMetaDataSchema } from '@/types/meta'; diff --git a/src/server/routers/lambda/sessionGroup.ts b/src/server/routers/lambda/sessionGroup.ts index 9a3b96b7b2bf..38d509825ef4 100644 --- a/src/server/routers/lambda/sessionGroup.ts +++ b/src/server/routers/lambda/sessionGroup.ts @@ -1,8 +1,8 @@ import { z } from 'zod'; +import { insertSessionGroupSchema } from '@/database/schemas'; import { serverDB } from '@/database/server'; import { SessionGroupModel } from '@/database/server/models/sessionGroup'; -import { insertSessionGroupSchema } from '@/database/schemas'; import { authedProcedure, router } from '@/libs/trpc'; import { SessionGroupItem } from '@/types/session'; diff --git a/src/server/routers/lambda/thread.ts b/src/server/routers/lambda/thread.ts index 05cb3eec0846..33903c4b9666 100644 --- a/src/server/routers/lambda/thread.ts +++ b/src/server/routers/lambda/thread.ts @@ -1,9 +1,9 @@ import { z } from 'zod'; +import { insertThreadSchema } from '@/database/schemas'; import { serverDB } from '@/database/server'; import { MessageModel } from '@/database/server/models/message'; import { ThreadModel } from '@/database/server/models/thread'; -import { insertThreadSchema } from '@/database/schemas'; import { authedProcedure, router } from '@/libs/trpc'; import { ThreadItem, createThreadSchema } from '@/types/topic/thread'; diff --git a/src/server/services/changelog/index.ts b/src/server/services/changelog/index.ts index 45e1e2014cad..dd5a464afc37 100644 --- a/src/server/services/changelog/index.ts +++ b/src/server/services/changelog/index.ts @@ -54,7 +54,7 @@ export class ChangelogService { return this.mergeChangelogs(data.cloud, data.community).slice(0, 5); } catch (e) { console.error('Error getting changelog lists:', e); - return false as any; + return []; } } diff --git a/src/server/services/nextAuthUser/index.ts b/src/server/services/nextAuthUser/index.ts index 85f972aab576..13833929c93e 100644 --- a/src/server/services/nextAuthUser/index.ts +++ b/src/server/services/nextAuthUser/index.ts @@ -1,8 +1,8 @@ import { NextResponse } from 'next/server'; +import { UserItem } from '@/database/schemas'; import { serverDB } from '@/database/server'; import { UserModel } from '@/database/server/models/user'; -import { UserItem } from '@/database/schemas'; import { pino } from '@/libs/logger'; import { LobeNextAuthDbAdapter } from '@/libs/next-auth/adapter'; diff --git a/src/store/user/slices/modelList/__snapshots__/action.test.ts.snap b/src/store/user/slices/modelList/__snapshots__/action.test.ts.snap index d1b25b895c83..7ca5b82eced8 100644 --- a/src/store/user/slices/modelList/__snapshots__/action.test.ts.snap +++ b/src/store/user/slices/modelList/__snapshots__/action.test.ts.snap @@ -2,11 +2,11 @@ exports[`LLMSettingsSliceAction > refreshModelProviderList > visible 1`] = ` { + "contextWindowTokens": 4096, "description": "LLaVA 是结合视觉编码器和 Vicuna 的多模态模型,用于强大的视觉和语言理解。", "displayName": "LLaVA 7B", "enabled": true, "id": "llava", - "tokens": 4096, "vision": true, } `; diff --git a/src/store/user/slices/modelList/action.test.ts b/src/store/user/slices/modelList/action.test.ts index cb110f6a01bf..011254240f1f 100644 --- a/src/store/user/slices/modelList/action.test.ts +++ b/src/store/user/slices/modelList/action.test.ts @@ -92,14 +92,14 @@ describe('LLMSettingsSliceAction', () => { displayName: 'ChatGPT-4', functionCall: true, id: 'gpt-4-0125-preview', - tokens: 128000, + contextWindowTokens: 128000, enabled: true, }, { displayName: 'ChatGPT-4 Vision', functionCall: true, id: 'gpt-4-turbo-2024-04-09', - tokens: 128000, + contextWindowTokens: 128000, vision: true, enabled: true, }, @@ -121,14 +121,14 @@ describe('LLMSettingsSliceAction', () => { enabled: true, functionCall: true, id: 'gpt-4-0125-preview', - tokens: 128000, + contextWindowTokens: 128000, }, { displayName: 'ChatGPT-4 Vision', enabled: true, functionCall: true, id: 'gpt-4-turbo-2024-04-09', - tokens: 128000, + contextWindowTokens: 128000, vision: true, }, ]); diff --git a/src/store/user/slices/modelList/reducers/customModelCard.test.ts b/src/store/user/slices/modelList/reducers/customModelCard.test.ts index 046a6f1756fe..119fa057b70f 100644 --- a/src/store/user/slices/modelList/reducers/customModelCard.test.ts +++ b/src/store/user/slices/modelList/reducers/customModelCard.test.ts @@ -21,7 +21,7 @@ describe('customModelCardsReducer', () => { isCustom: true, legacy: false, maxOutput: 1000, - tokens: 2048, + contextWindowTokens: 2048, vision: false, }, { @@ -33,7 +33,7 @@ describe('customModelCardsReducer', () => { isCustom: true, legacy: true, maxOutput: 500, - tokens: 1024, + contextWindowTokens: 1024, vision: true, }, ]; @@ -49,7 +49,7 @@ describe('customModelCardsReducer', () => { isCustom: true, legacy: false, maxOutput: 2000, - tokens: 4096, + contextWindowTokens: 4096, vision: false, }; @@ -75,7 +75,7 @@ describe('customModelCardsReducer', () => { isCustom: true, legacy: false, maxOutput: 1000, - tokens: 2048, + contextWindowTokens: 2048, vision: false, }; @@ -147,7 +147,7 @@ describe('customModelCardsReducer', () => { isCustom: true, legacy: false, maxOutput: 1500, - tokens: 2048, + contextWindowTokens: 2048, vision: false, }; @@ -187,7 +187,7 @@ describe('customModelCardsReducer', () => { isCustom: true, legacy: false, maxOutput: 1500, - tokens: 2048, + contextWindowTokens: 2048, vision: false, }; diff --git a/src/store/user/slices/modelList/selectors/modelProvider.ts b/src/store/user/slices/modelList/selectors/modelProvider.ts index 283d372212e8..57f2629b427a 100644 --- a/src/store/user/slices/modelList/selectors/modelProvider.ts +++ b/src/store/user/slices/modelList/selectors/modelProvider.ts @@ -116,9 +116,10 @@ const isModelEnabledUpload = (id: string) => (s: UserStore) => isModelEnabledVision(id)(s) || isModelEnabledFiles(id)(s); const isModelHasMaxToken = (id: string) => (s: UserStore) => - typeof getModelCardById(id)(s)?.tokens !== 'undefined'; + typeof getModelCardById(id)(s)?.contextWindowTokens !== 'undefined'; -const modelMaxToken = (id: string) => (s: UserStore) => getModelCardById(id)(s)?.tokens || 0; +const modelMaxToken = (id: string) => (s: UserStore) => + getModelCardById(id)(s)?.contextWindowTokens || 0; export const modelProviderSelectors = { defaultModelProviderList, diff --git a/src/tools/dalle/Render/Item/index.tsx b/src/tools/dalle/Render/Item/index.tsx index 356a755a858f..95a05401abbd 100644 --- a/src/tools/dalle/Render/Item/index.tsx +++ b/src/tools/dalle/Render/Item/index.tsx @@ -56,7 +56,7 @@ const ImageItem = memo( ); if (imageId || previewUrl) - return ; + return ; return ( diff --git a/src/types/files/index.ts b/src/types/files/index.ts index 280866cea0fc..105449372bd9 100644 --- a/src/types/files/index.ts +++ b/src/types/files/index.ts @@ -7,7 +7,6 @@ export enum FilesTabs { Websites = 'websites', } - export interface FileItem { createdAt: Date; enabled?: boolean; diff --git a/src/types/llm.ts b/src/types/llm.ts index 95e86af7c9c9..36bd5d77d53d 100644 --- a/src/types/llm.ts +++ b/src/types/llm.ts @@ -3,6 +3,10 @@ import { ReactNode } from 'react'; export type ModelPriceCurrency = 'CNY' | 'USD'; export interface ChatModelCard { + /** + * the context window (or input + output tokens limit) + */ + contextWindowTokens?: number; /** * only used in azure */ @@ -12,7 +16,6 @@ export interface ChatModelCard { * the name show for end user */ displayName?: string; - /** * whether model is enabled by default */ @@ -53,10 +56,6 @@ export interface ChatModelCard { writeCacheInput?: number; }; releasedAt?: string; - /** - * the context window (or input + output tokens limit) - */ - tokens?: number; /** * whether model supports vision diff --git a/src/utils/__snapshots__/parseModels.test.ts.snap b/src/utils/__snapshots__/parseModels.test.ts.snap index fa698597dd9c..a96639a72be2 100644 --- a/src/utils/__snapshots__/parseModels.test.ts.snap +++ b/src/utils/__snapshots__/parseModels.test.ts.snap @@ -82,6 +82,7 @@ exports[`parseModelString > only add the model 1`] = ` exports[`transformToChatModelCards > should have file with builtin models like gpt-4-0125-preview 1`] = ` [ { + "contextWindowTokens": 128000, "description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。", "displayName": "ChatGPT-4", "enabled": true, @@ -92,9 +93,9 @@ exports[`transformToChatModelCards > should have file with builtin models like g "input": 10, "output": 30, }, - "tokens": 128000, }, { + "contextWindowTokens": 128000, "description": "最新的 GPT-4 Turbo 模型具备视觉功能。现在,视觉请求可以使用 JSON 模式和函数调用。 GPT-4 Turbo 是一个增强版本,为多模态任务提供成本效益高的支持。它在准确性和效率之间找到平衡,适合需要进行实时交互的应用程序场景。", "displayName": "ChatGPT-4 Vision", "enabled": true, @@ -105,7 +106,6 @@ exports[`transformToChatModelCards > should have file with builtin models like g "input": 10, "output": 30, }, - "tokens": 128000, "vision": true, }, ] diff --git a/src/utils/genUserLLMConfig.test.ts b/src/utils/genUserLLMConfig.test.ts index 1ee8d7b338ca..6d9e50c9371a 100644 --- a/src/utils/genUserLLMConfig.test.ts +++ b/src/utils/genUserLLMConfig.test.ts @@ -1,15 +1,15 @@ import { describe, expect, it, vi } from 'vitest'; -import { genUserLLMConfig } from './genUserLLMConfig'; - import { ModelProviderCard } from '@/types/llm'; +import { genUserLLMConfig } from './genUserLLMConfig'; + // Mock ModelProvider enum vi.mock('@/libs/agent-runtime', () => ({ ModelProvider: { Ollama: 'ollama', OpenAI: 'openai', - } + }, })); // Mock ProviderCards and filterEnabledModels @@ -27,7 +27,7 @@ describe('genUserLLMConfig', () => { it('should generate correct LLM config for Ollama and OpenAI', () => { const specificConfig = { ollama: { enabled: true, fetchOnClient: true }, - openai: { enabled: true } + openai: { enabled: true }, }; const config = genUserLLMConfig(specificConfig); diff --git a/src/utils/genUserLLMConfig.ts b/src/utils/genUserLLMConfig.ts index 685e6698525b..9031346e4353 100644 --- a/src/utils/genUserLLMConfig.ts +++ b/src/utils/genUserLLMConfig.ts @@ -1,20 +1,22 @@ import * as ProviderCards from '@/config/modelProviders'; - import { ModelProvider } from '@/libs/agent-runtime'; - import { ModelProviderCard } from '@/types/llm'; import { UserModelProviderConfig } from '@/types/user/settings'; export const genUserLLMConfig = (specificConfig: Record): UserModelProviderConfig => { return Object.keys(ModelProvider).reduce((config, providerKey) => { const provider = ModelProvider[providerKey as keyof typeof ModelProvider]; - const providerCard = ProviderCards[`${providerKey}ProviderCard` as keyof typeof ProviderCards] as ModelProviderCard; + const providerCard = ProviderCards[ + `${providerKey}ProviderCard` as keyof typeof ProviderCards + ] as ModelProviderCard; const providerConfig = specificConfig[provider as keyof typeof specificConfig] || {}; config[provider] = { enabled: providerConfig.enabled !== undefined ? providerConfig.enabled : false, enabledModels: providerCard ? ProviderCards.filterEnabledModels(providerCard) : [], - ...(providerConfig.fetchOnClient !== undefined && { fetchOnClient: providerConfig.fetchOnClient }), + ...(providerConfig.fetchOnClient !== undefined && { + fetchOnClient: providerConfig.fetchOnClient, + }), }; return config; diff --git a/src/utils/parseModels.test.ts b/src/utils/parseModels.test.ts index 75fb77e65d8b..7adfca0925d0 100644 --- a/src/utils/parseModels.test.ts +++ b/src/utils/parseModels.test.ts @@ -37,7 +37,7 @@ describe('parseModelString', () => { expect(result.add[0]).toEqual({ displayName: 'ChatGLM 6B', id: 'chatglm-6b', - tokens: 4096, + contextWindowTokens: 4096, }); }); @@ -48,7 +48,7 @@ describe('parseModelString', () => { displayName: '讯飞星火 v3.5', functionCall: true, id: 'spark-v3.5', - tokens: 8192, + contextWindowTokens: 8192, }); }); @@ -62,7 +62,7 @@ describe('parseModelString', () => { displayName: 'Gemini 1.5 Flash', vision: true, id: 'gemini-1.5-flash-latest', - tokens: 16000, + contextWindowTokens: 16000, }, { displayName: 'ChatGPT Plus', @@ -70,7 +70,7 @@ describe('parseModelString', () => { functionCall: true, files: true, id: 'gpt-4-all', - tokens: 128000, + contextWindowTokens: 128000, }, ]); }); @@ -85,14 +85,14 @@ describe('parseModelString', () => { files: true, functionCall: true, id: 'gpt-4-0125-preview', - tokens: 128000, + contextWindowTokens: 128000, }, { displayName: 'ChatGPT-4 Vision', files: true, functionCall: true, id: 'gpt-4-turbo-2024-04-09', - tokens: 128000, + contextWindowTokens: 128000, vision: true, }, ]); @@ -100,27 +100,27 @@ describe('parseModelString', () => { it('should handle empty extension capability value', () => { const result = parseModelString('model1<1024:>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024 }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024 }); }); it('should handle empty extension capability name', () => { const result = parseModelString('model1<1024::file>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024, files: true }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024, files: true }); }); it('should handle duplicate extension capabilities', () => { const result = parseModelString('model1<1024:vision:vision>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024, vision: true }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024, vision: true }); }); it('should handle case-sensitive extension capability names', () => { const result = parseModelString('model1<1024:VISION:FC:file>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024, files: true }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024, files: true }); }); it('should handle case-sensitive extension capability values', () => { const result = parseModelString('model1<1024:vision:Fc:File>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024, vision: true }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024, vision: true }); }); it('should handle empty angle brackets', () => { @@ -155,27 +155,27 @@ describe('parseModelString', () => { it('should handle digits followed by non-colon characters inside angle brackets', () => { const result = parseModelString('model1<1024abc>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024 }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024 }); }); it('should handle digits followed by multiple colons inside angle brackets', () => { const result = parseModelString('model1<1024::>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024 }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024 }); }); it('should handle digits followed by a colon and non-letter characters inside angle brackets', () => { const result = parseModelString('model1<1024:123>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024 }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024 }); }); it('should handle digits followed by a colon and spaces inside angle brackets', () => { const result = parseModelString('model1<1024: vision>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024 }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024 }); }); it('should handle digits followed by multiple colons and spaces inside angle brackets', () => { const result = parseModelString('model1<1024: : vision>'); - expect(result.add[0]).toEqual({ id: 'model1', tokens: 1024 }); + expect(result.add[0]).toEqual({ id: 'model1', contextWindowTokens: 1024 }); }); }); diff --git a/src/utils/parseModels.ts b/src/utils/parseModels.ts index d3aac53850ca..7e965902998b 100644 --- a/src/utils/parseModels.ts +++ b/src/utils/parseModels.ts @@ -56,7 +56,7 @@ export const parseModelString = (modelString: string = '', withDeploymentName = if (capabilities.length > 0) { const [maxTokenStr, ...capabilityList] = capabilities[0].replace('>', '').split(':'); - model.tokens = parseInt(maxTokenStr, 10) || undefined; + model.contextWindowTokens = parseInt(maxTokenStr, 10) || undefined; for (const capability of capabilityList) { switch (capability) { diff --git a/src/utils/server/jwt.ts b/src/utils/server/jwt.ts index 589b3e207a3f..cde47b68f58a 100644 --- a/src/utils/server/jwt.ts +++ b/src/utils/server/jwt.ts @@ -1,10 +1,6 @@ import { importJWK, jwtVerify } from 'jose'; -import { - JWTPayload, - JWT_SECRET_KEY, - NON_HTTP_PREFIX, -} from '@/const/auth'; +import { JWTPayload, JWT_SECRET_KEY, NON_HTTP_PREFIX } from '@/const/auth'; export const getJWTPayload = async (token: string): Promise => { //如果是 HTTP 协议发起的请求,直接解析 token @@ -29,4 +25,4 @@ export const getJWTPayload = async (token: string): Promise => { const { payload } = await jwtVerify(token, jwkSecretKey); return payload as JWTPayload; -}; \ No newline at end of file +};