-
-
Notifications
You must be signed in to change notification settings - Fork 10k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
⚡️ perf: improve performance on long context text #3754
Conversation
The latest updates on your projects. Learn more about Vercel for Git ↗︎
|
👍 @arvinxx Thank you for raising your pull request and contributing to our Community |
Codecov ReportAttention: Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## main #3754 +/- ##
==========================================
- Coverage 91.80% 91.75% -0.06%
==========================================
Files 452 453 +1
Lines 30020 30056 +36
Branches 2077 2913 +836
==========================================
+ Hits 27559 27577 +18
- Misses 2461 2479 +18
Flags with carried forward coverage won't be shown. Click here to find out more. ☔ View full report in Codecov by Sentry. |
❤️ Great PR @arvinxx ❤️ The growth of project is inseparable from user feedback and contribution, thanks for your contribution! If you are interesting with the lobehub developer community, please join our discord and then dm @arvinxx or @canisminor1990. They will invite you to our private developer channel. We are talking about the lobe-chat development or sharing ai newsletter around the world. |
🎉 This PR is included in version 1.15.10 🎉 The release is available on: Your semantic-release bot 📦🚀 |
💻 变更类型 | Change Type
🔀 变更说明 | Description of Change
📝 补充信息 | Additional Information
refs: #1029
tokenizer 性能优化点:
/webapi/tokenizer
做 token 计算。 ([RFC] 058 - 服务端接口架构梳理 #3755 )smoothing 动画,默认输出速率从 2-> 4,gpt-4o 级别的模型输出会明显提速
发现长上下文(300k)另外的几个问题:
requestAnimationFrame
看看是否能缓解;