This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Request] llama对话过程能否保活? #2240
Comments
Thank you for raising an issue. We will investigate into the matter and get back to you as soon as possible. |
🥰 Description of requirementsI deployed ollama with docker, and the machine is i-3600k+rtx4090
🧐 SolutionIs it possible to keep a model alive after starting a conversation? Or are there other options? 📝 Supplementary informationNo response |
这可能要问下 ollama 的社区了 |
You might have to ask the ollama community about this. |
看Ollama 部署的文档,我这边 PR 过对应的环境变量可以让模型保活时间变长 |
Looking at the Ollama deployment documentation, I PRed the corresponding environment variables to make the model stay alive longer. |
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
🥰 需求描述
我是docker部署的ollama,机器是i-3600k+rtx4090
在用lobechat使用llama模型对话时有两处需要优化:
1.启动对话要很久的时间,一直加载中,应该是执行ollama run llama很费时间
2.对话过程暂停一段时间后,再次对话又需要等待很久
🧐 解决方案
能否在启动对话后保活模型?或者有其他方案?
📝 补充信息
No response
The text was updated successfully, but these errors were encountered: