You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Collecting environment information...
WARNING 10-29 04:40:34 cuda.py:22] You are using a deprecated `pynvml` package. Please install `nvidia-ml-py` instead, and make sure to uninstall `pynvml`. When both of them are installed, `pynvml` will take precedence and cause errors. See https://pypi.org/project/pynvml for more information.
Traceback (most recent call last):
File "/workspace/collect_env.py", line 743, in <module>
main()
File "/workspace/collect_env.py", line 722, in main
output = get_pretty_env_info()
File "/workspace/collect_env.py", line 717, in get_pretty_env_info
return pretty_str(get_env_info())
File "/workspace/collect_env.py", line 549, in get_env_info
vllm_version = get_vllm_version()
File "/workspace/collect_env.py", line 270, in get_vllm_version
from vllm import __version__, __version_tuple__
ImportError: cannot import name '__version_tuple__' from 'vllm' (/workspace/code/vllm/vllm/__init__.py)
but I am using v0.6.0 vllm
How would you like to use vllm
I tried to return logits part of output and like
{"text":["this is new vllm with logits bla bla....."],"logits":[1.0,2.0]}
I found each model (e.g.. llama.py) call and return logits from the function compute_logits. and then LocalOrDistributedWorkerBase::execute_model has logits value output = self.model_runner.execute_model(...) > 0 in worker_base.py
I lost connection to return output to RequestOutput to print like above. I appreciate if anyone can help. thanks
Before submitting a new issue...
Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
The text was updated successfully, but these errors were encountered:
Your current environment
How would you like to use vllm
I tried to return logits part of output and like
I found each model (e.g.. llama.py) call and return logits from the function compute_logits. and then LocalOrDistributedWorkerBase::execute_model has logits value output = self.model_runner.execute_model(...) > 0 in worker_base.py
I lost connection to return output to RequestOutput to print like above. I appreciate if anyone can help. thanks
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: