-
Notifications
You must be signed in to change notification settings - Fork 10
Add githash to nm-vllm #299
Conversation
We might want to add this info to |
Good idea, thanks for the suggestion. I added the githash, however I found that I cannot run the script from the repo root directory, otherwise I hit following error:
I think this is due to the vllm/ folder in the repo. If however I move collect_env.py to a different location and run it from there, it ran fine after I installed the nm-vllm wheel:
Solved this issue by making installed vllm taking precedence over the local vllm/ directory. |
I file this PR as a reference for nm-vllm. Since we're planning to make this change to the upstream, I'll file another PR for the upstream so people can compare the code between the PRs. There are some slight difference between the two. |
This looks good to me but the pybind stuff has just been switched over to torch library so that will probably require a few changes/merges. |
@dhuangnm This is probably fine as the end-user typically would not have the vllm source or run it from that directory. Instead, their support process has you download just that file ( They're actually likely to run it from the source directory since this python script is from the repo and they may just check out the repo first then run it from there. Anyways, the hack allows it to work no matter if it's called from the repo dir or not. |
Yes I'll make the changes for the PR against the upstream. I'll post the PR shortly. |
Can I get an approval? Reran the failed 38 job and it passed now. |
Thanks Bill. It looks I need another approval? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
thanks
It looks there are two failures due to OOM across the python versions: FAILED tests/models_core/test_magic_wand.py::test_magic_wand[5-32-half-model_format_extrablocks0] - torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 896.00 MiB. GPU These are also failing in the nightly so seems not caused by this PR. |
Add git hash information to nm-vllm: ``` >>> import vllm >>> vllm.githash() '106796861914146372aba9386aeff9361edfb34d' ``` --------- Co-authored-by: dhuangnm <dhuang@MacBook-Pro-2.local>
Add git hash information to nm-vllm: