-
-
Notifications
You must be signed in to change notification settings - Fork 5.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Installation]: no version of pip install vllm works - Failed to initialize NumPy: No Module named 'numpy' #11037
Comments
The actual error message is: (vllm) cephdon@ubuntu:~/sources/vllm$ VLLM_USE_PRECOMPILED=1 pip install --editable . × Getting requirements to build editable did not run successfully. |
(vllm) cephdon@ubuntu:~/sources/vllm$ pip list absl-py 2.1.0 |
I should note that this is on an NVIDIA AGX Orin 64gb machine, so aarch64. I have cuda 12.6, but building from source has this same issue that I cannot seem to get past. Installing numpy (any version) does not resolve the issue. I am using miniconda to manage virtual environments. pip list and conda list both show numpy. Note that the error message above is trying to use the CPU device, even if I have the cuda version of torch installed. Using the CPU version of torch does not resolve the issue. |
I have got the same problem, waiting for some possible solutions. 😢 ... UserWarning: Failed to initialize NumPy: No module named 'numpy' ... |
I have got the same problem, waiting for some possible solutions. 😢 |
same problem building from main (ea7bd68) |
@youkaichao @dtrifiro any idea about this? |
@cephdon a couple notes:
pip install -r requirements-build.txt -r requirements-cuda-arm64.txt
pip install --no-build-isolation --verbose . The reason for the error you're getting is that we currenty have There has been some effort in fixing the build process for isolated builds and different targets here, but it's been on hold for a while. Perhaps I'll try do do some more work on it next month. @chris-aeviator are you using |
@youkaichao @dtrifiro there is no requirements-cuda-arm64.txt on GitHub |
for building on arm64, please follow https://docs.vllm.ai/en/latest/serving/deploying_with_docker.html#building-for-arm64-aarch64
you are right, we deleted it. please follow the up-to-date doc. |
That is to build in a container, which is thoroughly useless for many use cases. We can't use it on our supercomputers, we can't create a pip-installable wheel for our users either. If there's an option to build from source, great. If there's an argument to give to |
I don't get it. Building in a container just makes things more isolated, you can of course use it without a container. We have clear doc on how to build from source on arm GPUs: https://docs.vllm.ai/en/latest/getting_started/installation.html#use-an-existing-pytorch-installation |
That's not how supercomputing centers work. They tend to compile everything to their platform, and try to extract every bit of performance they can. Running a linux kernel and a whole toolchain on top of a a linux kernel and a whole toolchain (repeat as needed) is not acceptable in many places. The overhead is not worth the convenience. Plus, the security nightmare that docker makes. I have read your documentation multiple times. I have managed to compile it, but never managed to make it work. It is got so bad that the market has been offering gh200 under cost because literally no one can run vllm on them properly because you can't bother to offer a precompiled wheel. Do you need links? |
we cannot provide wheels, because pytorch does not have official release wheels for gh200 (arm gpus). we have to wait for pytorch 2.6 release. technically, you can get wheels by copying from the docker container, but the wheel does not specify pytorch as dependency, and you need to install pytorch before installing vllm using that wheel.
can you elaborate? |
Your current environment
How you are installing vllm
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: