-
Notifications
You must be signed in to change notification settings - Fork 8.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Error when installing requirements #6
Comments
I believe this should be fixed now. Can you try again? |
No, it doesn't work WARNING: jax 0.4.25 does not provide the extra 'cuda12-pip' |
I think this is a problem in my system, I have Windows, there is no CUDA in jax for Windows as I understand it |
Hello @ibab, I'm getting same error while installing requirements in WSL-2 Kali. Looks like fix doesn't work or I'm doing some kind of mistake while installing requirements. Error message under bellow; Used command:
|
Hi @kesevone Are you installing via WSL or trying to install on windows? I use wsl and getting same error. |
I'm trying to install on windows, now I'll try on wsl |
If you successfully install on WSL can you tell me too |
pip install dm-haiku in requirements it's dm_haiku==0.0.12 with underscore ... |
I got the same error and try this It work on my case. |
I get an error on startup |
That just means you don't have the appropriate number of devices. The mesh_shape is the configuration for what is expected, in this case 8 devices to distribute the model over and run inference on. If you don't allocate exactly 8 gpus, it will not work, granted running inference with this model will require a minimum of 8 large GPUs anyway. |
@AndreSlavescu I have tried with 1. i did mesh_shape (1, 1) instead mesh_shape (1, 8) |
I also have the same problem. Have you fixed it now? |
I got the same error, but running only:
worked for me, so it's probably because of the dm_haiku problem as described above |
In theory that should work in terms of not breaking at that point, but you won't have enough memory to load onto a single GPU, so you will get a CUDA OOM errror. |
same issue |
yep, I have up.. 300GB dowloaded for nothing :D |
running these commands after the error, fix the installation issue
|
If you want to try it and don’t have access to a 8 gpu cluster, there are cloud compute solutions with AWS sagemaker EC2 instances, lambda labs, coreweave, and a few more where you might be able to get an 8xA100 80GB (640GB total) allocation. |
change requirements.txt |
How do you solve this? |
After changing
and then running
Did anyone solved it? |
I was able to run Grok-1 yesterday. As people have commented, what did the trick for us at CloudWalk (a Brazilian fintech) was to use our K8 cluster with at least 8xA100 GPUs (80 GB family). Grok-1 uses almost all the memory from the GPUs (so using only 1 or 2 GPUs will not give you enough memory). Another thing that solved our problems was running: Then, we just needed to run |
You can also pull this container to run grok: |
i have installed python 3.10 and venv. Trying to "pip install -r requirements.txt"
ERROR: Ignored the following versions that require a different python version: 1.6.2 Requires-Python >=3.7,<3.10; 1.6.3 Requires-Python >=3.7,<3.10; 1.7.0 Requires-Python >=3.7,<3.10; 1.7.1 Requires-Python >=3.7,<3.10
ERROR: Could not find a version that satisfies the requirement jaxlib==0.4.25+cuda12.cudnn89; extra == "cuda12_pip" (from jax[cuda12-pip]) (from versions: 0.4.13, 0.4.14, 0.4.16, 0.4.17, 0.4.18, 0.4.19, 0.4.20, 0.4.21, 0.4.22, 0.4.23, 0.4.24, 0.4.25)
ERROR: No matching distribution found for jaxlib==0.4.25+cuda12.cudnn89; extra == "cuda12_pip"
The text was updated successfully, but these errors were encountered: