New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
On m1 pro - "Distributed package doesn't have NCCL built in #37
Comments
Got the same error. I tried to set PYTORCH_ENABLE_MPS_FALLBACK=1, but no luck. I am running M3 macbook air. |
Attempting a few more things to see what might happen, hopefully, someone from the team can respond if this is just supposed to run on Nvida / windows |
+1 |
+1 on m2 |
+1 |
thats a negative, i actually run into the same issue when i try to run it on windows. Windows Subsystem for Linux (WSL) works fine though. one thing to check would be to see if you have the proper CPU only installation of torch. i dont think GPU support is there for mac. |
I identified the code that is forcing nccl in my issue, #132 One of the first things Llama.build() does is initialize torch distributed with a hard coded nccl initialization. https://pytorch.org/docs/stable/distributed.html |
+1 on M3 Pro |
Feel free to correct me if I'm wrong... but I'm pretty sure there is no official support for ARM Macs (M-series chips). |
Must be something torch package, related...
This is when trying to run the command
torchrun --nproc_per_node 1 example_chat_completion.py
--ckpt_dir Meta-Llama-3-8B-Instruct/
--tokenizer_path Meta-Llama-3-8B-Instruct/tokenizer.model
--max_seq_len 512 --max_batch_size 6
The text was updated successfully, but these errors were encountered: