You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Now I want to train larger model, so how can I perform fine-tuning of LLM using distributed training on a single client?
I would like to use multiple GPUs or multiple servers on one client. Can I use something like 'accelerate launch', 'torch run', or 'deepspeed' in client ?
What is your question?
I have successfully completed the federated training of LLM following the link below.
https://flower.ai/docs/framework/tutorial-quickstart-huggingface.html
Now I want to train larger model, so how can I perform fine-tuning of LLM using distributed training on a single client?
I would like to use multiple GPUs or multiple servers on one client. Can I use something like 'accelerate launch', 'torch run', or 'deepspeed' in client ?
https://huggingface.co/docs/trl/example_overview#distributed-training
Thanks !
The text was updated successfully, but these errors were encountered: