-
Notifications
You must be signed in to change notification settings - Fork 14
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Benchmarks for Metis #10
Comments
Thank you for your interest in this project.
|
Thank you for your reply:
Thank you! |
Thank you! |
I got it! |
Hi, @mgong-kang |
The DataLoadBalancer is implemented at the follwing path: |
@goeunee326 |
Thank you for your help, but it seems that the output of the Metis strategy does not reflect the batch size when data parallelism in different GPUs. What is more, I found some same strategies have different costs. |
Thank you. |
You can execute the process by modifying a specific part of the Alpa benchmark code. To run the results from Metis in Alpa, parameter mapping is required:
Since the concept of |
If there are no further discussions or additional points to address, we will proceed to close this issue. Please feel free to reopen it at any time if further discussions or inquiries are needed. Do not hesitate to share any additional comments or questions. |
It is the excellent work "Metis: Fast Automatic Distributed Training on Heterogeneous GPUs", however, I have a couple of questions about the code:
Could you provide the benchmarks for metis?
Thank you!
The text was updated successfully, but these errors were encountered: