-
Notifications
You must be signed in to change notification settings - Fork 3.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
triton.runtime.autotuner.OutOfResources #4688
Labels
pending
This problem is yet to be addressed
Comments
The reproduction command is not posted, we don't know what process you are doing. |
Hey @codemayq, Thanks for the help.
|
|
decrease the train batch size |
@hiyouga Thanks for the help. This didn't work either. But I figured out that the quantization creates the issue. It works when I'm not quantizing. Sounds like a bug, right? Failing with this:
This worked:
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Reminder
System Info
llamafactory-0.8.3.dev0, Ubuntu 22.04.3 LTS, py3.10, cuda11.8.0
Reproduction
Command:
Error
Expected behavior
Hello, I'm really not sure if this is a LLaMA Factory issue or the Cloud GPU provider. Does anyone knows what to do?
Others
No response
The text was updated successfully, but these errors were encountered: