-
Notifications
You must be signed in to change notification settings - Fork 563
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
torchtune dry-run feature request #2453
Comments
Great minds chai alike 😁 #2452 |
Hey @agunapal , thats a good request! We do not have bandwidth to look into at this moment, but if you want to propose an RFC (PR with high level ideas on how to implement it) we could review it. As a sanity check, you could clone the recipe and modify the training loop to inspect the model inputs + tokenizer.decode. Would that work for you? |
just saw your comment @init27 ! I will close your issue, since i have replied here already, so we can consolidate the conversation. I will check if someone from the community has interest in picking it up, since both of you are interested. |
Just so I can understand the request better: You guys want to do a full epoch on the dataloader, but without training, to see if there are dataset issues. And you want to be able to possibly print or store the input/output of that dataloader, to confirm that they look like they should. Is that it, or is there something else? |
Yes exactly, thanks for confirming! The idea is instead of the training loop crashing on us mid-loop, have a method to validate all message examples. For an example: Right now im using synthetic conversations, sometimes these have duplicate |
@felipemello1 Yes, for my use case, I am using a custom prompt template, a custom dataset and llama-guard (the tokenizer is slightly different). I want to make sure that the model is getting the current input and its not a case of garbage in, garbage out. Currently I am adding prints to get around this. It would be nice to have this utility to visual inspect the final prompt and the corresponding tokens. |
New feature request
I know
torchtune
does some validation checks to ensure the prompt is not malformed.But we have no way of knowing if SFT is happening with the right input/output format.
What would be great is to have a
torchtune dry-run
command, which does the followingThis way one can visually inspect and be sure that
torchtune
has been configured correctly for fine-tuning.The text was updated successfully, but these errors were encountered: