-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add a llama 3.1 toy test with cross entropy test #50
Conversation
If we pick the maximum token each decode step we should get a somewhat consistent cross entropy loss. This uses a prebaked llama model. The `irpa` file should stay static but the `mlir` model should update as features are enabled.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good! The top level sharktank_models
path SGTM. We can shuffle things around later as more tests are added too.
A few mostly structural and style comments.
Signed-off-by: Rob Suderman <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Very nice! Thanks for the changes!
If we pick the maximum token each decode step we should get a somewhat consistent cross entropy loss. This uses a prebaked llama model. The
irpa
file should stay static but themlir
model should update as features are enabled.