Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Llmblog2 #16

Merged
merged 36 commits into from
Feb 26, 2024
Merged

Llmblog2 #16

merged 36 commits into from
Feb 26, 2024

Conversation

KevinMusgrave
Copy link
Contributor

No description provided.

@KevinMusgrave
Copy link
Contributor Author

KevinMusgrave commented Feb 23, 2024

The test_model.py script downloads the best checkpoint if exp_id is given, and this was something I was able to run on my laptop for TinyLlama. I also ran it for the pretrained Mistral model, but I don't want to run it for a finetuned Mistral model because I think the checkpoint will be huge.

Part of the problem is that the checkpoint contains the optimizer state, which I don't need to evaluate the model. Maybe there is some way to download only part of the checkpoint? I'm not sure.

Copy link
Contributor

@aciborowska aciborowska left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM!

@KevinMusgrave KevinMusgrave merged commit 209c06d into main Feb 26, 2024
1 check passed
@KevinMusgrave KevinMusgrave deleted the llmblog2-profiler branch February 26, 2024 23:17
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants