Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Pruned model is same size as original #29

Open
virentakia opened this issue Dec 5, 2023 · 2 comments
Open

Pruned model is same size as original #29

virentakia opened this issue Dec 5, 2023 · 2 comments

Comments

@virentakia
Copy link

virentakia commented Dec 5, 2023

Great work on the project, really excited to see the outcomes.

However, After running the script below, the pruned model (output) seems to be of the same size as the original one (which is 6.38G)

!python /content/wanda/main.py
--model openlm-research/open_llama_3b_v2
--prune_method wanda
--sparsity_ratio 0.5
--sparsity_type unstructured
--save_model out/pruned
--save out/open_llama_3b_v2/unstructured/wanda/

Is this correct, or am I missing something?!

@Eric-mingjie
Copy link
Collaborator

Yes, this is correct and it has been true for unstructured pruning. To my understanding, unstructured sparsity won't save the memory footprint on modern GPU devices.

@virentakia
Copy link
Author

Thanks, any pruning options for reducing memory footprint?!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants