Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

evaluating has an extremely large value when quantize to 4bit. #105

Open
JiachuanDENG opened this issue Jun 7, 2023 · 1 comment
Open

Comments

@JiachuanDENG
Copy link

I followed the steps try to get 4bit version of llama7b by using command python -m llama.llama_quant decapoda-research/llama-7b-hf c4 --wbits 4 --groupsize 128 --save pyllama-7B4b.pt, the script works well, but at the evaluating stage, it got a very large number 251086.96875.

Screenshot2023_06_07_134554

And when I testing with the quantized .pt file, model returns un-readable results.

Screenshot 2023-06-07 at 13 50 05

Anyone has same problem?

@rapidAmbakar
Copy link

Yes same issue, exatly same

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants