Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Questions about quantization #81

Open
mxjmtxrm opened this issue Jun 10, 2024 · 1 comment
Open

Questions about quantization #81

mxjmtxrm opened this issue Jun 10, 2024 · 1 comment

Comments

@mxjmtxrm
Copy link

mxjmtxrm commented Jun 10, 2024

Hi, great work!
I met some problems during 4bit weight-only quantization(--lwc).

  1. Is there any problem if the norm is nan?
  2. what's the best lwc hyper-parameter of LLama2 with different scales? like lwc-lr and epochs?
  3. Does more calib data bring better results?

I quantized a llama model using different lwc hyper-parameters and received different results.

  1. nsamples=1000, batch_size=1, epoch=2, the ppl is correct.
  2. nsamples=2000, batch_size=8, epoch=10, the ppl is super large (40000+).
    What's the problem?
@SherrySwift
Copy link

I found NaN norm during training, too. I guess it is caused by AMP training.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants