-
Notifications
You must be signed in to change notification settings - Fork 163
[OpenVINO] Add gsm8k as a dataset option for CausalLM quantization #1547
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
|
@ljaljushkin, please take a look. |
ljaljushkin
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, Nikita! LGTM
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM ! isn't the perf increase to be expected since you are calibrating on on test data/distribution ?
That's a good question! To be honest some experiments show higher accuracy when calibrating on
|
rkazants
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Is it possible and reasonable to specify several datasets (e.g., wikitext2 and gsm8k) simultaneously for calibration? Can it provide higher accuracy?
I think it is reasonable to update documentation and mention in what cases to use gsm8k dataset.
Thanks for the idea! In theory it could, however there were no experiments done in this direction.
Experiments are still work-in-progress in order to determine when exactly it is beneficial to use gsm8k dataset. For now, the results are mixed, however in some cases gsm8k provides better quality. With this PR we add another option for users to try improve quantized model quality. |
What does this PR do?
In this PR "gsm8k" dataset option is added to apply data-aware quantization to causal language modes. Calibrating on "gsm8k" in some cases provides better results compared to "wikitext2":
Before submitting