Commit 08b1591
authored
Add AWQ-INT4 option to release script (#2906)
Summary:
Test Plan:
```
python quantize_and_upload.py --model_id Qwen/Qwen3-8B --quant AWQ-INT4 --push_to_hub --task bbh --calibration_limit 2
python quantize_and_upload.py --model_id microsoft/Phi-4-mini-instruct --quant AWQ-INT4 --push_to_hub --task mmlu_pro --calibration_limit 2
```
https://huggingface.co/pytorch/Qwen3-8B-AWQ-INT4
https://huggingface.co/pytorch/Phi-4-mini-instruct-AWQ-INT4
```
export TASK=bbh
export MODEL=pytorch/Qwen3-8B-AWQ-INT4
lm_eval --model hf --model_args pretrained=$MODEL --tasks $TASK --device cuda:0 --batch_size auto --limit 50
export MODEL=jerryzh168/Qwen3-8B-INT4
lm_eval --model hf --model_args pretrained=$MODEL --tasks $TASK --device cuda:0 --batch_size auto --limit 50
```
Qwen3-8B-INT4
hf (pretrained=jerryzh168/Qwen3-8B-INT4), gen_kwargs: (None), limit: 50.0, num_fewshot: None, batch_size: auto
| Tasks |Version| Filter |n-shot| Metric | |Value | |Stderr|
|----------------------------------------------------------|------:|----------|-----:|-----------|---|-----:|---|-----:|
|bbh | 3|get-answer| |exact_match|↑ |0.7444|± |0.0107|
| - bbh_cot_fewshot_boolean_expressions | 3|get-answer| 3|exact_match|↑ |0.9400|± |0.0339|
| - bbh_cot_fewshot_causal_judgement | 3|get-answer| 3|exact_match|↑ |0.5600|± |0.0709|
| - bbh_cot_fewshot_date_understanding | 3|get-answer| 3|exact_match|↑ |0.7600|± |0.0610|
| - bbh_cot_fewshot_disambiguation_qa | 3|get-answer| 3|exact_match|↑ |0.5600|± |0.0709|
| - bbh_cot_fewshot_dyck_languages | 3|get-answer| 3|exact_match|↑ |0.3000|± |0.0655|
| - bbh_cot_fewshot_formal_fallacies | 3|get-answer| 3|exact_match|↑ |0.6400|± |0.0686|
| - bbh_cot_fewshot_geometric_shapes | 3|get-answer| 3|exact_match|↑ |0.5400|± |0.0712|
| - bbh_cot_fewshot_hyperbaton | 3|get-answer| 3|exact_match|↑ |0.9800|± |0.0200|
| - bbh_cot_fewshot_logical_deduction_five_objects | 3|get-answer| 3|exact_match|↑ |0.6600|± |0.0677|
| - bbh_cot_fewshot_logical_deduction_seven_objects | 3|get-answer| 3|exact_match|↑ |0.3000|± |0.0655|
| - bbh_cot_fewshot_logical_deduction_three_objects | 3|get-answer| 3|exact_match|↑ |0.9400|± |0.0339|
| - bbh_cot_fewshot_movie_recommendation | 3|get-answer| 3|exact_match|↑ |0.6400|± |0.0686|
| - bbh_cot_fewshot_multistep_arithmetic_two | 3|get-answer| 3|exact_match|↑ |1.0000|± |0.0000|
| - bbh_cot_fewshot_navigate | 3|get-answer| 3|exact_match|↑ |0.8800|± |0.0464|
| - bbh_cot_fewshot_object_counting | 3|get-answer| 3|exact_match|↑ |0.8200|± |0.0549|
| - bbh_cot_fewshot_penguins_in_a_table | 3|get-answer| 3|exact_match|↑ |0.9000|± |0.0429|
| - bbh_cot_fewshot_reasoning_about_colored_objects | 3|get-answer| 3|exact_match|↑ |0.9000|± |0.0429|
| - bbh_cot_fewshot_ruin_names | 3|get-answer| 3|exact_match|↑ |0.7000|± |0.0655|
| - bbh_cot_fewshot_salient_translation_error_detection | 3|get-answer| 3|exact_match|↑ |0.5200|± |0.0714|
| - bbh_cot_fewshot_snarks | 3|get-answer| 3|exact_match|↑ |0.6000|± |0.0700|
| - bbh_cot_fewshot_sports_understanding | 3|get-answer| 3|exact_match|↑ |0.8200|± |0.0549|
| - bbh_cot_fewshot_temporal_sequences | 3|get-answer| 3|exact_match|↑ |0.9200|± |0.0388|
| - bbh_cot_fewshot_tracking_shuffled_objects_five_objects | 3|get-answer| 3|exact_match|↑ |0.8600|± |0.0496|
| - bbh_cot_fewshot_tracking_shuffled_objects_seven_objects| 3|get-answer| 3|exact_match|↑ |0.8200|± |0.0549|
| - bbh_cot_fewshot_tracking_shuffled_objects_three_objects| 3|get-answer| 3|exact_match|↑ |0.9400|± |0.0339|
| - bbh_cot_fewshot_web_of_lies | 3|get-answer| 3|exact_match|↑ |1.0000|± |0.0000|
| - bbh_cot_fewshot_word_sorting | 3|get-answer| 3|exact_match|↑ |0.6000|± |0.0700|
|Groups|Version| Filter |n-shot| Metric | |Value | |Stderr|
|------|------:|----------|------|-----------|---|-----:|---|-----:|
|bbh | 3|get-answer| |exact_match|↑ |0.7444|± |0.0107|
AWQ-INT4
hf (pretrained=jerryzh168/Qwen3-8B-AWQ-INT4), gen_kwargs: (None), limit: 50.0, num_fewshot: None, batch_size: auto
| Tasks |Version| Filter |n-shot| Metric | |Value | |Stderr|
|----------------------------------------------------------|------:|----------|-----:|-----------|---|-----:|---|-----:|
|bbh | 3|get-answer| |exact_match|↑ |0.7844|± |0.0101|
| - bbh_cot_fewshot_boolean_expressions | 3|get-answer| 3|exact_match|↑ |1.0000|± |0.0000|
| - bbh_cot_fewshot_causal_judgement | 3|get-answer| 3|exact_match|↑ |0.5800|± |0.0705|
| - bbh_cot_fewshot_date_understanding | 3|get-answer| 3|exact_match|↑ |0.8000|± |0.0571|
| - bbh_cot_fewshot_disambiguation_qa | 3|get-answer| 3|exact_match|↑ |0.5600|± |0.0709|
| - bbh_cot_fewshot_dyck_languages | 3|get-answer| 3|exact_match|↑ |0.5600|± |0.0709|
| - bbh_cot_fewshot_formal_fallacies | 3|get-answer| 3|exact_match|↑ |0.6000|± |0.0700|
| - bbh_cot_fewshot_geometric_shapes | 3|get-answer| 3|exact_match|↑ |0.4200|± |0.0705|
| - bbh_cot_fewshot_hyperbaton | 3|get-answer| 3|exact_match|↑ |0.9600|± |0.0280|
| - bbh_cot_fewshot_logical_deduction_five_objects | 3|get-answer| 3|exact_match|↑ |0.7000|± |0.0655|
| - bbh_cot_fewshot_logical_deduction_seven_objects | 3|get-answer| 3|exact_match|↑ |0.4000|± |0.0700|
| - bbh_cot_fewshot_logical_deduction_three_objects | 3|get-answer| 3|exact_match|↑ |0.9600|± |0.0280|
| - bbh_cot_fewshot_movie_recommendation | 3|get-answer| 3|exact_match|↑ |0.7000|± |0.0655|
| - bbh_cot_fewshot_multistep_arithmetic_two | 3|get-answer| 3|exact_match|↑ |1.0000|± |0.0000|
| - bbh_cot_fewshot_navigate | 3|get-answer| 3|exact_match|↑ |0.9400|± |0.0339|
| - bbh_cot_fewshot_object_counting | 3|get-answer| 3|exact_match|↑ |0.9200|± |0.0388|
| - bbh_cot_fewshot_penguins_in_a_table | 3|get-answer| 3|exact_match|↑ |0.8200|± |0.0549|
| - bbh_cot_fewshot_reasoning_about_colored_objects | 3|get-answer| 3|exact_match|↑ |0.9200|± |0.0388|
| - bbh_cot_fewshot_ruin_names | 3|get-answer| 3|exact_match|↑ |0.7400|± |0.0627|
| - bbh_cot_fewshot_salient_translation_error_detection | 3|get-answer| 3|exact_match|↑ |0.6400|± |0.0686|
| - bbh_cot_fewshot_snarks | 3|get-answer| 3|exact_match|↑ |0.6800|± |0.0666|
| - bbh_cot_fewshot_sports_understanding | 3|get-answer| 3|exact_match|↑ |0.8400|± |0.0524|
| - bbh_cot_fewshot_temporal_sequences | 3|get-answer| 3|exact_match|↑ |0.9400|± |0.0339|
| - bbh_cot_fewshot_tracking_shuffled_objects_five_objects | 3|get-answer| 3|exact_match|↑ |0.9600|± |0.0280|
| - bbh_cot_fewshot_tracking_shuffled_objects_seven_objects| 3|get-answer| 3|exact_match|↑ |0.9400|± |0.0339|
| - bbh_cot_fewshot_tracking_shuffled_objects_three_objects| 3|get-answer| 3|exact_match|↑ |0.9600|± |0.0280|
| - bbh_cot_fewshot_web_of_lies | 3|get-answer| 3|exact_match|↑ |1.0000|± |0.0000|
| - bbh_cot_fewshot_word_sorting | 3|get-answer| 3|exact_match|↑ |0.6400|± |0.0686|
|Groups|Version| Filter |n-shot| Metric | |Value | |Stderr|
|------|------:|----------|------|-----------|---|-----:|---|-----:|
|bbh | 3|get-answer| |exact_match|↑ |0.7844|± |0.0101|
```
Reviewers:
Subscribers:
Tasks:
Tags:1 parent 83a20c7 commit 08b1591
File tree
1 file changed
+125
-17
lines changed- .github/scripts/torchao_model_releases
1 file changed
+125
-17
lines changedLines changed: 125 additions & 17 deletions
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
10 | 10 | | |
11 | 11 | | |
12 | 12 | | |
| 13 | + | |
| 14 | + | |
| 15 | + | |
| 16 | + | |
13 | 17 | | |
14 | 18 | | |
15 | 19 | | |
| |||
19 | 23 | | |
20 | 24 | | |
21 | 25 | | |
| 26 | + | |
22 | 27 | | |
23 | 28 | | |
24 | 29 | | |
| |||
103 | 108 | | |
104 | 109 | | |
105 | 110 | | |
106 | | - | |
107 | | - | |
108 | 111 | | |
109 | 112 | | |
110 | 113 | | |
| |||
204 | 207 | | |
205 | 208 | | |
206 | 209 | | |
| 210 | + | |
| 211 | + | |
207 | 212 | | |
208 | 213 | | |
209 | 214 | | |
210 | 215 | | |
211 | 216 | | |
212 | 217 | | |
| 218 | + | |
| 219 | + | |
213 | 220 | | |
214 | 221 | | |
215 | 222 | | |
| |||
230 | 237 | | |
231 | 238 | | |
232 | 239 | | |
| 240 | + | |
| 241 | + | |
| 242 | + | |
| 243 | + | |
| 244 | + | |
| 245 | + | |
| 246 | + | |
| 247 | + | |
| 248 | + | |
| 249 | + | |
| 250 | + | |
| 251 | + | |
| 252 | + | |
| 253 | + | |
| 254 | + | |
| 255 | + | |
| 256 | + | |
| 257 | + | |
| 258 | + | |
| 259 | + | |
| 260 | + | |
| 261 | + | |
| 262 | + | |
| 263 | + | |
| 264 | + | |
| 265 | + | |
| 266 | + | |
| 267 | + | |
| 268 | + | |
| 269 | + | |
| 270 | + | |
| 271 | + | |
| 272 | + | |
| 273 | + | |
| 274 | + | |
| 275 | + | |
| 276 | + | |
233 | 277 | | |
234 | 278 | | |
| 279 | + | |
235 | 280 | | |
236 | 281 | | |
237 | 282 | | |
| |||
568 | 613 | | |
569 | 614 | | |
570 | 615 | | |
571 | | - | |
| 616 | + | |
| 617 | + | |
| 618 | + | |
572 | 619 | | |
573 | 620 | | |
574 | 621 | | |
| |||
580 | 627 | | |
581 | 628 | | |
582 | 629 | | |
583 | | - | |
| 630 | + | |
584 | 631 | | |
585 | 632 | | |
586 | 633 | | |
| |||
593 | 640 | | |
594 | 641 | | |
595 | 642 | | |
| 643 | + | |
596 | 644 | | |
597 | 645 | | |
598 | | - | |
599 | | - | |
600 | | - | |
| 646 | + | |
601 | 647 | | |
602 | 648 | | |
603 | 649 | | |
604 | 650 | | |
605 | | - | |
606 | | - | |
607 | | - | |
608 | | - | |
609 | | - | |
610 | | - | |
611 | | - | |
612 | | - | |
| 651 | + | |
| 652 | + | |
| 653 | + | |
| 654 | + | |
| 655 | + | |
| 656 | + | |
| 657 | + | |
| 658 | + | |
| 659 | + | |
| 660 | + | |
| 661 | + | |
| 662 | + | |
| 663 | + | |
| 664 | + | |
| 665 | + | |
| 666 | + | |
| 667 | + | |
| 668 | + | |
| 669 | + | |
| 670 | + | |
| 671 | + | |
| 672 | + | |
| 673 | + | |
| 674 | + | |
| 675 | + | |
| 676 | + | |
| 677 | + | |
| 678 | + | |
| 679 | + | |
| 680 | + | |
| 681 | + | |
| 682 | + | |
| 683 | + | |
| 684 | + | |
| 685 | + | |
| 686 | + | |
| 687 | + | |
| 688 | + | |
| 689 | + | |
| 690 | + | |
| 691 | + | |
| 692 | + | |
| 693 | + | |
| 694 | + | |
613 | 695 | | |
614 | 696 | | |
615 | 697 | | |
| |||
702 | 784 | | |
703 | 785 | | |
704 | 786 | | |
705 | | - | |
| 787 | + | |
| 788 | + | |
| 789 | + | |
| 790 | + | |
| 791 | + | |
| 792 | + | |
| 793 | + | |
| 794 | + | |
| 795 | + | |
| 796 | + | |
| 797 | + | |
| 798 | + | |
| 799 | + | |
| 800 | + | |
| 801 | + | |
| 802 | + | |
| 803 | + | |
| 804 | + | |
| 805 | + | |
| 806 | + | |
706 | 807 | | |
707 | 808 | | |
708 | 809 | | |
| |||
711 | 812 | | |
712 | 813 | | |
713 | 814 | | |
714 | | - | |
| 815 | + | |
| 816 | + | |
| 817 | + | |
| 818 | + | |
| 819 | + | |
| 820 | + | |
| 821 | + | |
| 822 | + | |
0 commit comments