Skip to content

Commit

Permalink
Update README.md
Browse files Browse the repository at this point in the history
  • Loading branch information
SuperBruceJia authored Dec 9, 2024
1 parent 2786e11 commit 570564e
Showing 1 changed file with 10 additions and 0 deletions.
10 changes: 10 additions & 0 deletions quantization/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,16 @@ For `quantization.py`, we are using Python [AutoGPTQ](https://github.com/AutoGPT
python quantization.py "meta-llama/Meta-Llama-3-70B-Instruct" "./gptq_model" --bits 4 --group_size 128 --desc_act 1 --dtype bfloat16 --seqlen 2048 --damp 0.01
```

For `quantization_GPTQModel.py`, we are using Python [GPTQModel](https://github.com/ModelCloud/GPTQModel) package to conduct quantization.
```shell
pip install -v gptqmodel --no-build-isolation
```

Then,
```shell
python quantization_GPTQModel.py "meta-llama/Llama-3.3-70B-Instruct" "./gptq_model" --bits 4 --group_size 128 --seqlen 2048 --damp 0.01 --desc_act 1 --dtype bfloat16
```

For `quantization_HF.py`, we are using Hugging Face [transformers](https://github.com/huggingface/transformers) package to conduct quantization.
```shell
python quantization_HF.py --repo "meta-llama/Meta-Llama-3-70B-Instruct" --bits 4 --group_size 128
Expand Down

0 comments on commit 570564e

Please sign in to comment.