mirror of
https://github.com/0xSojalSec/airllm.git
synced 2026-03-07 22:33:47 +00:00
update readme
This commit is contained in:
@@ -93,7 +93,7 @@ We just added model compression based on block-wise quantization based model com
|
||||
|
||||
```python
|
||||
model = AirLLMLlama2("garage-bAInd/Platypus2-70B-instruct",
|
||||
compression='4bit' # specify '8bit' for 8-bit block-wise quantization
|
||||
compression='4bit' # specify '8bit' for 8-bit block-wise quantization
|
||||
)
|
||||
```
|
||||
|
||||
|
||||
Reference in New Issue
Block a user