From c62818e84e71749afd916f79f5de70a5b210e7d4 Mon Sep 17 00:00:00 2001 From: Yu Li Date: Thu, 7 Dec 2023 19:07:03 -0600 Subject: [PATCH] update --- air_llm/README.md | 25 +++++++++++++++++++++++++ 1 file changed, 25 insertions(+) diff --git a/air_llm/README.md b/air_llm/README.md index c3e4ba0..c1a104b 100644 --- a/air_llm/README.md +++ b/air_llm/README.md @@ -269,6 +269,31 @@ If you run into this error, most possible cause is you run out of disk space. Th 如果你碰到这个error,很有可能是空间不足。可以参考一下[这个](https://huggingface.co/TheBloke/guanaco-65B-GPTQ/discussions/12) 可能需要扩大硬盘空间,删除huggingface的[.cache](https://huggingface.co/docs/datasets/cache),然后重新run。 +### 2. ValueError: max() arg is an empty sequence + +Most likely you are loading QWen or ChatGLM model with Llama2 class. Try the following: + +For QWen model: + +```python +from airllm import AirLLMQWen #<----- instead of AirLLMLlama2 +``` + +For ChatGLM model: + +```python +from airllm import AirLLM ChatGLM #<----- instead of AirLLMLlama2 +``` + +### 3. 401 Client Error....Repo model ... is gated. + +Some models are gated models, needs huggingface api token. You can provide hf_token: + +```python +model = AirLLMLlama2("meta-llama/Llama-2-7b-hf", #hf_token='HF_API_TOKEN') +``` + + ## Contribution Welcome contribution, ideas and discussions!