From 6eb999fbcd13185d05943bc24010f01f87089adf Mon Sep 17 00:00:00 2001 From: Yu Li Date: Sat, 18 Nov 2023 08:13:57 -0600 Subject: [PATCH] update readme --- README.md | 2 ++ air_llm/setup.py | 2 +- 2 files changed, 3 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 81199d4..997a21c 100644 --- a/README.md +++ b/README.md @@ -21,6 +21,8 @@ This is the first open source 33B Chinese LLM, we also support DPO alignment tra [![Generic badge](https://img.shields.io/badge/wechat-Anima-brightgreen?logo=wechat)](https://static.aicompose.cn/static/wecom_barcode.png?t=1671918938) [![Generic badge](https://img.shields.io/badge/🤗-Huggingface%20Repo-green.svg)](https://huggingface.co/lyogavin/Anima33B-merged) [![Generic badge](https://img.shields.io/badge/🤗-Huggingface%20Repo-green.svg)](https://huggingface.co/lyogavin/Anima-7B-100K) +[![Discord](https://img.shields.io/discord/1175437549783760896)](https://img.shields.io/discord/1175437549783760896) + ## 🔄 更新 Updates diff --git a/air_llm/setup.py b/air_llm/setup.py index 1730565..acb74dc 100644 --- a/air_llm/setup.py +++ b/air_llm/setup.py @@ -5,7 +5,7 @@ with open("README.md", "r") as fh: setuptools.setup( name="airllm", - version="0.9.3", + version="0.9.4", author="Gavin Li", author_email="gavinli@animaai.cloud", description="AirLLM allows single 4GB GPU card to run 70B large language models without quantization, distillation or pruning.",