diff --git a/README.md b/README.md index 1f82c4c..e8ac570 100644 --- a/README.md +++ b/README.md @@ -10,6 +10,8 @@ So it's combining the best of RNN and transformer - **great performance, fast in Raven (7B finetuned on Alpaca) Demo: https://huggingface.co/spaces/BlinkDL/Raven-RWKV-7B +**ChatRWKV v2:** with "stream" and "split" strategies and INT8. **3G VRAM is enough to run RWKV 14B :)** https://github.com/BlinkDL/ChatRWKV/tree/main/v2 + **RWKV pip package**: https://pypi.org/project/rwkv/ ```python @@ -34,8 +36,6 @@ print(out.detach().cpu().numpy()) # same result as above **RWKV in 150 lines** (model, inference, text generation): https://github.com/BlinkDL/ChatRWKV/blob/main/RWKV_in_150_lines.py -**ChatRWKV v2:** with "stream" and "split" strategies and INT8. **3G VRAM is enough to run RWKV 14B :)** https://github.com/BlinkDL/ChatRWKV/tree/main/v2 - ![RWKV-chat](RWKV-chat.png) You are welcome to join the RWKV discord https://discord.gg/bDSBUMeFpc to build upon it. We have plenty of potential compute (A100 40Gs) now (thanks to Stability and EleutherAI), so if you have interesting ideas I can run them.