From 123536b2a7045b22939c9ff28aec25a58ba33fef Mon Sep 17 00:00:00 2001 From: PENG Bo <33809201+BlinkDL@users.noreply.github.com> Date: Wed, 8 Mar 2023 13:38:56 +0800 Subject: [PATCH] Update README.md --- README.md | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index 454ce27..e682609 100644 --- a/README.md +++ b/README.md @@ -12,7 +12,8 @@ So it's combining the best of RNN and transformer - **great performance, fast in **RWKV in 150 lines** (model, inference, text generation): https://github.com/BlinkDL/ChatRWKV/blob/main/RWKV_in_150_lines.py -**ChatRWKV v2:** with "stream" and "split" strategies. **3G VRAM is enough to run RWKV 14B :)** https://github.com/BlinkDL/ChatRWKV/tree/main/v2 +**ChatRWKV v2:** with "stream" and "split" strategies and INT8. **3G VRAM is enough to run RWKV 14B :)** https://github.com/BlinkDL/ChatRWKV/tree/main/v2 +![RWKV-chat](RWKV-chat.png) ```python os.environ["RWKV_JIT_ON"] = '1' os.environ["RWKV_CUDA_ON"] = '0' # if '1' then use CUDA kernel for seq mode (much faster) @@ -28,8 +29,6 @@ print(out.detach().cpu().numpy()) # same result as above ``` **Hugging Face space**: https://huggingface.co/spaces/BlinkDL/ChatRWKV-gradio -![RWKV-chat](RWKV-chat.png) - ## Join our Discord: https://discord.gg/bDSBUMeFpc :) You are welcome to join the RWKV discord https://discord.gg/bDSBUMeFpc to build upon it. We have plenty of potential compute (A100 40Gs) now (thanks to Stability and EleutherAI), so if you have interesting ideas I can run them.