diff --git a/README.md b/README.md index 4fcf9a2..e54e5c5 100644 --- a/README.md +++ b/README.md @@ -8,18 +8,6 @@ So it's combining the best of RNN and transformer - **great performance, fast in **RWKV pip package**: https://pypi.org/project/rwkv/ -**Download RWKV-4 0.1/0.4/1.5/3/7/14B weights**: https://huggingface.co/BlinkDL - -**Discord**: https://discord.gg/bDSBUMeFpc - -**Twitter**: https://twitter.com/BlinkDL_AI - -**RWKV in 150 lines** (model, inference, text generation): https://github.com/BlinkDL/ChatRWKV/blob/main/RWKV_in_150_lines.py - -**ChatRWKV v2:** with "stream" and "split" strategies and INT8. **3G VRAM is enough to run RWKV 14B :)** https://github.com/BlinkDL/ChatRWKV/tree/main/v2 - -![RWKV-chat](RWKV-chat.png) - ```python os.environ["RWKV_JIT_ON"] = '1' os.environ["RWKV_CUDA_ON"] = '0' # if '1' then use CUDA kernel for seq mode (much faster) @@ -33,6 +21,19 @@ out, state = model.forward([1563], state) # RNN has state (use deepcop out, state = model.forward([310, 247], state) print(out.detach().cpu().numpy()) # same result as above ``` + +**Download RWKV-4 0.1/0.4/1.5/3/7/14B weights**: https://huggingface.co/BlinkDL + +**Discord**: https://discord.gg/bDSBUMeFpc + +**Twitter**: https://twitter.com/BlinkDL_AI + +**RWKV in 150 lines** (model, inference, text generation): https://github.com/BlinkDL/ChatRWKV/blob/main/RWKV_in_150_lines.py + +**ChatRWKV v2:** with "stream" and "split" strategies and INT8. **3G VRAM is enough to run RWKV 14B :)** https://github.com/BlinkDL/ChatRWKV/tree/main/v2 + +![RWKV-chat](RWKV-chat.png) + **Hugging Face space**: https://huggingface.co/spaces/BlinkDL/ChatRWKV-gradio You are welcome to join the RWKV discord https://discord.gg/bDSBUMeFpc to build upon it. We have plenty of potential compute (A100 40Gs) now (thanks to Stability and EleutherAI), so if you have interesting ideas I can run them.