From a9007581d06e121aa81c61163b4dc1f6bd855b6b Mon Sep 17 00:00:00 2001 From: PENG Bo <33809201+BlinkDL@users.noreply.github.com> Date: Tue, 28 Mar 2023 19:27:14 +0800 Subject: [PATCH] Update README.md --- README.md | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index c01fa06..1f82c4c 100644 --- a/README.md +++ b/README.md @@ -6,6 +6,10 @@ RWKV is an RNN with Transformer-level LLM performance, which can also be directl So it's combining the best of RNN and transformer - **great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding** (using the final hidden state). +**HuggingFace Gradio demo (14B ctx8192)**: https://huggingface.co/spaces/BlinkDL/ChatRWKV-gradio + +Raven (7B finetuned on Alpaca) Demo: https://huggingface.co/spaces/BlinkDL/Raven-RWKV-7B + **RWKV pip package**: https://pypi.org/project/rwkv/ ```python @@ -34,8 +38,6 @@ print(out.detach().cpu().numpy()) # same result as above ![RWKV-chat](RWKV-chat.png) -**Hugging Face space**: https://huggingface.co/spaces/BlinkDL/ChatRWKV-gradio - You are welcome to join the RWKV discord https://discord.gg/bDSBUMeFpc to build upon it. We have plenty of potential compute (A100 40Gs) now (thanks to Stability and EleutherAI), so if you have interesting ideas I can run them. ![RWKV-eval2](RWKV-eval2.png)