From c59ee12659136149eee520d10600fdd373861675 Mon Sep 17 00:00:00 2001 From: randaller Date: Sat, 11 Mar 2023 00:33:24 +0300 Subject: [PATCH] Update README.md --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 2216c5e..e3c22af 100644 --- a/README.md +++ b/README.md @@ -21,7 +21,7 @@ One may run with 32 Gb of RAM, but inference will be slow (with the speed of you I am running this on 12700k/128 Gb RAM/NVIDIA 3070ti 8Gb/fast huge nvme and getting one token from 30B model in a few seconds. -For example, **30B model uses around 70 Gb of RAM**. 7B model fits into 18 Gb. +For example, **30B model uses around 70 Gb of RAM**. 7B model fits into 18 Gb. 13B model uses 48 Gb. If you do not have powerful videocard, you may use another repo for cpu-only inference: https://github.com/randaller/llama-cpu