diff --git a/README.md b/README.md index 1b0cbbd..c03efdc 100644 --- a/README.md +++ b/README.md @@ -21,11 +21,11 @@ Share your best prompts, chats or generations here in this issue: https://github One may run with 32 Gb of RAM, but inference will be slow (with the speed of your swap file reading) -I am running this on 12700k/128 Gb RAM/NVIDIA 3070ti 8Gb/fast huge nvme and getting one token from 30B model in a few seconds. +I am running this on a [12700k/128 Gb RAM/NVIDIA 3070ti 8Gb/fast huge nvme with 256 Gb swap for 65B model] and getting one token from 30B model in a few seconds. For example, **30B model uses around 70 Gb of RAM**. 7B model fits into 18 Gb. 13B model uses 48 Gb. -If you do not have powerful videocard, you may use another repo for cpu-only inference: https://github.com/randaller/llama-cpu +If you do not have nvidia videocard, you may use another repo for cpu-only inference: https://github.com/randaller/llama-cpu ### Conda Environment Setup Example for Windows 10+ Download and install Anaconda Python https://www.anaconda.com and run Anaconda Prompt