diff --git a/README.md b/README.md index 0e31191..9d622d5 100755 --- a/README.md +++ b/README.md @@ -1,4 +1,4 @@ -# LLaMA +# Inference LLaMA models using CPU only This repository is intended as a minimal, hackable and readable example to load [LLaMA](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/) ([arXiv](https://arxiv.org/abs/2302.13971v1)) models and run inference. In order to download the checkpoints and tokenizer, fill this [google form](https://forms.gle/jk851eBVbX1m5TAv5) @@ -43,3 +43,8 @@ See [MODEL_CARD.md](MODEL_CARD.md) ### License See the [LICENSE](LICENSE) file. + +### CPU Inference +Place tokenizer.model and tokenizer_checklist.chk into /tokenizer folder +Place three files of 7B model into /model folder +Run python example-cpu.py