Place tokenizer.model file from torrent into repo's [/tokenizer] folder.
Place tokenizer.model file from torrent into repo's [/tokenizer] folder.
Place consolidated.00.pth and params.json from 7B torrent folder into repo's [/model] folder.
Place model files from torrent folder (for example, [/13B]) into repo's [/model] folder.
Run the example:
```
python example-cpu.py
```
### CPU Inference of 13B, 30B and 65B models
A little bit tricky part is that we need to unshard the checkpoints first. In this example, D:\Downloads\LLaMA is a root folder of downloaded torrent with models. Run the following command to create merged weights checkpoint: