mirror of https://github.com/novarobot/llama.cpp
Respect the maximum number of tokens in interactive. (#298)
Co-authored-by: Johnman <johnman@github> Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>llama-patch-enable-fma-msvc master-368d0c8
parent
50fae10d03
commit
368d0c8a9e
Loading…
Reference in New Issue