Using LLaMA.cpp
Learn how to set up and configure LLaMA.cpp as a local AI backend for Arcania
⚠️ Dieser Inhalt ist noch nicht in Ihrer Sprache verfügbar.
You can find the full llama.cpp documentation here.
Installation Steps
-
Clone the repo:
-
Download the model: For example, we will use OpenChat 3.5 model, which is what is used on the demo instance. There are many models to choose from.
Navigate to TheBloke/openchat_3.5-GGUF and download one of the models, such as
openchat_3.5.Q5_K_M.gguf
. Place this file inside the./models
directory. -
Build the server:
-
Run the server: Read the llama.cpp documentation for more information on the server options. Or run
./server --help
.
Configuration
Enable the server in the client: