Get llama.cpp up and running in Wave AI
brew install ggerganov/ggerganov/llama.cpp
. This command will install two tools: llama-cli
for command-line interaction and llama-server
for running the llama.cpp HTTP server — both of which use the same options and syntax as the manually built llama.cpp.aibaseurl
. This parameter can be set either through the UI or from the command line, but please note that the parameter names are slightly different depending on the method you choose.
8080
may be different depending on your specific installation. For remote llama.cpp server instances, replace localhost
with the appropriate hostname or IP address of the server where llama.cpp is running. If the port number is different from the default 8080
, update it accordingly in the URL.AI Base URL
parameter as described in the previous section.
aibaseurl
parameter using the /client:set command, as shown in the example below.
ctrl + space
shortcut. This will open an interactive chat session where you can have a continuous conversation with the AI assistant powered by your llama.cpp model.aitimeout
parameter to a higher value. This will give your llama.cpp server more time to process and respond to your requests, especially if you are running it on a system with limited hardware resources.aibaseurl
parameter points to the correct URL and port number where the llama.cpp server is running. If you have changed the default port or are running llama.cpp on a remote server, update the URL accordingly.aibaseurl
and aimodel
parameters to their default values and reconfiguring llama.cpp from scratch. This can help rule out any configuration issues that might be causing problems.aibaseurl
and aimodel
parameters to their default state by using the following commands.