llama-cli inference batching 2025-08-08
llama.cpp Created 2025-07-16 Updated 2025-07-16
askubuntu.com/questions/1461564/install-llama-cpp-locally has some tutorials for Ubuntu. There was no nicely pre-packaged one for Ubuntu 25.04, but build worked on 79e0b68c178656bb0632cb8602d2940b755077f8 In particular it exposed Vulkan support before Ollama did: github.com/ollama/ollama/pull/5059 and it did seem to work, using up my AMD GPU.
Ollama deterministic output Created 2025-03-20 Updated 2025-07-16