On Fri, 2025-01-31 at 12:31 +0100, Simon Josefsson wrote: > Package: wnpp > Severity: wishlist > Owner: Simon Josefsson <[email protected]> > X-Debbugs-CC: [email protected], [email protected] > > * Package name : ollama > Version : 0.5.7-1 > Upstream Author : Ollama > * URL : https://github.com/ollama/ollama > * License : Expat > Programming Lang: Go > Description : large language model tools > > Ollama: Get up and running with large language models. > > https://salsa.debian.org/go-team/packages/ollama > https://salsa.debian.org/jas/ollama/-/pipelines
@ckk is planning to package llama.cpp within debian deep learning team ([email protected]). Maybe you want to discuss with the team whether you want to deal with the embedded copy of llama.cpp inside ollama source tree? I did not look into how ollama enables ROCm and CUDA support, but that's also something the team care about. Do you want to enable any of them?

