Hi Salvatore, On 2025-07-13 07:49, Salvatore Bonaccorso wrote: > On Sat, Jul 12, 2025 at 12:04:34AM +0200, Christian Kastner wrote: >> Nevertheless, I really need to figure out a better way to deal with >> llama.cpp, whisper.cpp, and ggml triad. Re-embedding isn't an option as >> the ggml build is already pretty complicated by itself, adding another >> layer would be a pain. > > Thanks. The gguf.cpp as emmbedded in llama.cpp is compiled and used, > is that correct? Do we use the external ggml in the system?
That's how llama.cpp is primarily developed distributed, but for Debian we ignore the embedded version and build a standalone src:ggml. This is mainly because the ggml build is already quite complex (multiple CPU and GPU backends), and redundant between llama.cpp and whisper.cpp (still in NEW). llama.cpp and whisper.cpp should eventually +ds clean their embedded copies, I just want to wait a bit more to see if a standalone src:ggml remains viable. Best, Christian

