Bug#1109124: llama.cpp: CVE-2025-53630
Hi Salvatore,
On 2025-07-13 07:49, Salvatore Bonaccorso wrote:
> On Sat, Jul 12, 2025 at 12:04:34AM +0200, Christian Kastner wrote:
>> Nevertheless, I really need to figure out a better way to deal with
>> llama.cpp, whisper.cpp, and ggml triad. Re-embedding isn't an option as
>> the ggml build is already pretty complicated by itself, adding another
>> layer would be a pain.
>
> Thanks. The gguf.cpp as emmbedded in llama.cpp is compiled and used,
> is that correct? Do we use the external ggml in the system?
That's how llama.cpp is primarily developed distributed, but for Debian
we ignore the embedded version and build a standalone src:ggml.
This is mainly because the ggml build is already quite complex (multiple
CPU and GPU backends), and redundant between llama.cpp and whisper.cpp
(still in NEW).
llama.cpp and whisper.cpp should eventually +ds clean their embedded
copies, I just want to wait a bit more to see if a standalone src:ggml
remains viable.
Best,
Christian
Reply to: