ports@,
I'd like to update misc/llama.cpp and devel/libggml to an actual
"synchronization" between them.
Tested on -current/amd64, no regression on my usecase.
I also run make test on whisper.cpp, works as before update.
Ok?
Index: misc/llama.cpp/Makefile
===================================================================
RCS file: /home/cvs/ports/misc/llama.cpp/Makefile,v
diff -u -p -r1.14 Makefile
--- misc/llama.cpp/Makefile 14 Dec 2025 12:52:17 -0000 1.14
+++ misc/llama.cpp/Makefile 30 Jan 2026 20:40:29 -0000
@@ -2,7 +2,7 @@ COMMENT = LLM inference system
GH_ACCOUNT = ggml-org
GH_PROJECT = llama.cpp
-GH_TAGNAME = b7360
+GH_TAGNAME = b7883
PKGNAME = llama.cpp-0.0.${GH_TAGNAME:S/b//}
SHARED_LIBS += llama 3.0
@@ -13,14 +13,13 @@ CATEGORIES = misc
# MIT
PERMIT_PACKAGE = Yes
-WANTLIB += ${COMPILER_LIBCXX} c curl ggml ggml-base m
+WANTLIB += ${COMPILER_LIBCXX} c crypto ggml ggml-base m ssl
MODULES = devel/cmake
COMPILER = base-clang ports-gcc
# some tests need network access
-LIB_DEPENDS = devel/libggml \
- net/curl
+LIB_DEPENDS = devel/libggml
CONFIGURE_ARGS += -DLLAMA_USE_SYSTEM_GGML=on
CFLAGS += -I${LOCALBASE}/include
Index: misc/llama.cpp/distinfo
===================================================================
RCS file: /home/cvs/ports/misc/llama.cpp/distinfo,v
diff -u -p -r1.7 distinfo
--- misc/llama.cpp/distinfo 14 Dec 2025 12:52:17 -0000 1.7
+++ misc/llama.cpp/distinfo 30 Jan 2026 20:32:35 -0000
@@ -1,2 +1,2 @@
-SHA256 (llama.cpp-b7360.tar.gz) = nML2VgV6sqaB7UYkpy2ltjI7nRlGxoElcMOmWqx17Gk=
-SIZE (llama.cpp-b7360.tar.gz) = 28115754
+SHA256 (llama.cpp-b7883.tar.gz) = eR/Bt8eMbw4FM/0/By7RUyg/WmuI/hgYbwAf5Q1auwA=
+SIZE (llama.cpp-b7883.tar.gz) = 28882544
Index: misc/llama.cpp/pkg/PLIST
===================================================================
RCS file: /home/cvs/ports/misc/llama.cpp/pkg/PLIST,v
diff -u -p -r1.7 PLIST
--- misc/llama.cpp/pkg/PLIST 14 Dec 2025 12:52:17 -0000 1.7
+++ misc/llama.cpp/pkg/PLIST 21 Jan 2026 21:52:57 -0000
@@ -7,17 +7,18 @@ bin/convert_hf_to_gguf.py
@bin bin/llama-bench
@bin bin/llama-cli
@bin bin/llama-completion
+@bin bin/llama-debug
@bin bin/llama-diffusion-cli
@bin bin/llama-embedding
@bin bin/llama-eval-callback
@bin bin/llama-finetune
+@bin bin/llama-fit-params
@bin bin/llama-gen-docs
@bin bin/llama-gguf
@bin bin/llama-gguf-hash
@bin bin/llama-gguf-split
@bin bin/llama-idle
@bin bin/llama-imatrix
-@bin bin/llama-logits
@bin bin/llama-lookahead
@bin bin/llama-lookup
@bin bin/llama-lookup-create
@@ -29,7 +30,6 @@ bin/convert_hf_to_gguf.py
@bin bin/llama-perplexity
@bin bin/llama-quantize
@bin bin/llama-retrieval
-@bin bin/llama-run
@bin bin/llama-save-load-state
@bin bin/llama-server
@bin bin/llama-simple
Index: devel/libggml/Makefile
===================================================================
RCS file: /home/cvs/ports/devel/libggml/Makefile,v
diff -u -p -r1.6 Makefile
--- devel/libggml/Makefile 14 Dec 2025 12:51:12 -0000 1.6
+++ devel/libggml/Makefile 30 Jan 2026 20:02:31 -0000
@@ -2,12 +2,12 @@ COMMENT= tensor library for machine lea
GH_ACCOUNT= ggml-org
GH_PROJECT= ggml
-GH_COMMIT= f5425c0ee5e582a7d64411f06139870bff3e52e0
-DISTNAME= ggml-0.9.4pl20251211
+GH_COMMIT= ad04d9446254f6b5e352a944fece859940c0b50f
+DISTNAME= ggml-0.9.5pl20260130
PKGNAME= lib${DISTNAME}
-SHARED_LIBS += ggml 3.2
-SHARED_LIBS += ggml-base 3.2
+SHARED_LIBS += ggml 3.3
+SHARED_LIBS += ggml-base 3.3
CATEGORIES= devel
Index: devel/libggml/distinfo
===================================================================
RCS file: /home/cvs/ports/devel/libggml/distinfo,v
diff -u -p -r1.4 distinfo
--- devel/libggml/distinfo 14 Dec 2025 12:51:12 -0000 1.4
+++ devel/libggml/distinfo 30 Jan 2026 20:02:51 -0000
@@ -1,2 +1,2 @@
-SHA256 (ggml-0.9.4pl20251211-f5425c0e.tar.gz) =
l0I374JdNBkL1IVIYNQBgXRCSZLkc9F1rE/ZOm1Jy0g=
-SIZE (ggml-0.9.4pl20251211-f5425c0e.tar.gz) = 2407321
+SHA256 (ggml-0.9.5pl20260130-ad04d944.tar.gz) =
fcfr2CQUoGDuO7QGQTHOiC5UnSiEWSuSbCeslF2EIqU=
+SIZE (ggml-0.9.5pl20260130-ad04d944.tar.gz) = 2543796
Index: devel/libggml/patches/patch-CMakeLists_txt
===================================================================
RCS file: /home/cvs/ports/devel/libggml/patches/patch-CMakeLists_txt,v
diff -u -p -r1.1.1.1 patch-CMakeLists_txt
--- devel/libggml/patches/patch-CMakeLists_txt 1 Oct 2025 19:42:10 -0000
1.1.1.1
+++ devel/libggml/patches/patch-CMakeLists_txt 21 Jan 2026 21:10:56 -0000
@@ -2,7 +2,7 @@ Index: CMakeLists.txt
--- CMakeLists.txt.orig
+++ CMakeLists.txt
@@ -7,7 +7,7 @@ set(GGML_VERSION_MINOR 9)
- set(GGML_VERSION_PATCH 4)
+ set(GGML_VERSION_PATCH 5)
set(GGML_VERSION_BASE
"${GGML_VERSION_MAJOR}.${GGML_VERSION_MINOR}.${GGML_VERSION_PATCH}")
-find_program(GIT_EXE NAMES git git.exe NO_CMAKE_FIND_ROOT_PATH)
Index: devel/libggml/pkg/PFRAG.amd64
===================================================================
RCS file: /home/cvs/ports/devel/libggml/pkg/PFRAG.amd64,v
diff -u -p -r1.1 PFRAG.amd64
--- devel/libggml/pkg/PFRAG.amd64 5 Nov 2025 08:59:44 -0000 1.1
+++ devel/libggml/pkg/PFRAG.amd64 21 Jan 2026 21:15:09 -0000
@@ -1,8 +1,14 @@
@so lib/libggml-cpu-alderlake.so
+@so lib/libggml-cpu-cannonlake.so
+@so lib/libggml-cpu-cascadelake.so
+@so lib/libggml-cpu-cooperlake.so
@so lib/libggml-cpu-haswell.so
@so lib/libggml-cpu-icelake.so
+@so lib/libggml-cpu-ivybridge.so
+@so lib/libggml-cpu-piledriver.so
@so lib/libggml-cpu-sandybridge.so
@so lib/libggml-cpu-sapphirerapids.so
@so lib/libggml-cpu-skylakex.so
@so lib/libggml-cpu-sse42.so
@so lib/libggml-cpu-x64.so
+@so lib/libggml-cpu-zen4.so
Index: devel/libggml/pkg/PLIST
===================================================================
RCS file: /home/cvs/ports/devel/libggml/pkg/PLIST,v
diff -u -p -r1.3 PLIST
--- devel/libggml/pkg/PLIST 14 Dec 2025 12:51:12 -0000 1.3
+++ devel/libggml/pkg/PLIST 30 Jan 2026 20:13:34 -0000
@@ -11,6 +11,7 @@ include/ggml-metal.h
include/ggml-opt.h
include/ggml-rpc.h
include/ggml-sycl.h
+include/ggml-virtgpu.h
include/ggml-vulkan.h
include/ggml-webgpu.h
include/ggml-zendnn.h
--
wbr, Kirill