git: 15a257d5f8b3 - main - misc/llama-cpp: update 8132 → 8182
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Sat, 28 Feb 2026 16:57:37 UTC
The branch main has been updated by yuri:
URL: https://cgit.FreeBSD.org/ports/commit/?id=15a257d5f8b3fd93cf23cf61361253024a651413
commit 15a257d5f8b3fd93cf23cf61361253024a651413
Author: Yuri Victorovich <yuri@FreeBSD.org>
AuthorDate: 2026-02-28 16:54:53 +0000
Commit: Yuri Victorovich <yuri@FreeBSD.org>
CommitDate: 2026-02-28 16:57:34 +0000
misc/llama-cpp: update 8132 → 8182
It now uses the the ggml package as dependency instead of the bundled one.
PR: 293496 (for ggml part)
Reported by: mord0d@firemail.cc
---
misc/llama-cpp/Makefile | 8 +++++---
misc/llama-cpp/distinfo | 6 +++---
misc/llama-cpp/pkg-plist | 35 ++---------------------------------
3 files changed, 10 insertions(+), 39 deletions(-)
diff --git a/misc/llama-cpp/Makefile b/misc/llama-cpp/Makefile
index c2612c1e54b8..bd914adab036 100644
--- a/misc/llama-cpp/Makefile
+++ b/misc/llama-cpp/Makefile
@@ -1,6 +1,6 @@
PORTNAME= llama-cpp
DISTVERSIONPREFIX= b
-DISTVERSION= 8132
+DISTVERSION= 8182
CATEGORIES= misc # machine-learning
MAINTAINER= yuri@FreeBSD.org
@@ -13,6 +13,8 @@ LICENSE_FILE= ${WRKSRC}/LICENSE
BROKEN_armv7= clang crashes, see https://bugs.freebsd.org/bugzilla/show_bug.cgi?id=278810
BROKEN_i386= compilation fails, see https://github.com/ggerganov/llama.cpp/issues/9545
+LIB_DEPENDS= libggml-base.so:misc/ggml
+
USES= cmake:testing compiler:c++11-lang python:run shebangfix
USE_LDCONFIG= yes
@@ -23,8 +25,8 @@ GH_TUPLE= nomic-ai:kompute:4565194:kompute/kompute
SHEBANG_GLOB= *.py
-CMAKE_ON= BUILD_SHARED_LIBS
- #LLAMA_USE_SYSTEM_GGML
+CMAKE_ON= BUILD_SHARED_LIBS \
+ LLAMA_USE_SYSTEM_GGML
CMAKE_OFF= GGML_NATIVE \
FREEBSD_ALLOW_ADVANCED_CPU_FEATURES \
LLAMA_BUILD_TESTS
diff --git a/misc/llama-cpp/distinfo b/misc/llama-cpp/distinfo
index d75ba99c1802..e88a96bf93ca 100644
--- a/misc/llama-cpp/distinfo
+++ b/misc/llama-cpp/distinfo
@@ -1,5 +1,5 @@
-TIMESTAMP = 1771831071
-SHA256 (ggerganov-llama.cpp-b8132_GH0.tar.gz) = 346e3ec5c8146947d4cdda6bf340384006c590b0c4f8a3c89f265dd076351f57
-SIZE (ggerganov-llama.cpp-b8132_GH0.tar.gz) = 29059325
+TIMESTAMP = 1772296554
+SHA256 (ggerganov-llama.cpp-b8182_GH0.tar.gz) = 3bfffdc4fde15b90578010714ca0721c33f13c544cf2d8dc170c71a139b60777
+SIZE (ggerganov-llama.cpp-b8182_GH0.tar.gz) = 29080587
SHA256 (nomic-ai-kompute-4565194_GH0.tar.gz) = 95b52d2f0514c5201c7838348a9c3c9e60902ea3c6c9aa862193a212150b2bfc
SIZE (nomic-ai-kompute-4565194_GH0.tar.gz) = 13540496
diff --git a/misc/llama-cpp/pkg-plist b/misc/llama-cpp/pkg-plist
index 4a4098d35727..75c675c1a959 100644
--- a/misc/llama-cpp/pkg-plist
+++ b/misc/llama-cpp/pkg-plist
@@ -38,47 +38,16 @@ bin/convert_hf_to_gguf.py
%%EXAMPLES%%bin/llama-speculative-simple
%%EXAMPLES%%bin/llama-tokenize
%%EXAMPLES%%bin/llama-tts
-include/ggml-alloc.h
-include/ggml-backend.h
-include/ggml-blas.h
-include/ggml-cann.h
-include/ggml-cpp.h
-include/ggml-cpu.h
-include/ggml-cuda.h
-include/ggml-metal.h
-include/ggml-opt.h
-include/ggml-rpc.h
-include/ggml-sycl.h
-include/ggml-virtgpu.h
-include/ggml-vulkan.h
-include/ggml-webgpu.h
-include/ggml-zendnn.h
-include/ggml.h
-include/gguf.h
include/llama-cpp.h
include/llama.h
include/mtmd-helper.h
include/mtmd.h
-lib/cmake/ggml/ggml-config.cmake
-lib/cmake/ggml/ggml-version.cmake
lib/cmake/llama/llama-config.cmake
lib/cmake/llama/llama-version.cmake
-lib/libggml-base.so
-lib/libggml-base.so.0
-lib/libggml-base.so.0.9.7
-lib/libggml-cpu.so
-lib/libggml-cpu.so.0
-lib/libggml-cpu.so.0.9.7
-lib/libggml-vulkan.so
-lib/libggml-vulkan.so.0
-lib/libggml-vulkan.so.0.9.7
-lib/libggml.so
-lib/libggml.so.0
-lib/libggml.so.0.9.7
lib/libllama.so
lib/libllama.so.0
-lib/libllama.so.0.0.8132
+lib/libllama.so.0.0.8182
lib/libmtmd.so
lib/libmtmd.so.0
-lib/libmtmd.so.0.0.8132
+lib/libmtmd.so.0.0.8182
libdata/pkgconfig/llama.pc