git: fb4ebd8c85c6 - main - misc/llama-cpp: update 7709 → 8132
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Mon, 23 Feb 2026 07:25:32 UTC
The branch main has been updated by yuri:
URL: https://cgit.FreeBSD.org/ports/commit/?id=fb4ebd8c85c66924eb2957415541d09b18dd0884
commit fb4ebd8c85c66924eb2957415541d09b18dd0884
Author: Yuri Victorovich <yuri@FreeBSD.org>
AuthorDate: 2026-02-23 07:25:16 +0000
Commit: Yuri Victorovich <yuri@FreeBSD.org>
CommitDate: 2026-02-23 07:25:23 +0000
misc/llama-cpp: update 7709 → 8132
Reported by: portscout
---
misc/llama-cpp/Makefile | 2 +-
misc/llama-cpp/distinfo | 6 +++---
misc/llama-cpp/pkg-plist | 17 +++++++++--------
3 files changed, 13 insertions(+), 12 deletions(-)
diff --git a/misc/llama-cpp/Makefile b/misc/llama-cpp/Makefile
index 12023ccb8981..c2612c1e54b8 100644
--- a/misc/llama-cpp/Makefile
+++ b/misc/llama-cpp/Makefile
@@ -1,6 +1,6 @@
PORTNAME= llama-cpp
DISTVERSIONPREFIX= b
-DISTVERSION= 7709
+DISTVERSION= 8132
CATEGORIES= misc # machine-learning
MAINTAINER= yuri@FreeBSD.org
diff --git a/misc/llama-cpp/distinfo b/misc/llama-cpp/distinfo
index 13090992cbe0..d75ba99c1802 100644
--- a/misc/llama-cpp/distinfo
+++ b/misc/llama-cpp/distinfo
@@ -1,5 +1,5 @@
-TIMESTAMP = 1768206387
-SHA256 (ggerganov-llama.cpp-b7709_GH0.tar.gz) = 8aa5d02ec90c70fa496cc878ef3962733e74184a8e43f191db5471288f5cf911
-SIZE (ggerganov-llama.cpp-b7709_GH0.tar.gz) = 28712220
+TIMESTAMP = 1771831071
+SHA256 (ggerganov-llama.cpp-b8132_GH0.tar.gz) = 346e3ec5c8146947d4cdda6bf340384006c590b0c4f8a3c89f265dd076351f57
+SIZE (ggerganov-llama.cpp-b8132_GH0.tar.gz) = 29059325
SHA256 (nomic-ai-kompute-4565194_GH0.tar.gz) = 95b52d2f0514c5201c7838348a9c3c9e60902ea3c6c9aa862193a212150b2bfc
SIZE (nomic-ai-kompute-4565194_GH0.tar.gz) = 13540496
diff --git a/misc/llama-cpp/pkg-plist b/misc/llama-cpp/pkg-plist
index 84f8c70f2bbd..4a4098d35727 100644
--- a/misc/llama-cpp/pkg-plist
+++ b/misc/llama-cpp/pkg-plist
@@ -49,6 +49,7 @@ include/ggml-metal.h
include/ggml-opt.h
include/ggml-rpc.h
include/ggml-sycl.h
+include/ggml-virtgpu.h
include/ggml-vulkan.h
include/ggml-webgpu.h
include/ggml-zendnn.h
@@ -64,20 +65,20 @@ lib/cmake/llama/llama-config.cmake
lib/cmake/llama/llama-version.cmake
lib/libggml-base.so
lib/libggml-base.so.0
-lib/libggml-base.so.0.9.5
+lib/libggml-base.so.0.9.7
lib/libggml-cpu.so
lib/libggml-cpu.so.0
-lib/libggml-cpu.so.0.9.5
-%%VULKAN%%lib/libggml-vulkan.so
-%%VULKAN%%lib/libggml-vulkan.so.0
-%%VULKAN%%lib/libggml-vulkan.so.0.9.5
+lib/libggml-cpu.so.0.9.7
+lib/libggml-vulkan.so
+lib/libggml-vulkan.so.0
+lib/libggml-vulkan.so.0.9.7
lib/libggml.so
lib/libggml.so.0
-lib/libggml.so.0.9.5
+lib/libggml.so.0.9.7
lib/libllama.so
lib/libllama.so.0
-lib/libllama.so.0.0.7709
+lib/libllama.so.0.0.8132
lib/libmtmd.so
lib/libmtmd.so.0
-lib/libmtmd.so.0.0.7709
+lib/libmtmd.so.0.0.8132
libdata/pkgconfig/llama.pc