git: 03e4b270df01 - main - misc/py-llama-cpp-python: update 0.3.2 → 0.3.16

From: Yuri Victorovich <yuri_at_FreeBSD.org>
Date: Sat, 27 Sep 2025 16:33:16 UTC
The branch main has been updated by yuri:

URL: https://cgit.FreeBSD.org/ports/commit/?id=03e4b270df0112b60ad3d0f116f21d8c5c70978a

commit 03e4b270df0112b60ad3d0f116f21d8c5c70978a
Author:     Yuri Victorovich <yuri@FreeBSD.org>
AuthorDate: 2025-09-26 20:34:54 +0000
Commit:     Yuri Victorovich <yuri@FreeBSD.org>
CommitDate: 2025-09-27 16:33:12 +0000

    misc/py-llama-cpp-python: update 0.3.2 → 0.3.16
---
 misc/py-llama-cpp-python/Makefile                             |  6 +++---
 misc/py-llama-cpp-python/distinfo                             | 10 +++++-----
 .../patch-vendor_llama.cpp_ggml_src_ggml-cpu_ggml-cpu-impl.h  | 11 -----------
 3 files changed, 8 insertions(+), 19 deletions(-)

diff --git a/misc/py-llama-cpp-python/Makefile b/misc/py-llama-cpp-python/Makefile
index 2cc013916c94..6b4387d5bf18 100644
--- a/misc/py-llama-cpp-python/Makefile
+++ b/misc/py-llama-cpp-python/Makefile
@@ -1,6 +1,6 @@
 PORTNAME=	llama-cpp-python
 DISTVERSIONPREFIX=	v
-DISTVERSION=	0.3.2
+DISTVERSION=	0.3.16
 CATEGORIES=	misc # machine-learning
 PKGNAMEPREFIX=	${PYTHON_PKGNAMEPREFIX}
 
@@ -12,7 +12,7 @@ WWW=		https://llama-cpp-python.readthedocs.io/en/latest/ \
 LICENSE=	MIT
 LICENSE_FILE=	${WRKSRC}/LICENSE.md
 
-BUILD_DEPENDS=	${PYTHON_PKGNAMEPREFIX}scikit-build-core>0:devel/py-scikit-build-core@${PY_FLAVOR} \
+BUILD_DEPENDS=	${PYTHON_PKGNAMEPREFIX}scikit-build-core>=0.9.2:devel/py-scikit-build-core@${PY_FLAVOR} \
 		cmake:devel/cmake-core
 LIB_DEPENDS=	libvulkan.so:graphics/vulkan-loader
 RUN_DEPENDS=	${PYTHON_PKGNAMEPREFIX}diskcache>=5.6.1:devel/py-diskcache@${PY_FLAVOR} \
@@ -34,7 +34,7 @@ USE_PYTHON=	pep517 autoplist pytest
 
 USE_GITHUB=	yes
 GH_ACCOUNT=	abetlen
-GH_TUPLE=	ggerganov:llama.cpp:74d73dc:cpp/vendor/llama.cpp
+GH_TUPLE=	ggerganov:llama.cpp:b6598:cpp/vendor/llama.cpp
 
 SHEBANG_GLOB=	*.py
 
diff --git a/misc/py-llama-cpp-python/distinfo b/misc/py-llama-cpp-python/distinfo
index a25d0e5c74ba..7f4849c553a3 100644
--- a/misc/py-llama-cpp-python/distinfo
+++ b/misc/py-llama-cpp-python/distinfo
@@ -1,5 +1,5 @@
-TIMESTAMP = 1731907964
-SHA256 (abetlen-llama-cpp-python-v0.3.2_GH0.tar.gz) = 491057624a095ff6d4eae1bcc139b579f72d03d9deee1a2ab05d2f980ec824ed
-SIZE (abetlen-llama-cpp-python-v0.3.2_GH0.tar.gz) = 274530
-SHA256 (ggerganov-llama.cpp-74d73dc_GH0.tar.gz) = 9cc0aad9746b93b461c53d7038211ea7cbe70df2aa1a23b2daac07af935f6990
-SIZE (ggerganov-llama.cpp-74d73dc_GH0.tar.gz) = 19564318
+TIMESTAMP = 1758915707
+SHA256 (abetlen-llama-cpp-python-v0.3.16_GH0.tar.gz) = d6ae5a6ac40dda4d14c6bb8f5e9504d28f442cf810263661d457c948c386f2a4
+SIZE (abetlen-llama-cpp-python-v0.3.16_GH0.tar.gz) = 279565
+SHA256 (ggerganov-llama.cpp-b6598_GH0.tar.gz) = cd296792f49695bd44e885a5c1fcefe4ef72f3da7f8933be1378f944116515a3
+SIZE (ggerganov-llama.cpp-b6598_GH0.tar.gz) = 25833773
diff --git a/misc/py-llama-cpp-python/files/patch-vendor_llama.cpp_ggml_src_ggml-cpu_ggml-cpu-impl.h b/misc/py-llama-cpp-python/files/patch-vendor_llama.cpp_ggml_src_ggml-cpu_ggml-cpu-impl.h
deleted file mode 100644
index 376142c64242..000000000000
--- a/misc/py-llama-cpp-python/files/patch-vendor_llama.cpp_ggml_src_ggml-cpu_ggml-cpu-impl.h
+++ /dev/null
@@ -1,11 +0,0 @@
---- vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu-impl.h.orig	2025-08-02 23:17:06 UTC
-+++ vendor/llama.cpp/ggml/src/ggml-cpu/ggml-cpu-impl.h
-@@ -319,8 +319,6 @@ inline static int32x4_t ggml_vdotq_s32(int32x4_t acc, 
- #else
- #ifdef __POWER9_VECTOR__
- #include <altivec.h>
--#undef bool
--#define bool _Bool
- #else
- #if defined(_MSC_VER) || defined(__MINGW32__)
- #include <intrin.h>