git: 10580f9c51f6 - main - misc/llama-cpp: update 3135 → 3145
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Fri, 14 Jun 2024 09:11:59 UTC
The branch main has been updated by yuri:
URL: https://cgit.FreeBSD.org/ports/commit/?id=10580f9c51f6d16af473cffc0ed622effc046e86
commit 10580f9c51f6d16af473cffc0ed622effc046e86
Author: Yuri Victorovich <yuri@FreeBSD.org>
AuthorDate: 2024-06-14 08:55:03 +0000
Commit: Yuri Victorovich <yuri@FreeBSD.org>
CommitDate: 2024-06-14 09:11:39 +0000
misc/llama-cpp: update 3135 → 3145
Reported by: portscout
---
misc/llama-cpp/Makefile | 2 +-
misc/llama-cpp/distinfo | 6 ++---
misc/llama-cpp/pkg-plist | 67 ++++++++++++++++++++++++------------------------
3 files changed, 38 insertions(+), 37 deletions(-)
diff --git a/misc/llama-cpp/Makefile b/misc/llama-cpp/Makefile
index 5700d57653b5..ab01926ca833 100644
--- a/misc/llama-cpp/Makefile
+++ b/misc/llama-cpp/Makefile
@@ -1,6 +1,6 @@
PORTNAME= llama-cpp
DISTVERSIONPREFIX= b
-DISTVERSION= 3135
+DISTVERSION= 3145
CATEGORIES= misc # machine-learning
MAINTAINER= yuri@FreeBSD.org
diff --git a/misc/llama-cpp/distinfo b/misc/llama-cpp/distinfo
index bc8ebdb50bbf..8121e0b72d7e 100644
--- a/misc/llama-cpp/distinfo
+++ b/misc/llama-cpp/distinfo
@@ -1,5 +1,5 @@
-TIMESTAMP = 1718166366
-SHA256 (ggerganov-llama.cpp-b3135_GH0.tar.gz) = 4e5e5284177bf261545899c986df2f195e7703ee1647589a26ccf2e1389b6218
-SIZE (ggerganov-llama.cpp-b3135_GH0.tar.gz) = 20552678
+TIMESTAMP = 1718344488
+SHA256 (ggerganov-llama.cpp-b3145_GH0.tar.gz) = 8fbd54ce330227b8d6fe157e114dfa42d4b11da9d70924840365bce91fe88670
+SIZE (ggerganov-llama.cpp-b3145_GH0.tar.gz) = 20555853
SHA256 (nomic-ai-kompute-4565194_GH0.tar.gz) = 95b52d2f0514c5201c7838348a9c3c9e60902ea3c6c9aa862193a212150b2bfc
SIZE (nomic-ai-kompute-4565194_GH0.tar.gz) = 13540496
diff --git a/misc/llama-cpp/pkg-plist b/misc/llama-cpp/pkg-plist
index 0b5bc0e84cf6..1851cb1e6ee2 100644
--- a/misc/llama-cpp/pkg-plist
+++ b/misc/llama-cpp/pkg-plist
@@ -1,38 +1,39 @@
-%%EXAMPLES%%bin/baby-llama
-%%EXAMPLES%%bin/batched
-%%EXAMPLES%%bin/batched-bench
-%%EXAMPLES%%bin/benchmark
%%EXAMPLES%%bin/convert-hf-to-gguf.py
-%%EXAMPLES%%bin/convert-llama2c-to-ggml
-%%EXAMPLES%%bin/embedding
-%%EXAMPLES%%bin/eval-callback
-%%EXAMPLES%%bin/export-lora
-%%EXAMPLES%%bin/finetune
-%%EXAMPLES%%bin/gguf
-%%EXAMPLES%%bin/gguf-split
-%%EXAMPLES%%bin/gritlm
-%%EXAMPLES%%bin/imatrix
-%%EXAMPLES%%bin/infill
+%%EXAMPLES%%bin/llama-baby-llama
+%%EXAMPLES%%bin/llama-batched
+%%EXAMPLES%%bin/llama-batched-bench
%%EXAMPLES%%bin/llama-bench
-%%EXAMPLES%%bin/llava-cli
-%%EXAMPLES%%bin/lookahead
-%%EXAMPLES%%bin/lookup
-%%EXAMPLES%%bin/lookup-create
-%%EXAMPLES%%bin/lookup-merge
-%%EXAMPLES%%bin/lookup-stats
-%%EXAMPLES%%bin/main
-%%EXAMPLES%%bin/parallel
-%%EXAMPLES%%bin/passkey
-%%EXAMPLES%%bin/perplexity
-%%EXAMPLES%%bin/quantize
-%%EXAMPLES%%bin/quantize-stats
-%%EXAMPLES%%bin/retrieval
-%%EXAMPLES%%bin/save-load-state
-%%EXAMPLES%%bin/server
-%%EXAMPLES%%bin/simple
-%%EXAMPLES%%bin/speculative
-%%EXAMPLES%%bin/tokenize
-%%EXAMPLES%%bin/train-text-from-scratch
+%%EXAMPLES%%bin/llama-bench-matmult
+%%EXAMPLES%%bin/llama-cli
+%%EXAMPLES%%bin/llama-convert-llama2c-to-ggml
+%%EXAMPLES%%bin/llama-embedding
+%%EXAMPLES%%bin/llama-eval-callback
+%%EXAMPLES%%bin/llama-export-lora
+%%EXAMPLES%%bin/llama-finetune
+%%EXAMPLES%%bin/llama-gbnf-validator
+%%EXAMPLES%%bin/llama-gguf
+%%EXAMPLES%%bin/llama-gguf-split
+%%EXAMPLES%%bin/llama-gritlm
+%%EXAMPLES%%bin/llama-imatrix
+%%EXAMPLES%%bin/llama-infill
+%%EXAMPLES%%bin/llama-llava-cli
+%%EXAMPLES%%bin/llama-lookahead
+%%EXAMPLES%%bin/llama-lookup
+%%EXAMPLES%%bin/llama-lookup-create
+%%EXAMPLES%%bin/llama-lookup-merge
+%%EXAMPLES%%bin/llama-lookup-stats
+%%EXAMPLES%%bin/llama-parallel
+%%EXAMPLES%%bin/llama-passkey
+%%EXAMPLES%%bin/llama-perplexity
+%%EXAMPLES%%bin/llama-quantize
+%%EXAMPLES%%bin/llama-quantize-stats
+%%EXAMPLES%%bin/llama-retrieval
+%%EXAMPLES%%bin/llama-save-load-state
+%%EXAMPLES%%bin/llama-server
+%%EXAMPLES%%bin/llama-simple
+%%EXAMPLES%%bin/llama-speculative
+%%EXAMPLES%%bin/llama-tokenize
+%%EXAMPLES%%bin/llama-train-text-from-scratch
include/ggml-alloc.h
include/ggml-backend.h
include/ggml.h