git: 2afd8305a7c3 - main - misc/llama-cpp: update 3452 → 3465

From: Yuri Victorovich <yuri_at_FreeBSD.org>
Date: Fri, 26 Jul 2024 08:58:11 UTC
The branch main has been updated by yuri:

URL: https://cgit.FreeBSD.org/ports/commit/?id=2afd8305a7c323e00db1814af1a37dfc81857509

commit 2afd8305a7c323e00db1814af1a37dfc81857509
Author:     Yuri Victorovich <yuri@FreeBSD.org>
AuthorDate: 2024-07-26 08:44:36 +0000
Commit:     Yuri Victorovich <yuri@FreeBSD.org>
CommitDate: 2024-07-26 08:57:59 +0000

    misc/llama-cpp: update 3452 → 3465
    
    Reported by:    portscout
---
 misc/llama-cpp/Makefile  |  2 +-
 misc/llama-cpp/distinfo  |  6 +++---
 misc/llama-cpp/pkg-plist | 22 ++++++++++------------
 3 files changed, 14 insertions(+), 16 deletions(-)

diff --git a/misc/llama-cpp/Makefile b/misc/llama-cpp/Makefile
index 002d765add3d..0f41c02e6645 100644
--- a/misc/llama-cpp/Makefile
+++ b/misc/llama-cpp/Makefile
@@ -1,6 +1,6 @@
 PORTNAME=	llama-cpp
 DISTVERSIONPREFIX=	b
-DISTVERSION=	3452
+DISTVERSION=	3465
 CATEGORIES=	misc # machine-learning
 
 MAINTAINER=	yuri@FreeBSD.org
diff --git a/misc/llama-cpp/distinfo b/misc/llama-cpp/distinfo
index f280ce7f3f42..15d42ee9b5ac 100644
--- a/misc/llama-cpp/distinfo
+++ b/misc/llama-cpp/distinfo
@@ -1,5 +1,5 @@
-TIMESTAMP = 1721889252
-SHA256 (ggerganov-llama.cpp-b3452_GH0.tar.gz) = 552b43db840cd8d501e1940560098bc8d41ccfb3334130b9ca64253e288ba664
-SIZE (ggerganov-llama.cpp-b3452_GH0.tar.gz) = 19034057
+TIMESTAMP = 1721973886
+SHA256 (ggerganov-llama.cpp-b3465_GH0.tar.gz) = 2ccc493ce196c8cecaf97dcb15238ce5396c4ae0380c5b7e0c18a9c9f020b2fd
+SIZE (ggerganov-llama.cpp-b3465_GH0.tar.gz) = 19001687
 SHA256 (nomic-ai-kompute-4565194_GH0.tar.gz) = 95b52d2f0514c5201c7838348a9c3c9e60902ea3c6c9aa862193a212150b2bfc
 SIZE (nomic-ai-kompute-4565194_GH0.tar.gz) = 13540496
diff --git a/misc/llama-cpp/pkg-plist b/misc/llama-cpp/pkg-plist
index c386202d9b63..7f6d4ac506e6 100644
--- a/misc/llama-cpp/pkg-plist
+++ b/misc/llama-cpp/pkg-plist
@@ -1,16 +1,15 @@
-%%EXAMPLES%%bin/convert_hf_to_gguf.py
-%%EXAMPLES%%bin/llama-baby-llama
-%%EXAMPLES%%bin/llama-batched
-%%EXAMPLES%%bin/llama-batched-bench
-%%EXAMPLES%%bin/llama-bench
-%%EXAMPLES%%bin/llama-bench-matmult
-%%EXAMPLES%%bin/llama-cli
-%%EXAMPLES%%bin/llama-convert-llama2c-to-ggml
-%%EXAMPLES%%bin/llama-cvector-generator
-%%EXAMPLES%%bin/llama-embedding
+bin/convert_hf_to_gguf.py
+bin/llama-baby-llama
+bin/llama-batched
+bin/llama-batched-bench
+bin/llama-bench
+bin/llama-bench-matmult
+bin/llama-cli
+bin/llama-convert-llama2c-to-ggml
+bin/llama-cvector-generator
+bin/llama-embedding
 %%EXAMPLES%%bin/llama-eval-callback
 %%EXAMPLES%%bin/llama-export-lora
-%%EXAMPLES%%bin/llama-finetune
 %%EXAMPLES%%bin/llama-gbnf-validator
 %%EXAMPLES%%bin/llama-gguf
 %%EXAMPLES%%bin/llama-gguf-hash
@@ -35,7 +34,6 @@
 %%EXAMPLES%%bin/llama-simple
 %%EXAMPLES%%bin/llama-speculative
 %%EXAMPLES%%bin/llama-tokenize
-%%EXAMPLES%%bin/llama-train-text-from-scratch
 include/ggml-alloc.h
 include/ggml-backend.h
 include/ggml-blas.h