git: 2afd8305a7c3 - main - misc/llama-cpp: update 3452 → 3465
- Go to: [ bottom of page ] [ top of archives ] [ this month ]
Date: Fri, 26 Jul 2024 08:58:11 UTC
The branch main has been updated by yuri: URL: https://cgit.FreeBSD.org/ports/commit/?id=2afd8305a7c323e00db1814af1a37dfc81857509 commit 2afd8305a7c323e00db1814af1a37dfc81857509 Author: Yuri Victorovich <yuri@FreeBSD.org> AuthorDate: 2024-07-26 08:44:36 +0000 Commit: Yuri Victorovich <yuri@FreeBSD.org> CommitDate: 2024-07-26 08:57:59 +0000 misc/llama-cpp: update 3452 → 3465 Reported by: portscout --- misc/llama-cpp/Makefile | 2 +- misc/llama-cpp/distinfo | 6 +++--- misc/llama-cpp/pkg-plist | 22 ++++++++++------------ 3 files changed, 14 insertions(+), 16 deletions(-) diff --git a/misc/llama-cpp/Makefile b/misc/llama-cpp/Makefile index 002d765add3d..0f41c02e6645 100644 --- a/misc/llama-cpp/Makefile +++ b/misc/llama-cpp/Makefile @@ -1,6 +1,6 @@ PORTNAME= llama-cpp DISTVERSIONPREFIX= b -DISTVERSION= 3452 +DISTVERSION= 3465 CATEGORIES= misc # machine-learning MAINTAINER= yuri@FreeBSD.org diff --git a/misc/llama-cpp/distinfo b/misc/llama-cpp/distinfo index f280ce7f3f42..15d42ee9b5ac 100644 --- a/misc/llama-cpp/distinfo +++ b/misc/llama-cpp/distinfo @@ -1,5 +1,5 @@ -TIMESTAMP = 1721889252 -SHA256 (ggerganov-llama.cpp-b3452_GH0.tar.gz) = 552b43db840cd8d501e1940560098bc8d41ccfb3334130b9ca64253e288ba664 -SIZE (ggerganov-llama.cpp-b3452_GH0.tar.gz) = 19034057 +TIMESTAMP = 1721973886 +SHA256 (ggerganov-llama.cpp-b3465_GH0.tar.gz) = 2ccc493ce196c8cecaf97dcb15238ce5396c4ae0380c5b7e0c18a9c9f020b2fd +SIZE (ggerganov-llama.cpp-b3465_GH0.tar.gz) = 19001687 SHA256 (nomic-ai-kompute-4565194_GH0.tar.gz) = 95b52d2f0514c5201c7838348a9c3c9e60902ea3c6c9aa862193a212150b2bfc SIZE (nomic-ai-kompute-4565194_GH0.tar.gz) = 13540496 diff --git a/misc/llama-cpp/pkg-plist b/misc/llama-cpp/pkg-plist index c386202d9b63..7f6d4ac506e6 100644 --- a/misc/llama-cpp/pkg-plist +++ b/misc/llama-cpp/pkg-plist @@ -1,16 +1,15 @@ -%%EXAMPLES%%bin/convert_hf_to_gguf.py -%%EXAMPLES%%bin/llama-baby-llama -%%EXAMPLES%%bin/llama-batched -%%EXAMPLES%%bin/llama-batched-bench -%%EXAMPLES%%bin/llama-bench -%%EXAMPLES%%bin/llama-bench-matmult -%%EXAMPLES%%bin/llama-cli -%%EXAMPLES%%bin/llama-convert-llama2c-to-ggml -%%EXAMPLES%%bin/llama-cvector-generator -%%EXAMPLES%%bin/llama-embedding +bin/convert_hf_to_gguf.py +bin/llama-baby-llama +bin/llama-batched +bin/llama-batched-bench +bin/llama-bench +bin/llama-bench-matmult +bin/llama-cli +bin/llama-convert-llama2c-to-ggml +bin/llama-cvector-generator +bin/llama-embedding %%EXAMPLES%%bin/llama-eval-callback %%EXAMPLES%%bin/llama-export-lora -%%EXAMPLES%%bin/llama-finetune %%EXAMPLES%%bin/llama-gbnf-validator %%EXAMPLES%%bin/llama-gguf %%EXAMPLES%%bin/llama-gguf-hash @@ -35,7 +34,6 @@ %%EXAMPLES%%bin/llama-simple %%EXAMPLES%%bin/llama-speculative %%EXAMPLES%%bin/llama-tokenize -%%EXAMPLES%%bin/llama-train-text-from-scratch include/ggml-alloc.h include/ggml-backend.h include/ggml-blas.h