We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent f2c7af6 commit a7792baCopy full SHA for a7792ba
packages/llama-cpp/build.sh
@@ -2,9 +2,9 @@ TERMUX_PKG_HOMEPAGE=https://github.com/ggml-org/llama.cpp
2
TERMUX_PKG_DESCRIPTION="LLM inference in C/C++"
3
TERMUX_PKG_LICENSE="MIT"
4
TERMUX_PKG_MAINTAINER=@termux
5
-TERMUX_PKG_VERSION="0.0.0-b6989"
+TERMUX_PKG_VERSION="0.0.0-b6992"
6
TERMUX_PKG_SRCURL=https://github.com/ggml-org/llama.cpp/archive/refs/tags/${TERMUX_PKG_VERSION#*-}.tar.gz
7
-TERMUX_PKG_SHA256=0112c067c11ea122089b0b0e85dfd163083f7ec6ad572fc4374268eac3e92fbc
+TERMUX_PKG_SHA256=65c4cf114c838d5849d91c7d00507561bccca0dace5d8d6f1dc841c9046a8483
8
TERMUX_PKG_AUTO_UPDATE=true
9
TERMUX_PKG_DEPENDS="libc++, libcurl"
10
TERMUX_PKG_BUILD_DEPENDS="vulkan-headers, opencl-headers, ocl-icd"
0 commit comments