# Generated by abuild 3.16.0-r0
# using fakeroot version 1.37.2
pkgname = llama.cpp
pkgver = 0.0.8697-r0
pkgdesc = LLM inference in C/C++ (with Vulkan GPU acceleration)
url = https://github.com/ggml-org/llama.cpp
builddate = 1775614197
packager = Buildozer <alpine-devel@lists.alpinelinux.org>
size = 35448278
arch = aarch64
origin = llama.cpp
commit = e1346e394a7f97646fd3b1591e4d055902deee86
maintainer = Hugo Osvaldo Barrera <hugo@whynothugo.nl>
license = MIT
depend = /bin/sh
# automatically detected:
provides = cmd:export-graph-ops=0.0.8697-r0
provides = cmd:llama-bench=0.0.8697-r0
provides = cmd:llama-cli=0.0.8697-r0
provides = cmd:llama-debug-template-parser=0.0.8697-r0
provides = cmd:llama-results=0.0.8697-r0
provides = cmd:llama-server=0.0.8697-r0
provides = cmd:llama-simple-chat=0.0.8697-r0
provides = cmd:llama-template-analysis=0.0.8697-r0
provides = cmd:llama-tts=0.0.8697-r0
provides = so:llama.cpp:libggml-base.so.0=0.9.11
provides = so:llama.cpp:libggml-blas.so=0
provides = so:llama.cpp:libggml-rpc.so=0
provides = so:llama.cpp:libggml.so.0=0.9.11
provides = so:llama.cpp:libllama.so.0=0.0.8697
provides = so:llama.cpp:libmtmd.so.0=0.0.8697
depend = openblas
depend = so:libc.musl-aarch64.so.1
depend = so:libcrypto.so.3
depend = so:libgcc_s.so.1
depend = so:libssl.so.3
depend = so:libstdc++.so.6
datahash = 6d53a472d689075000868e2bf424124985025a60586f035fec1907e3713da50a
