# Generated by makepkg 7.1.0
# using fakeroot version 1.37.2
pkgname = llama.cpp
pkgbase = llama.cpp
xdata = pkgtype=pkg
pkgver = b8185-1
pkgdesc = Port of Facebook's LLaMA model in C/C++ (with system ggml support)
url = https://github.com/ggerganov/llama.cpp
builddate = 1772455808
packager = lilac (on behalf of 依云) <lilydjwg@gmail.com>
size = 32464797
arch = x86_64
license = MIT
depend = curl
depend = gcc-libs
depend = python
depend = ggml
optdepend = python-numpy: needed for convert_hf_to_gguf.py
optdepend = python-safetensors: needed for convert_hf_to_gguf.py
optdepend = python-sentencepiece: needed for convert_hf_to_gguf.py
optdepend = python-pytorch: needed for convert_hf_to_gguf.py
optdepend = python-transformers: needed for convert_hf_to_gguf.py
makedepend = cmake
makedepend = git
makedepend = vulkan-radeon
