Search Criteria
Package Details: llama.cpp-cuda-f16 b4304-1
Package Actions
Git Clone URL: | https://aur.archlinux.org/llama.cpp-cuda-f16.git (read-only, click to copy) |
---|---|
Package Base: | llama.cpp-cuda-f16 |
Description: | Port of Facebook's LLaMA model in C/C++ (with NVIDIA CUDA optimizations and f16) |
Upstream URL: | https://github.com/ggerganov/llama.cpp |
Licenses: | MIT |
Conflicts: | libggml, llama.cpp |
Provides: | llama.cpp |
Submitter: | txtsd |
Maintainer: | txtsd |
Last Packager: | txtsd |
Votes: | 1 |
Popularity: | 0.90 |
First Submitted: | 2024-12-06 13:35 (UTC) |
Last Updated: | 2024-12-11 14:48 (UTC) |
Dependencies (12)
- blas-openblas
- blas64-openblas
- cuda (cuda11.1AUR, cuda-12.2AUR, cuda12.0AUR, cuda11.4AUR, cuda11.4-versionedAUR, cuda12.0-versionedAUR)
- curl (curl-quiche-gitAUR, curl-http3-ngtcp2AUR, curl-gitAUR, curl-c-aresAUR)
- gcc-libs (gcc-libs-gitAUR, gccrs-libs-gitAUR, gcc11-libsAUR, gcc-libs-snapshotAUR)
- glibc (glibc-gitAUR, glibc-linux4AUR, glibc-eacAUR, glibc-eac-binAUR, glibc-eac-rocoAUR)
- openmp
- python (python37AUR, python311AUR, python310AUR)
- python-numpy (python-numpy-flameAUR, python-numpy-gitAUR, python-numpy1AUR, python-numpy-mkl-binAUR, python-numpy-mkl-tbbAUR, python-numpy-mklAUR)
- python-sentencepieceAUR (python-sentencepiece-gitAUR)
- cmake (cmake-gitAUR) (make)
- git (git-gitAUR, git-glAUR) (make)
Latest Comments
txtsd commented on 2024-12-06 14:15 (UTC)
Alternate versions
llama.cpp
llama.cpp-vulkan
llama.cpp-sycl-fp16
llama.cpp-sycl-fp32
llama.cpp-cuda
llama.cpp-cuda-f16
llama.cpp-hip