Not working currently due to lack of python 3.13 support in vllm-flash-attention. Try python-vllm-bin or the cpu version python-vllm
Search Criteria
Package Details: python-vllm-cuda 0.6.5-2
Package Actions
Git Clone URL: | https://aur.archlinux.org/python-vllm-cuda.git (read-only, click to copy) |
---|---|
Package Base: | python-vllm-cuda |
Description: | high-throughput and memory-efficient inference and serving engine for LLMs |
Upstream URL: | https://github.com/vllm-project/vllm |
Licenses: | Apache-2.0 |
Conflicts: | python-vllm |
Provides: | python-vllm |
Submitter: | envolution |
Maintainer: | envolution |
Last Packager: | envolution |
Votes: | 0 |
Popularity: | 0.000000 |
First Submitted: | 2024-12-01 16:12 (UTC) |
Last Updated: | 2025-01-01 17:16 (UTC) |
Dependencies (7)
- python (python37AUR, python311AUR, python310AUR)
- python-installer
- python-pytorch (python-pytorch-cxx11abiAUR, python-pytorch-cxx11abi-optAUR, python-pytorch-cxx11abi-cudaAUR, python-pytorch-cxx11abi-opt-cudaAUR, python-pytorch-cxx11abi-rocmAUR, python-pytorch-cxx11abi-opt-rocmAUR, python-pytorch-rocm-binAUR, python-pytorch-cuda, python-pytorch-opt, python-pytorch-opt-cuda, python-pytorch-opt-rocm, python-pytorch-rocm)
- cuda (cuda11.1AUR, cuda-12.2AUR, cuda12.0AUR, cuda11.4AUR, cuda11.4-versionedAUR, cuda12.0-versionedAUR) (make)
- cuda-tools (cuda11.1-toolsAUR, cuda12.0-toolsAUR, cuda11.4-toolsAUR, cuda11.4-versioned-toolsAUR, cuda12.0-versioned-toolsAUR) (make)
- gcc13 (make)
- git (git-gitAUR, git-glAUR) (make)
Pinned Comments
envolution commented on 2024-12-28 04:47 (UTC) (edited on 2024-12-28 04:51 (UTC) by envolution)
Not working currently due to lack of python 3.13 support in vllm-flash-attention. Try python-vllm-bin or the cpu version python-vllm