Package Details: python-vllm-cuda 0.6.5-2

Git Clone URL: https://aur.archlinux.org/python-vllm-cuda.git (read-only, click to copy)
Package Base: python-vllm-cuda
Description: high-throughput and memory-efficient inference and serving engine for LLMs
Upstream URL: https://github.com/vllm-project/vllm
Licenses: Apache-2.0
Conflicts: python-vllm
Provides: python-vllm
Submitter: envolution
Maintainer: envolution
Last Packager: envolution
Votes: 0
Popularity: 0.000000
First Submitted: 2024-12-01 16:12 (UTC)
Last Updated: 2025-01-01 17:16 (UTC)

Pinned Comments

envolution commented on 2024-12-28 04:47 (UTC) (edited on 2024-12-28 04:51 (UTC) by envolution)

Not working currently due to lack of python 3.13 support in vllm-flash-attention. Try python-vllm-bin or the cpu version python-vllm

Latest Comments

envolution commented on 2024-12-28 04:47 (UTC) (edited on 2024-12-28 04:51 (UTC) by envolution)

Not working currently due to lack of python 3.13 support in vllm-flash-attention. Try python-vllm-bin or the cpu version python-vllm