Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

recipe for vllm #24710

Open
wants to merge 67 commits into
base: main
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
67 commits
Select commit Hold shift + click to select a range
67a8dc7
recipe for vllm
mediocretech Dec 7, 2023
a3e6578
fixed ray conda package name ref
mediocretech Dec 21, 2023
de9e058
moved to ray-core
mediocretech Dec 21, 2023
ba1d0dd
Update build_steps.sh
mediocretech Jan 19, 2024
f62f10c
GPU requirement
mediocretech Jan 19, 2024
68f4f5f
Build compiler deps
mediocretech Jan 19, 2024
d27b5eb
setting ninja max jobs in attempt to stop build cycling
mediocretech Jan 19, 2024
4f19559
added overlink requirements
mediocretech Jan 19, 2024
911519a
libtorch_python overlink
mediocretech Jan 19, 2024
b932232
latest source version
mediocretech Jan 19, 2024
e68a28c
sha mismatch
mediocretech Jan 19, 2024
8c0a076
build test
mediocretech Jan 19, 2024
b5ae846
lint
mediocretech Jan 19, 2024
2aeb452
debugging; trying remove build for yaml formating issue
mediocretech Jan 25, 2024
942f98f
simplified build
mediocretech Jan 25, 2024
acaf85c
Update build_steps.sh
mediocretech Jan 25, 2024
a2d33d7
lowered build version due to pytorch version requirement not in conda…
mediocretech Jan 25, 2024
f9d9e2d
removed pip check due to conda-forge version differences
mediocretech Jan 25, 2024
000f766
returned pip check, runaway build
mediocretech Jan 25, 2024
c2d4774
Merge branch 'main' into vllm
mediocretech Feb 3, 2024
bd0f153
No test for pip check
mediocretech Mar 1, 2024
4307b81
Merge branch 'main' into vllm
mediocretech Mar 1, 2024
2b0bce6
updated dependancy list
mediocretech Mar 1, 2024
6e10d68
the more, the merrier
mediocretech Mar 1, 2024
bfba786
Merge branch 'main' into vllm
mediocretech Mar 1, 2024
a5ca474
no noarch
mediocretech Mar 1, 2024
04064e1
Merge branch 'vllm' of github.com:mediocretech/staged-recipes into vllm
mediocretech Mar 1, 2024
dc76ebb
torch is needed for build
mediocretech Mar 1, 2024
33b4b2f
add vllm
iamthebot Mar 4, 2024
d5ec5f2
Merge branch 'vllm' into al--vllm
iamthebot Mar 4, 2024
35b4ff9
Merge branch 'main' into vllm
mediocretech Mar 4, 2024
388a4fa
Merge branch 'vllm' into al--vllm
mediocretech Mar 4, 2024
ef0472c
Merge pull request #1 from iamthebot/al--vllm
mediocretech Mar 4, 2024
43e2dbc
whitelist libcuda
mediocretech Mar 6, 2024
4f5d04b
Merge branch 'main' into vllm
mediocretech Mar 6, 2024
a6b1e5f
force build
mediocretech Mar 7, 2024
593f989
Merge branch 'main' into vllm
mediocretech Mar 21, 2024
42b1ba3
Merge branch 'main' into vllm
mediocretech Mar 28, 2024
671910b
Merge branch 'main' into vllm
mediocretech Apr 4, 2024
ae258f4
worth a shot
mediocretech Apr 4, 2024
7a6e5e5
Merge branch 'main' into vllm
mediocretech Apr 4, 2024
6e50aa8
Merge branch 'main' into vllm
mediocretech May 2, 2024
6435f40
Merge branch 'main' into vllm
mediocretech May 2, 2024
0a9a043
simplified build deps
mediocretech May 2, 2024
c41c3bf
libcusparse
mediocretech May 2, 2024
2742768
removed libcuseparse from build
mediocretech May 2, 2024
42350b4
libcusparse-dev
mediocretech May 2, 2024
27ea238
versioning for cusparse
mediocretech May 2, 2024
49a3b5e
touchvision dep
mediocretech May 2, 2024
b122ae9
version bump
mediocretech May 21, 2024
2101b09
updated sha for newer version
mediocretech May 21, 2024
eae0ea1
triton runtime req
mediocretech May 21, 2024
6b9c7fc
Merge branch 'main' into vllm
mediocretech Aug 1, 2024
8738492
latest conda-build is incompatible
mediocretech Aug 1, 2024
3a0dcdc
update to latest release
mediocretech Aug 1, 2024
2a25f9b
unpinned dependancies for latest vllm release
mediocretech Aug 1, 2024
a5a87d9
corrected sha for 0.5.3
mediocretech Aug 1, 2024
bf73d7e
Update requirements.txt
mediocretech Aug 1, 2024
7281d5f
Merge branch 'main' into vllm
mediocretech Aug 1, 2024
89e18d0
hope for the future...0.6.2
mediocretech Sep 26, 2024
2a8254d
hash for 0.6.2
mediocretech Sep 26, 2024
c5b616c
Merge branch 'main' into vllm
mediocretech Sep 26, 2024
ab8ff82
setuptools_scm
mediocretech Sep 26, 2024
c843952
changes to build
mediocretech Sep 26, 2024
a4f65c0
0.6.4 bump
mediocretech Dec 17, 2024
72aa925
hash update
mediocretech Dec 17, 2024
d8ab547
build is now 3.12
mediocretech Dec 17, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
8 changes: 8 additions & 0 deletions recipes/vllm/build.sh
Original file line number Diff line number Diff line change
@@ -0,0 +1,8 @@
#!/bin/bash -e
export MAX_JOBS=1
# needed to find libcrypt headers
export CPATH=${BUILD_PREFIX}/include
export VLLM_TARGET_DEVICE="cuda"

${PYTHON} use_existing_torch.py
${PYTHON} -m pip install . --no-deps -vv --no-deps --no-build-isolation
99 changes: 99 additions & 0 deletions recipes/vllm/meta.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,99 @@
{% set name = "vllm" %}
{% set version = "0.6.4" %}

package:
name: {{ name|lower }}
version: {{ version }}

source:
url: https://pypi.io/packages/source/{{ name[0] }}/{{ name }}/vllm-{{ version }}.tar.gz
sha256: 5d1452c44e6fde35084219a8396a79c77f1202a6fc69fdb966d15a1a8933b762

build:
# noarch: python

rpaths:
- lib/

script_env:
- TORCH_CUDA_ARCH_LIST=3.5;5.0;6.0;6.1;7.0;7.5;8.0;8.6+PTX # [cuda_compiler_version == "11.2"]
- TORCH_CUDA_ARCH_LIST=3.5;5.0;6.0;6.1;7.0;7.5;8.0;8.6;8.9+PTX # [cuda_compiler_version == "11.8"]
- TORCH_CUDA_ARCH_LIST=5.0;6.0;6.1;7.0;7.5;8.0;8.6;8.9;9.0+PTX # [(cuda_compiler_version or "").startswith("12")]

number: 0
skip: True # [cuda_compiler_version == "None"]
skip: True # [osx or win]
skip: True # [py < 312]
missing_dso_whitelist:
- '*/libtorch_python.so'
- '*/libcuda.so.1'

requirements:
build:
- cmake
- make
- {{ stdlib("c") }}
- {{ compiler('c') }}
- {{ compiler('cxx') }}
- {{ compiler('cuda') }}

host:
mediocretech marked this conversation as resolved.
Show resolved Hide resolved
- cuda-version {{ cuda_compiler_version }}
- python
mediocretech marked this conversation as resolved.
Show resolved Hide resolved
- ninja
- packaging
- setuptools
- setuptools-scm
- pytorch
- torchvision
- wheel
- pip
- libxcrypt

run:
- python
- ninja
- psutil
# - ray-core >=2.9.0
- pandas
- pyarrow
- sentencepiece
- numpy
- einops
- pytorch
- transformers >=4.34.0
- xformers
- fastapi
- uvicorn
- pydantic
- aioprometheus
- prometheus_client
- triton
# - libtorch
# - cuda-version {{ cuda_compiler_version }} # [cuda_compiler_version != "None"]
# - cuda-cupti-dev # [(cuda_compiler_version or "").startswith("12")]
# - cuda-cudart-dev # [(cuda_compiler_version or "").startswith("12")]
# - cuda-nvml-dev # [(cuda_compiler_version or "").startswith("12")]
# - cuda-nvtx-dev # [(cuda_compiler_version or "").startswith("12")]
# - libstdcxx-ng
# - libgcc-ng
# - libxcrypt

test:
imports:
- vllm
# commands:
# - pip check
requires:
- pip

about:
home: https://github.com/vllm-project/vllm
summary: A high-throughput and memory-efficient inference and serving engine for LLMs
license: Apache-2.0
license_file: LICENSE

extra:
recipe-maintainers:
- mediocretech
- iamthebot
Loading