Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

feat: support vllm in controller #635

Open
wants to merge 2 commits into
base: main
Choose a base branch
from
Open

Conversation

zhuangqh
Copy link
Collaborator

@zhuangqh zhuangqh commented Oct 17, 2024

Reason for Change:

  • support vllm runtime deployments
  • add a feature gate setting vllm as the default runtime
  • able to select runtime by annotation tag

Requirements

  • added unit tests and e2e tests (if applicable).

@@ -5,31 +5,39 @@ ARG MODEL_TYPE
ARG VERSION
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

.
├── chat_templates
│   ├── alpaca.jinja
│   ├── amberchat.jinja
│   ├── chatml.jinja
│   ├── chatqa.jinja
│   ├── falcon-instruct.jinja
│   ├── gemma-it.jinja
│   ├── llama-2-chat.jinja
│   ├── llama-3-instruct.jinja
│   ├── mistral-instruct.jinja
│   ├── openchat-3.5.jinja
│   ├── phi-3-small.jinja
│   ├── phi-3.jinja
│   ├── saiga.jinja
│   ├── solar-instruct.jinja
│   ├── vicuna.jinja
│   └── zephyr.jinja
├── tfs
│   ├── cli.py
│   ├── dataset.py
│   ├── fine_tuning.py
│   ├── inference-requirements.txt
│   ├── inference_api.py
│   ├── metrics_server.py
│   ├── parser.py
│   ├── tuning-requirements.txt
│   └── weights -> /workspace/weights
├── version.txt
├── vllm
│   ├── inference-requirements.txt
│   ├── inference_api.py
│   └── weights -> /workspace/weights
└── weights

@zhuangqh zhuangqh marked this pull request as ready for review October 28, 2024 11:13
ModelRunParams map[string]string // Parameters for running the model training/inference.
}

func (p *PresetParam) DeepCopy() *PresetParam {
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

we may update some params according to the node counts. Thus, we must deepcopy it at first.

- set vllm as the default runtime

Signed-off-by: jerryzhuang <[email protected]>
Signed-off-by: jerryzhuang <[email protected]>
TorchRunRdzvParams map[string]string // Optional rendezvous parameters for distributed training/inference using torchrun (elastic).
BaseCommand string // The initial command (e.g., 'torchrun', 'accelerate launch') used in the command line.
ModelRunParams map[string]string // Parameters for running the model training/inference.
Tag string // The model image tag
Copy link
Collaborator

@ishaansehgal99 ishaansehgal99 Nov 14, 2024

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

does this tag field get used

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants