-
Notifications
You must be signed in to change notification settings - Fork 30
Issues: intel/llm-on-ray
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Calculate correct input length for every prompt in a single batch
#222
opened May 14, 2024 by
kira-lin
Openai API not allow temperature=0.0 for llama-2-7b-chat-hf
#139
opened Mar 12, 2024 by
yutianchen666
Add ipex extra in pyproject.toml to use restricted transformers version
#127
opened Feb 29, 2024 by
jiafuzha
Support functions/tools in OpenAI API
enhancement
New feature or request
#121
opened Feb 23, 2024 by
carsonwang
Support and validate model Mixtral-8x7B
enhancement
New feature or request
#119
opened Feb 23, 2024 by
carsonwang
[Quantization] Support loading AWQ, GPTQ, GGUF/GGML quantized models
#85
opened Jan 26, 2024 by
xwu99
[Benchmark] Load config from yaml and output results with multiple formats
#82
opened Jan 24, 2024 by
xwu99
Getting error while executing query_openai_sdk.py to test the inference
#66
opened Jan 18, 2024 by
dkiran1
Not able to run inference server for mistral 7b model, mpt-7b model on Ray
#65
opened Jan 18, 2024 by
dkiran1
[Serving] Add a table of models and corresponding supported parameters
#51
opened Jan 11, 2024 by
KepingYan
Previous Next
ProTip!
What’s not been updated in a month: updated:<2024-10-05.