Welcome to our Model Repository! This repository houses a collection of machine learning models designed to perform various AI tasks. They are all prepared and ready to be seamlessly served on Instill Core or Instill Cloud via our MLOps/LLMOps platform Instill Model.
We have a diverse set of models, each optimized for different AI tasks. Please refer to the table below to gain more insight into a specific model, including its configuration, implementation details, and usage. Feel free to check out the README files in the respective model folders:
Model Name | Task Type | Description |
---|---|---|
phi-3.5-vision-instruct | Chat | Phi-3.5-vision is a lightweight, state-of-the-art open multimodal model. |
gte-Qwen2-1.5B-instruct | Embedding | gte-Qwen2-1.5B-instruct is the latest model in the gte (General Text Embedding) model family. |
jina-clip-v1 | Embedding | jina-clip-v1 is a state-of-the-art English multimodal (text-image) embedding model. |
llama-2-7b-chat | Chat | llama-2-7b-chat is optimized for dialogue use cases. |
llama-3-8b-instruct | Chat | llama-3-8b-instruct is an instruction tuned generative text model. |
llamacode-7b | Completion | llamacode-7b is designed for general code synthesis and understanding. |
llava-1-6-13b | Chat | llava-1-6-13b is an open-source chatbot trained by fine-tuning LLM on multimodal instruction-following data. |
mobilenetv2 | Classification | mobilenetv2 is a lightweight 53-layer deep CNN model with a smaller number of parameters and an input size of 224×224. |
stable-diffusion-xl | Text to Image | stable-diffusion-xl is a a latent diffusion model for text-to-image synthesis. |
stella-en-1.5B-v5 | Embedding | stella-en-1.5B-v5 is trained based on Alibaba-NLP/gte-large-en-v1.5 and Alibaba-NLP/gte-Qwen2-1.5B-instruct. |
tinyllama | Chat | tinyllama is a chat model finetuned on top of TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T. |
yolov7 | Object Detection | yolov7 is a state-of-the-art real-time object detector. |
yolov7-stomata | Instance Segmentation | yolov7-stomata is designed for stomata detection and segmentation. |
zephyr-7b | Chat | zephyr-7b is a series of language models that are trained to act as helpful assistants. |
gemma-2-27b-it | Chat | Gemma is a family of lightweight, state-of-the-art open models from Google. |
qwen-2-5-72b-instruct | Chat | Qwen2.5 is the latest series of Qwen large language models. |
qwen-2-5-coder-32b-instructt | Chat | Qwen2.5-Coder is the latest series of Code-Specific Qwen large language models (formerly known as CodeQwen). |
llama-3-2-3b-instruct | Chat | The Llama 3.2 collection of multilingual large language models (LLMs) is a collection of pretrained and instruction-tuned generative models in 1B and 3B sizes (text in/text out). |
llama-3-2-11b-vision-instruct | Chat | The Llama 3.2-Vision collection of multimodal large language models (LLMs) is a collection of pretrained and instruction-tuned image reasoning generative models in 11B and 90B sizes (text + images in / text out). |
llama-3-2-90b-vision-instruct | Chat | The Llama 3.2-Vision collection of multimodal large language models (LLMs) is a collection of pretrained and instruction-tuned image reasoning generative models in 11B and 90B sizes (text + images in / text out). |
functionary-medium-v3.2 | Chat | Functionary is a language model that can interpret and execute functions/plugins. |
yi-vl-6b | Chat | Yi Vision Language (Yi-VL) model is the open-source, multimodal version of the Yi Large Language Model (LLM) series, enabling content comprehension, recognition, and multi-round conversations about images. |
stable-diffusion-3.5-large | Text to Image | Stable Diffusion 3.5 Large is a Multimodal Diffusion Transformer (MMDiT) text-to-image model. |
qwen-2-vl-72b-instruct | Chat | Qwen2-VL, the latest iteration of our Qwen-VL model, representing nearly a year of innovation. |
llama-3-3-70b-instruct | Chat | The Meta Llama 3.3 multilingual large language model (LLM) is a pretrained and instruction tuned generative model in 70B (text in/text out). |
We leverage Instill Core to provide a seamless experience for serving models. Follow the steps below to quickly get started:
Follow this section of our quick start guide to get it up and running with self-hosting Instill Core on a local or remote instance.
Follow this section of our quick start guide to get it up and running on Instill Cloud, our fully managed public cloud service that provides you with access to all the features of Instill Core without the burden of infrastructure management.
To create a model namespace, follow the steps on the Create Namespace page.
Find the model you want to serve and download the desired version folder. Also, make sure to check out the particular model folder README to obtain other necessary files, model weights or perform any additional required steps.
Follow the steps on the Build Model Image page, and remember to install the python-sdk
version according to the compatibility matrix in each model's README.
Follow the steps on the Push Model Image page to deploy the model to your choice of Instill Core or Instill Cloud.
Follow the steps on the Prepare Model page to see how to implement your own custom model that can be served on Instill Core and Instill Cloud! You can also checkout the step-by-step tutorial which walks you through the process of serving your own custom model on Instill Core.
We welcome contributions! Please see our CONTRIBUTING.md file for more details on how to get started.
If you encounter any issues, please check our Documentation or open an issue on GitHub.
This project is licensed under the MIT License - see the LICENSE file for details.