Simple scaffolding, testbed, and API endpoints for building, testing, and deploying LLM chains.
Supports local files, as well as those passed in directly or via API calls.
** consider experimental and unstable **
For fastest setup, start the devcontainer in VS code.
- Install dependencies via poetry.
- Add a
.env
file with your api keys (wandb, openai, etc) - Use datasets_sample as a template for the hierarchy of local files for use with chains. Best numbered and laid out in the format shown in
datasets_sample
- Set a port in
serve.py
, open the port (using ngrok, for example), and runserve.py
. - LLMitless uses FastAPI: once running, you can find documentation and make test calls by visiting /docs
- Tests run automatically on pushes to main
- The manual Deploy action will dockerize the app, store the image in Google Artifact Registry, and deploy from there to Cloud Run.
parsers/
contains modules that parse formats for use with LLMs
services/summarize.py
collects docs, prepares them for use in a chain, and calls the LLM
chains/
contains core langauge processing chains
serve.py
Fast API endpoints