From b1595e2a508fbb1db1e74c59f46f0fff93237590 Mon Sep 17 00:00:00 2001 From: ChihYu Yeh Date: Sat, 1 Feb 2025 15:43:03 +0800 Subject: [PATCH] update docs --- .../docs/config_examples/config.deepseek.yaml | 31 ++++++++++--------- .../config.google_ai_studio.yaml | 30 +++++++++--------- .../docs/config_examples/config.groq.yaml | 31 ++++++++++--------- .../docs/config_examples/config.ollama.yaml | 4 +-- 4 files changed, 49 insertions(+), 47 deletions(-) diff --git a/wren-ai-service/docs/config_examples/config.deepseek.yaml b/wren-ai-service/docs/config_examples/config.deepseek.yaml index 4e9844d2e..21bd5ba7e 100644 --- a/wren-ai-service/docs/config_examples/config.deepseek.yaml +++ b/wren-ai-service/docs/config_examples/config.deepseek.yaml @@ -30,12 +30,13 @@ models: --- type: embedder -provider: openai_embedder +provider: litellm_embedder models: - # find EMBEDDER_OPENAI_API_KEY and fill in value of api key in ~/.wrenai/.env - - model: text-embedding-3-large # put your openai compatible embedder model name here -url: https://api.openai.com/v1 # change this according to your openai compatible embedder model -timeout: 120 + # define OPENAI_API_KEY= in ~/.wrenai/.env if you are using openai embedding model + # please refer to LiteLLM documentation for more details: https://docs.litellm.ai/docs/providers + - model: text-embedding-3-large # put your embedding model name here, if it is not openai embedding model, should be / + api_base: https://api.openai.com/v1 # change this according to your embedding model + timeout: 120 --- type: engine @@ -57,20 +58,20 @@ recreate_index: false type: pipeline pipes: - name: db_schema_indexing - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: historical_question_indexing - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: table_description_indexing - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: db_schema_retrieval llm: litellm_llm.deepseek/deepseek-coder - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: historical_question_retrieval - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: sql_generation llm: litellm_llm.deepseek/deepseek-coder @@ -106,7 +107,7 @@ pipes: llm: litellm_llm.deepseek/deepseek-coder - name: question_recommendation_db_schema_retrieval llm: litellm_llm.deepseek/deepseek-coder - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: question_recommendation_sql_generation llm: litellm_llm.deepseek/deepseek-coder @@ -117,19 +118,19 @@ pipes: llm: litellm_llm.deepseek/deepseek-coder - name: intent_classification llm: litellm_llm.deepseek/deepseek-coder - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: data_assistance llm: litellm_llm.deepseek/deepseek-chat - name: sql_pairs_indexing document_store: qdrant - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large - name: sql_pairs_deletion document_store: qdrant - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large - name: sql_pairs_retrieval document_store: qdrant - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large llm: litellm_llm.deepseek/deepseek-coder - name: preprocess_sql_data llm: litellm_llm.deepseek/deepseek-coder diff --git a/wren-ai-service/docs/config_examples/config.google_ai_studio.yaml b/wren-ai-service/docs/config_examples/config.google_ai_studio.yaml index 5ab55e7d0..e2ecdbbbb 100644 --- a/wren-ai-service/docs/config_examples/config.google_ai_studio.yaml +++ b/wren-ai-service/docs/config_examples/config.google_ai_studio.yaml @@ -13,12 +13,12 @@ models: --- type: embedder -provider: openai_embedder +provider: litellm_embedder models: - # find EMBEDDER_OPENAI_API_KEY and fill in value of api key in ~/.wrenai/.env - - model: text-embedding-004 # put your openai compatible embedder model name here -url: https://generativelanguage.googleapis.com/v1beta/openai # change this according to your openai compatible embedder model -timeout: 120 + # put GEMINI_API_KEY= in ~/.wrenai/.env + - model: gemini/text-embedding-004 # gemini/ + api_base: https://generativelanguage.googleapis.com/v1beta/openai # change this according to your embedding model + timeout: 120 --- type: engine @@ -40,20 +40,20 @@ recreate_index: false type: pipeline pipes: - name: db_schema_indexing - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 document_store: qdrant - name: historical_question_indexing - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 document_store: qdrant - name: table_description_indexing - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 document_store: qdrant - name: db_schema_retrieval llm: litellm_llm.gemini/gemini-2.0-flash-exp - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 document_store: qdrant - name: historical_question_retrieval - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 document_store: qdrant - name: sql_generation llm: litellm_llm.gemini/gemini-2.0-flash-exp @@ -89,7 +89,7 @@ pipes: llm: litellm_llm.gemini/gemini-2.0-flash-exp - name: question_recommendation_db_schema_retrieval llm: litellm_llm.gemini/gemini-2.0-flash-exp - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 document_store: qdrant - name: question_recommendation_sql_generation llm: litellm_llm.gemini/gemini-2.0-flash-exp @@ -100,19 +100,19 @@ pipes: llm: litellm_llm.gemini/gemini-2.0-flash-exp - name: intent_classification llm: litellm_llm.gemini/gemini-2.0-flash-exp - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 document_store: qdrant - name: data_assistance llm: litellm_llm.gemini/gemini-2.0-flash-exp - name: sql_pairs_indexing document_store: qdrant - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 - name: sql_pairs_deletion document_store: qdrant - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 - name: sql_pairs_retrieval document_store: qdrant - embedder: openai_embedder.text-embedding-004 + embedder: litellm_embedder.text-embedding-004 llm: litellm_llm.gemini/gemini-2.0-flash-exp - name: preprocess_sql_data llm: litellm_llm.gemini/gemini-2.0-flash-exp diff --git a/wren-ai-service/docs/config_examples/config.groq.yaml b/wren-ai-service/docs/config_examples/config.groq.yaml index fc40afbdc..13bead784 100644 --- a/wren-ai-service/docs/config_examples/config.groq.yaml +++ b/wren-ai-service/docs/config_examples/config.groq.yaml @@ -14,12 +14,13 @@ models: --- type: embedder -provider: openai_embedder +provider: litellm_embedder models: - # find EMBEDDER_OPENAI_API_KEY and fill in value of api key in ~/.wrenai/.env - - model: text-embedding-3-large # put your openai compatible embedder model name here -url: https://api.openai.com/v1 # change this according to your openai compatible embedder model -timeout: 120 + # define OPENAI_API_KEY= in ~/.wrenai/.env if you are using openai embedding model + # please refer to LiteLLM documentation for more details: https://docs.litellm.ai/docs/providers + - model: text-embedding-3-large # put your embedding model name here, if it is not openai embedding model, should be / + api_base: https://api.openai.com/v1 # change this according to your embedding model + timeout: 120 --- type: engine @@ -41,20 +42,20 @@ recreate_index: false type: pipeline pipes: - name: db_schema_indexing - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: historical_question_indexing - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: table_description_indexing - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: db_schema_retrieval llm: litellm_llm.groq/llama-3.3-70b-specdec - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: historical_question_retrieval - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: sql_generation llm: litellm_llm.groq/llama-3.3-70b-specdec @@ -90,7 +91,7 @@ pipes: llm: litellm_llm.groq/llama-3.3-70b-specdec - name: question_recommendation_db_schema_retrieval llm: litellm_llm.groq/llama-3.3-70b-specdec - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: question_recommendation_sql_generation llm: litellm_llm.groq/llama-3.3-70b-specdec @@ -101,19 +102,19 @@ pipes: llm: litellm_llm.groq/llama-3.3-70b-specdec - name: intent_classification llm: litellm_llm.groq/llama-3.3-70b-specdec - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large document_store: qdrant - name: data_assistance llm: litellm_llm.groq/llama-3.3-70b-specdec - name: sql_pairs_indexing document_store: qdrant - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large - name: sql_pairs_deletion document_store: qdrant - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large - name: sql_pairs_retrieval document_store: qdrant - embedder: openai_embedder.text-embedding-3-large + embedder: litellm_embedder.text-embedding-3-large llm: litellm_llm.groq/llama-3.3-70b-specdec - name: preprocess_sql_data llm: litellm_llm.groq/llama-3.3-70b-specdec diff --git a/wren-ai-service/docs/config_examples/config.ollama.yaml b/wren-ai-service/docs/config_examples/config.ollama.yaml index c6c781c4c..de09ff16e 100644 --- a/wren-ai-service/docs/config_examples/config.ollama.yaml +++ b/wren-ai-service/docs/config_examples/config.ollama.yaml @@ -16,8 +16,8 @@ models: type: embedder provider: litellm_embedder models: -- model: openai/nomic-embed-text # put your ollama embedder model name here - api_base: http://host.docker.internal:11434/v1 # change this to your ollama host, url should be +- model: openai/nomic-embed-text # put your ollama embedder model name here, openai/ + api_base: http://host.docker.internal:11434/v1 # change this to your ollama host, api_base should be /v1 timeout: 120 ---