-
Notifications
You must be signed in to change notification settings - Fork 443
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Can't start wren-ai-service for a qdrant interaction error #1121
Comments
@nemesisq3a could you provide your config.yaml? Thanks |
Sure! Thanks
|
:39:04 return await val |
@wisonzhu could u try to use litellm_llm as provider? azure_openai will be deprecated soon |
Since Wren AI is running in containers, you may need to add Ollama to the same docker network of WrenAI's, and change the url of embedding model accordingly |
16:59:04 I0116 08:59:04.712 12 wren-ai-service:263] Question Recommendation pipeline is running... |
Ollama is running on a different server, in the same lan network (is a GPU dedicated server). I configured the http://192.168.200.111:11434 url (the ollama server machine lan ip) in the config.yaml file and in the .env file in accordance with the guide I found here.
From my prospective seems that the problem is beetween wren-ai-service and qdrant in itself. |
@nemesisq3a could you join discord, let's schedule a time so that you could you share your screen with me and we could look at the issue together? |
All the containers are running on a corporate cloud, so I can't share this with you. Let me see if I can replicate this thing on my home desktop pc, I've a popos installation and a RTX 3080, all should work. When I'm ready I'll join you on discord. Thanks for now |
I am getting the exact same error. It's raised from
and my WrenAI info as below:
Configured env vars:
I listed relevant vars only above. And the initialization logs of ai service as below:
All these configuration and settings are verified, and they are working. |
How can this problem be solved? |
@nemesisq3a @thiner @wisonzhu @kyyz147 Could you try to follow the config.yaml examples here and adapt to your usecases first? feel free to reach out to me if there are further issues. Thank you! |
Ok, I replicated the configuration on my desktop PC and same get_document_embedder error..
In this case I have to keep the openai name even if it's not?
right? |
@nemesisq3a yes, as I said above, as long as you are using ollama LLMs, please follow the format: |
@nemesisq3a btw the reason is that I found performance of using |
As I wrote on discord I tried using the provided example.
Still same error as before:
|
I am still getting the same error. My ai-service configuration as below:
The error message:
|
__
please use api_base for openai_embedder instead of url, thanks |
The error remains the same. I am going to provide the whole log:
|
Please give me the config.yaml. There must be some config issues |
@cyyeh The config.yaml file as below: type: llm
provider: litellm_llm
timeout: 120
models:
- model: openai/qwen2.5-72b-instruct
api_base: https://llm.localai/v1
api_key_name: LLM_OPENAI_API_KEY
kwargs:
temperature: 0
n: 1
# for better consistency of llm response, refer: https://platform.openai.com/docs/api-reference/chat/create#chat-create-seed
seed: 0
max_tokens: 32768
response_format:
type: json_object
- model: openai/qwen2.5-coder-instruct
api_base: https://llm.localai/v1
api_key_name: LLM_OPENAI_API_KEY
kwargs:
temperature: 0
n: 1
# for better consistency of llm response, refer: https://platform.openai.com/docs/api-reference/chat/create#chat-create-seed
seed: 0
max_tokens: 16384
response_format:
type: json_object
---
type: embedder
provider: openai_embedder
timeout: 120
models:
- model: openai/bce-embedding-base
dimension: 768
api_base: http://bce-embedding.localai:8080/v1
---
type: engine
provider: wren_ui
endpoint: http://wren-ui:3000
---
type: document_store
provider: qdrant
location: http://qdrant:6333
embedding_model_dim: 3072
timeout: 120
recreate_index: true
---
type: pipeline
pipes:
- name: db_schema_indexing
embedder: litellm_llm.openai/bce-embedding-base
document_store: qdrant
- name: historical_question_indexing
embedder: litellm_llm.openai/bce-embedding-base
document_store: qdrant
- name: table_description_indexing
embedder: litellm_llm.openai/bce-embedding-base
document_store: qdrant
- name: db_schema_retrieval
llm: litellm_llm.openai/qwen2.5-72b-instruct
embedder: litellm_llm.openai/bce-embedding-base
document_store: qdrant
- name: historical_question_retrieval
embedder: litellm_llm.openai/bce-embedding-base
document_store: qdrant
- name: sql_generation
llm: litellm_llm.openai/qwen2.5-coder-instruct
engine: wren_ui
- name: sql_correction
llm: litellm_llm.openai/qwen2.5-coder-instruct
engine: wren_ui
- name: followup_sql_generation
llm: litellm_llm.openai/qwen2.5-coder-instruct
engine: wren_ui
- name: sql_summary
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: sql_answer
llm: litellm_llm.openai/qwen2.5-72b-instruct
engine: wren_ui
- name: sql_breakdown
llm: litellm_llm.openai/qwen2.5-coder-instruct
engine: wren_ui
- name: sql_expansion
llm: litellm_llm.openai/qwen2.5-coder-instruct
engine: wren_ui
- name: sql_explanation
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: sql_regeneration
llm: litellm_llm.openai/qwen2.5-coder-instruct
engine: wren_ui
- name: semantics_description
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: relationship_recommendation
llm: litellm_llm.openai/qwen2.5-72b-instruct
engine: wren_ui
- name: question_recommendation
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: intent_classification
llm: litellm_llm.openai/qwen2.5-72b-instruct
embedder: litellm_llm.openai/bce-embedding-base
document_store: qdrant
- name: data_assistance
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: sql_pairs_preparation
document_store: qdrant
embedder: litellm_llm.openai/bce-embedding-base
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: sql_pairs_deletion
document_store: qdrant
embedder: litellm_llm.openai/bce-embedding-base
- name: sql_pairs_retrieval
document_store: qdrant
embedder: litellm_llm.openai/bce-embedding-base
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: preprocess_sql_data
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: sql_executor
engine: wren_ui
- name: chart_generation
llm: litellm_llm.openai/qwen2.5-72b-instruct
- name: chart_adjustment
llm: litellm_llm.openai/qwen2.5-72b-instruct
---
settings:
column_indexing_batch_size: 50
table_retrieval_size: 10
table_column_retrieval_size: 100
allow_using_db_schemas_without_pruning: false
query_cache_maxsize: 1000
query_cache_ttl: 3600
langfuse_host: http://langfuse.dify:3000
langfuse_enable: true
logging_level: DEBUG
development: false |
Describe the bug
Unable to start wren-ai-service from docker compose startup
To Reproduce
Download docker-compose.yaml, .env and config.yaml examples from https://github.com/Canner/WrenAI/tree/main/docker and configure them to use a ollama embedded LLM (self-hosted). I'm using a docker image of ollama running on a network machine with the Qwen2.5-coder LLM loaded.
Expected behavior
wren-ai-service positive startup
Desktop (please complete the following information):
Wren AI Information
Additional context
ERROR LOG:
The text was updated successfully, but these errors were encountered: