WrenAI icon indicating copy to clipboard operation
WrenAI copied to clipboard

wren-ai-service - ERROR - ask pipeline - NO_RELEVANT_SQL:

Open gabilanbrc opened this issue 1 year ago • 6 comments

Hi Team, I have just installed Wren with Ollama with this config LLM_PROVIDER=ollama_llm GENERATION_MODEL=mistral-nemo:latest EMBEDDER_PROVIDER=ollama_embedder EMBEDDING_MODEL=mxbai-embed-large EMBEDDING_MODEL_DIMENSION=1024 None of the three examples of the E-Commerce (or NBA) worked for me with (after more than one minute) the error: 2024-09-03 17:02:29 2024-09-03 20:02:29,912 - wren-ai-service - ERROR - ask pipeline - NO_RELEVANT_SQL: What are the top 3 value for orders placed by customers in each city? (ask.py:233)

I have managed to make it work only with a simple question like How many orders are there? That also took more than 1 minute to reply 181

Should I change the EMBEDDING_MODEL_DIMENSION or something else? Thanks in advance

gabilanbrc avatar Sep 03 '24 20:09 gabilanbrc

Also after switching to gemma2:2b i get faster results but the error remains 2024-09-03 17:53:22 2024-09-03 20:53:22,200 - wren-ai-service - ERROR - Error in GenerationPostProcessor: 'results' (post_processors.py:48) 2024-09-03 17:53:22 Traceback (most recent call last): 2024-09-03 17:53:22 File "/src/pipelines/ask/components/post_processors.py", line 29, in run 2024-09-03 17:53:22 cleaned_generation_result = orjson.loads( 2024-09-03 17:53:22 ^^^^^^^^^^^^^ 2024-09-03 17:53:22 KeyError: 'results' 2024-09-03 17:53:22 2024-09-03 20:53:22,200 - wren-ai-service - ERROR - ask pipeline - NO_RELEVANT_SQL: Which is the oldest team? (ask.py:233)

gabilanbrc avatar Sep 03 '24 20:09 gabilanbrc

@gabilanbrc

thanks for reaching out! Actually it's a known issue and the root cause is some parts of the ai pipeline are too complex for weaker LLMs. The good news is we have improved the ai pipeline these days and will alleviate this issue in the next release! Stay tuned!

cyyeh avatar Sep 04 '24 00:09 cyyeh

@gabilanbrc could u try install the latest version and try again?

cyyeh avatar Sep 05 '24 14:09 cyyeh

Hi! I did and now the error is different 2024-09-05 13:58:06 ******************************************************************************** 2024-09-05 13:58:06 > construct_retrieval_results [src.pipelines.ask.retrieval.construct_retrieval_results()] encountered an error< 2024-09-05 13:58:06 > Node inputs: 2024-09-05 13:58:06 {'construct_db_schemas': "<Task finished name='Task-275' " 2024-09-05 13:58:06 'coro=<AsyncGraphAda...', 2024-09-05 13:58:06 'dbschema_retrieval': "<Task finished name='Task-274' coro=<AsyncGraphAda...", 2024-09-05 13:58:06 'filter_columns_in_tables': "<Task finished name='Task-277' " 2024-09-05 13:58:06 'coro=<AsyncGraphAda...'} 2024-09-05 13:58:06 ******************************************************************************** 2024-09-05 13:58:06 Traceback (most recent call last): 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 122, in new_fn 2024-09-05 13:58:06 await fn(**fn_kwargs) if asyncio.iscoroutinefunction(fn) else fn(**fn_kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/src/utils.py", line 95, in wrapper_timer 2024-09-05 13:58:06 return func(*args, **kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 225, in sync_wrapper 2024-09-05 13:58:06 self._handle_exception(observation, e) 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 428, in _handle_exception 2024-09-05 13:58:06 raise e 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 223, in sync_wrapper 2024-09-05 13:58:06 result = func(*args, **kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/src/pipelines/ask/retrieval.py", line 243, in construct_retrieval_results 2024-09-05 13:58:06 columns_and_tables_needed = orjson.loads(filter_columns_in_tables["replies"][0])[ 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 KeyError: 'results' 2024-09-05 13:58:06 ------------------------------------------------------------------- 2024-09-05 13:58:06 Oh no an error! Need help with Hamilton? 2024-09-05 13:58:06 Join our slack and ask for help! https://join.slack.com/t/hamilton-opensource/shared_invite/zt-1bjs72asx-wcUTgH7q7QX1igiQ5bbdcg 2024-09-05 13:58:06 ------------------------------------------------------------------- 2024-09-05 13:58:06 2024-09-05 13:58:06 2024-09-05 16:58:06,703 - wren-ai-service - ERROR - ask pipeline - OTHERS: 'results' (ask.py:307) 2024-09-05 13:58:06 Traceback (most recent call last): 2024-09-05 13:58:06 File "/src/web/v1/services/ask.py", line 145, in ask 2024-09-05 13:58:06 retrieval_result = await self._pipelines["retrieval"].run( 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/src/utils.py", line 119, in wrapper_timer 2024-09-05 13:58:06 return await process(func, *args, **kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/src/utils.py", line 103, in process 2024-09-05 13:58:06 return await func(*args, **kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 188, in async_wrapper 2024-09-05 13:58:06 self._handle_exception(observation, e) 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 428, in _handle_exception 2024-09-05 13:58:06 raise e 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 186, in async_wrapper 2024-09-05 13:58:06 result = await func(*args, **kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/src/pipelines/ask/retrieval.py", line 341, in run 2024-09-05 13:58:06 return await self._pipe.execute( 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 368, in execute 2024-09-05 13:58:06 raise e 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 359, in execute 2024-09-05 13:58:06 outputs = await self.raw_execute(final_vars, overrides, display_graph, inputs=inputs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 320, in raw_execute 2024-09-05 13:58:06 raise e 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 315, in raw_execute 2024-09-05 13:58:06 results = await await_dict_of_tasks(task_dict) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 23, in await_dict_of_tasks 2024-09-05 13:58:06 coroutines_gathered = await asyncio.gather(*coroutines) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 36, in process_value 2024-09-05 13:58:06 return await val 2024-09-05 13:58:06 ^^^^^^^^^ 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 122, in new_fn 2024-09-05 13:58:06 await fn(**fn_kwargs) if asyncio.iscoroutinefunction(fn) else fn(**fn_kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/src/utils.py", line 95, in wrapper_timer 2024-09-05 13:58:06 return func(*args, **kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 225, in sync_wrapper 2024-09-05 13:58:06 self._handle_exception(observation, e) 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 428, in _handle_exception 2024-09-05 13:58:06 raise e 2024-09-05 13:58:06 File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 223, in sync_wrapper 2024-09-05 13:58:06 result = func(*args, **kwargs) 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 File "/src/pipelines/ask/retrieval.py", line 243, in construct_retrieval_results 2024-09-05 13:58:06 columns_and_tables_needed = orjson.loads(filter_columns_in_tables["replies"][0])[ 2024-09-05 13:58:06 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ 2024-09-05 13:58:06 KeyError: 'results' 2024-09-05 13:58:07 INFO: 172.19.0.6:50196 - "GET /v1/asks/e3c5cc21-beee-43af-8300-c202c8686656/result HTTP/1.1" 200 OK

The config I used was

LLM_PROVIDER=ollama_llm # ollama_llm, azure_openai_llm, ollama_llm GENERATION_MODEL=llama3.1:8b # gemma2:2b, llama3.1:8b GENERATION_MODEL_KWARGS={"temperature": 0, "n": 1, "max_tokens": 4096, "response_format": {"type": "json_object"}} COLUMN_INDEXING_BATCH_SIZE=50 TABLE_RETRIEVAL_SIZE=10 TABLE_COLUMN_RETRIEVAL_SIZE=1000

LLM_OLLAMA_URL=http://host.docker.internal:11434

EMBEDDER_PROVIDER=ollama_embedder # openai_embedder, azure_openai_embedder, ollama_embedder EMBEDDING_MODEL=mxbai-embed-large EMBEDDING_MODEL_DIMENSION=1024

For the purpose of initial testing I can try to use the Free API from Gemini, can someone give me an example of a configuration for that?

gabilanbrc avatar Sep 05 '24 17:09 gabilanbrc

I`m have the same error...

nyeeldzn avatar Sep 14 '24 13:09 nyeeldzn

@gabilanbrc @nyeeldzn could you try the latest version and check if the error happen again?

@gabilanbrc for setting up gemini, here is instructions: https://docs.getwren.ai/oss/installation/custom_llm#running-wren-ai-with-your-custom-llm-or-document-store, and now we only support text generation models

relevant instructions to setup gemini: https://cloud.google.com/vertex-ai/generative-ai/docs/multimodal/call-vertex-using-openai-library#generativeaionvertexai_gemini_chat_completions_non_streaming-python_vertex_ai_sdk

and the .env.ai should look like this:

## LLM
# openai_llm, azure_openai_llm, ollama_llm
LLM_PROVIDER=openai_llm
LLM_TIMEOUT=120
GENERATION_MODEL=google/gemini-1.5-flash-002
GENERATION_MODEL_KWARGS={"temperature": 0, "n": 1, "max_tokens": 4096, "response_format": {"type": "json_object"}}

# openai or openai-api-compatible
LLM_OPENAI_API_KEY={API_KEY}
LLM_OPENAI_API_BASE=https://{LOCATION}-aiplatform.googleapis.com/v1beta1/projects/{PROJECT_ID}/locations/{LOCATION}/endpoints/openapi

# azure_openai
LLM_AZURE_OPENAI_API_KEY=
LLM_AZURE_OPENAI_API_BASE=
LLM_AZURE_OPENAI_VERSION=

# ollama
LLM_OLLAMA_URL=http://host.docker.internal:11434/

## EMBEDDER
# openai_embedder, azure_openai_embedder, ollama_embedder
EMBEDDER_PROVIDER=openai_embedder
EMBEDDER_TIMEOUT=120
# supported embedding models providers by qdrant: https://qdrant.tech/documentation/embeddings/
EMBEDDING_MODEL=text-embedding-3-large
EMBEDDING_MODEL_DIMENSION=3072

# openai or openai-api-compatible
EMBEDDER_OPENAI_API_KEY=<API_KEY>
EMBEDDER_OPENAI_API_BASE=https://api.openai.com/v1

# azure_openai
EMBEDDER_AZURE_OPENAI_API_KEY=
EMBEDDER_AZURE_OPENAI_API_BASE=
EMBEDDER_AZURE_OPENAI_VERSION=

# ollama
EMBEDDER_OLLAMA_URL=http://host.docker.internal:11434/


## DOCUMENT_STORE
DOCUMENT_STORE_PROVIDER=qdrant

QDRANT_HOST=qdrant


## Langfuse: https://langfuse.com/
# empty means disabled
LANGFUSE_ENABLE=
LANGFUSE_SECRET_KEY=
LANGFUSE_PUBLIC_KEY=
LANGFUSE_HOST=https://cloud.langfuse.com

cyyeh avatar Oct 01 '24 15:10 cyyeh

@cyyeh I'm also receiving similar error with local llm OLLAMA, it seems configured llm is not able to generate response on expected format. ERROR at : onstruct_retrieval_results [src.pipelines.retrieval.retrieval.construct_retrieval_results()] encountered an error. I tried with demo E-commerce dataset with basic recommended question. 3 models are explored all 3 models give same error:

  1. llama3.370b-instruct-q4_K_M
  2. llama3-8b-instruct-fp16
  3. phi4 Question: Which are the top 3 cities with the highest number of orders? model config files are attached. Stacktrace is here.

INFO: 172.30.0.21:53848 - "POST /v1/asks HTTP/1.1" 200 OK I0221 03:44:41.622 10 wren-ai-service:132] HistoricalQuestion Retrieval pipeline is running... I0221 03:44:41.641 10 wren-ai-service:318] Intent Classification pipeline is running... INFO: 172.30.0.21:53854 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:53864 - "GET /v1/ask

config.llama3.370b-instruct-q4_K_M.yaml.txt config.llama3-8b-instruct-fp16.yaml.txt config.phi4.yaml.txt

s/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK I0221 03:44:43.715 10 wren-ai-service:157] dbschema_retrieval with table_names: ['olist_order_items_dataset', 'olist_geolocation_dataset', 'olist_orders_dataset', 'olist_order_payments_dataset', 'olist_sellers_dataset', 'olist_order_reviews_dataset', 'product_category_name_translation', 'olist_products_dataset', 'olist_customers_dataset'] INFO: 172.30.0.21:53870 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:53886 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:58234 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:58238 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:58242 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:58244 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:58250 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:58260 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:58276 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:42826 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK I0221 03:44:55.890 10 wren-ai-service:440] Ask Retrieval pipeline is running... I0221 03:44:55.969 10 wren-ai-service:267] db_schemas token count is greater than 100,000, so we will prune columns INFO: 172.30.0.21:42838 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:42848 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:42854 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:42868 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:42874 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK


construct_retrieval_results [src.pipelines.retrieval.retrieval.construct_retrieval_results()] encountered an error< Node inputs: {'check_using_db_schemas_without_pruning': "<Task finished name='Task-187' " 'coro=<AsyncGraphAda...', 'construct_db_schemas': "<Task finished name='Task-186' " 'coro=<AsyncGraphAda...', 'dbschema_retrieval': "<Task finished name='Task-185' coro=<AsyncGraphAda...", 'filter_columns_in_tables': "<Task finished name='Task-189' " 'coro=<AsyncGraphAda...'}


Traceback (most recent call last): File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 122, in new_fn await fn(**fn_kwargs) if asyncio.iscoroutinefunction(fn) else fn(**fn_kwargs) ^^^^^^^^^^^^^^^ File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 256, in sync_wrapper self._handle_exception(observation, e) File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 517, in _handle_exception raise e File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 254, in sync_wrapper result = func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "/src/pipelines/retrieval/retrieval.py", line 306, in construct_retrieval_results columns_and_tables_needed = orjson.loads( ^^^^^^^^^^^^^ orjson.JSONDecodeError: unexpected character: line 1 column 1 (char 0)

Oh no an error! Need help with Hamilton? Join our slack and ask for help! https://join.slack.com/t/hamilton-opensource/shared_invite/zt-2niepkra8-DGKGf_tTYhXuJWBTXtIs4g

E0221 03:45:03.247 10 wren-ai-service:425] ask pipeline - OTHERS: unexpected character: line 1 column 1 (char 0) Traceback (most recent call last): File "/src/web/v1/services/ask.py", line 246, in ask retrieval_result = await self._pipelines["retrieval"].run( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 219, in async_wrapper self._handle_exception(observation, e) File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 517, in _handle_exception raise e File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 217, in async_wrapper result = await func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/src/pipelines/retrieval/retrieval.py", line 441, in run return await self._pipe.execute( ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 375, in execute raise e File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 366, in execute outputs = await self.raw_execute(_final_vars, overrides, display_graph, inputs=inputs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 326, in raw_execute raise e File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 321, in raw_execute results = await await_dict_of_tasks(task_dict) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 23, in await_dict_of_tasks coroutines_gathered = await asyncio.gather(*coroutines) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 36, in process_value return await val ^^^^^^^^^ File "/app/.venv/lib/python3.12/site-packages/hamilton/async_driver.py", line 122, in new_fn await fn(**fn_kwargs) if asyncio.iscoroutinefunction(fn) else fn(**fn_kwargs) ^^^^^^^^^^^^^^^ File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 256, in sync_wrapper self._handle_exception(observation, e) File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 517, in _handle_exception raise e File "/app/.venv/lib/python3.12/site-packages/langfuse/decorators/langfuse_decorator.py", line 254, in sync_wrapper result = func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "/src/pipelines/retrieval/retrieval.py", line 306, in construct_retrieval_results columns_and_tables_needed = orjson.loads( ^^^^^^^^^^^^^ orjson.JSONDecodeError: unexpected character: line 1 column 1 (char 0) INFO: 172.30.0.21:42884 - "GET /v1/asks/76f1b952-e587-4e68-b5ed-9e3c3596c119/result HTTP/1.1" 200 OK INFO: 172.30.0.21:42896 - "POST /v1/question-recommendations HTTP/1.1" 200 OK I0221 03:45:03.903 10 wren-ai-service:154] Request d043bb29-b6db-4838-a958-c76da098be55: Generate Question Recommendation pipeline is running... I0221 03:45:03.911 10 wren-ai-service:263] Question Recommendation pipeline is running... INFO: 172.30.0.21:42904 - "GET /v1/question-recommendations/d043bb29-b6db-4838-a958-c76da098be55 HTTP/1.1" 200 OK INFO: 172.30.0.21:42914 - "GET /v1/question-recommendations/d043bb29-b6db-4838-a958-c76da098be55 HTTP/1.1" 200 OK INFO: 172.30.0.21:42530 - "GET /v1/question-recommendations/d043bb29-b6db-4838-a958-c76da098be55 HTTP/1.1" 200 OK INFO: 172.30.0.21:42544 - "GET /v1/question-recommendations/d043bb29-b6db-4838-a958-c76da098be55 HTTP/1.1" 200 OK INFO: 172.30.0.21:42558 - "GET /v1/question-recommendations/d043bb29-b6db-4838-a958-c76da098be55 HTTP/1.1" 200 OK INFO: 172.30.0.21:42568 - "GET /v1/question-recommendations/d043bb29-b6db-4838-a958-c76da098be55 HTTP/1.1" 200 OK INFO: 172.30.0.21:42574 - "GET /v1/question-recommendations/d043bb29-b6db-4838-a958-c76da098be55 HTTP/1.1" 200 OK E0221 03:45:12.368 10 wren-ai-service:60] Error decoding JSON: unexpected character: line 1 column 1 (char 0) E0221 03:45:12.369 10 wren-ai-service:60] An error occurred during question recommendation generation: 'list' object has no attribute 'get' INFO: 172.30.0.21:42578 - "GET /v1/question-recommendations/d043bb29-b6db-4838-a958-c76da098be55 HTTP/1.1" 200 OK

I would like to know how to fix this error, Which open source model can work with this.

numberonewastefellow avatar Feb 21 '25 05:02 numberonewastefellow