'Invalid or incomplete response' when using Mistral API
I have spent lot of time to understanding the Mistral API within the CrewAI framework, specifically for 'mistral-medium' and 'mistral-small'. Although I am able to execute the code smoothly, I am encountering an issue: it returns the error 'Invalid or incomplete response. Parsing LLM output produced both a final answer and a parse-able action.'
To address this, I have attempted the following solutions, but without success:
Updating the versions of the relevant packages: langchain to 0.1.1, langchain_mistralai to 0.0.3, and crewai to 0.1.32. Adjusting the parameters such as temperature, max_tokens, and top_p across various ranges.
Does anyone have a solution for this problem?" Thanks
I've wrote email to Mistral AI and the reply is they don't support Stop parameter at this moment. They're still working on it.
Also having an issue with this.
can you try the new RC and tell me if it works better?
pip install crewai==0.17.0rc2
João Moura @joaomdmoura http://twitter.com/joaomdmoura
Em sáb., 2 de mar. de 2024 às 17:05, Joaquin Dominguez < @.***> escreveu:
Also having an issue with this.
— Reply to this email directly, view it on GitHub https://github.com/joaomdmoura/crewAI/issues/155#issuecomment-1974895013, or unsubscribe https://github.com/notifications/unsubscribe-auth/AAFC3N5E6FJJ4O3KKVNNW7TYWIWHPAVCNFSM6AAAAABCASIGXCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTSNZUHA4TKMBRGM . You are receiving this because you are subscribed to this thread.Message ID: @.***>
I just tried it with the new rc you specified and still getting same error: