@matrixmayhem if you have the first message sent, the rephrase prompt will not be used. When you have a 2nd message sent, then it will be used.
@matrixmayhem if you have the first message sent, the rephrase prompt will not be used. When you have a 2nd message sent, then it will be used.
Originally posted by @HenryHengZJ in https://github.com/FlowiseAI/Flowise/issues/1742#issuecomment-1952745253
I have tired it with multiple message and it still uses the user mesage instead of the 'rephrased message'. This problem has started after the update to the "Conversational Retrival QA Chain" node.
Your chat doesn't have a memory node. Add a simple memory node, and you'll see the chat_history value include prior messages.
I am using upstash (though not in the demo above). I have also tried buffer, buffer window etc.
Do you have screenshots from the prior version where this was working? Or can you rollback and show screenshots of the prior node working as you're expecting it to?
"I installed version 1.3.9 and it behaved exactly as it should. Here are the screenshots: The second question, "Who is Luke?" gets rephrased and asked as, "Can you please provide more information on who Luke is?" Then the third question is asked, "Who is his father?" gets rephrased and asked as, "Who is Anakin Skywalker?" (screenshot attached below): The problem in version 1.5.0 is that the question does get rephrased but is not passed to the language model. This is clearly a bug. What's the point of this intermediate step of rephrasing the question if it is not passed to the language model?"
@matrixmayhem Thanks for the additional information. That has helped me identify that the QA chain was working up to at least version 1.4.9.
the rephrased question get passed to vector store to get relevant context, not LLM
1.) Rephrased question:
2.) Vector Retriever: