Describe the bug
Streaming no longer works in a Chatflow using the Conversational Retrieval QA Chain after upgrading Flowise from v3.0.13 to v3.1.1.
The response is returned only once fully completed (no progressive streaming), whereas streaming was working correctly in v3.0.13.
The issue occurs across multiple Chat Models, so it does not appear to be provider-specific. This suggests a regression in the Conversational Retrieval QA Chain implementation.
To Reproduce
- Run Flowise v3.1.1
- Create a Chatflow with:
- A Conversational Retrieval QA Chain
- A Chat Model with streaming enabled
- A retriever (e.g., vector store)
- Open the chat UI
- Send any query
Expected behavior
The response should be streamed progressively in the UI (token-by-token or chunked output), as in v3.0.13.
Screenshots
No response
Flow
No response
Use Method
Docker
Flowise Version
3.1.1
Operating System
None
Browser
None
Additional context
- ✅ Streaming works in:
- Agentflow
- Chatflow using Conversation Chain
- ❌ Streaming does NOT work in:
- Chatflow using Conversational Retrieval QA Chain
- Tested with multiple LLM providers → same behavior
- No specific error in logs (streaming silently not triggered)
Possible regression source:
PR #5893 — “Feat/Reasoning Support, Migrate LangChain to v1” (Flowise 3.1.0)
This PR explicitly migrated:
- ConversationalRetrievalQAChain.ts
from langchain/chains → @langchain/classic/chains
and langchain/memory → @langchain/classic/memory
Since:
- streaming still works in Agentflow
- streaming still works with Conversation Chain
- the issue is isolated to Conversational Retrieval QA Chain
it seems likely the regression is related to how streaming callbacks / handlers are wired after the LangChain v1 migration for this specific chain.
Describe the bug
Streaming no longer works in a Chatflow using the Conversational Retrieval QA Chain after upgrading Flowise from v3.0.13 to v3.1.1.
The response is returned only once fully completed (no progressive streaming), whereas streaming was working correctly in v3.0.13.
The issue occurs across multiple Chat Models, so it does not appear to be provider-specific. This suggests a regression in the Conversational Retrieval QA Chain implementation.
To Reproduce
Expected behavior
The response should be streamed progressively in the UI (token-by-token or chunked output), as in v3.0.13.
Screenshots
No response
Flow
No response
Use Method
Docker
Flowise Version
3.1.1
Operating System
None
Browser
None
Additional context
Possible regression source:
PR #5893 — “Feat/Reasoning Support, Migrate LangChain to v1” (Flowise 3.1.0)
This PR explicitly migrated:
from langchain/chains → @langchain/classic/chains
and langchain/memory → @langchain/classic/memory
Since:
it seems likely the regression is related to how streaming callbacks / handlers are wired after the LangChain v1 migration for this specific chain.