I've built a simple AI-chat interface for one of my products and I had experienced a latency of 6000 ms by adding a router LLM to selectively inject context to the main LLM. Dropped that idea completely.
EDIT: my bad, just read the term "multi-agent workflow" in the title. Still, some idea about the added latency would be helpful.