The "telephone between two agents" problem you describe is painfully real. I've been building a decentralized GPU inference network and coordination between distributed agents is a constant headache.
How do the LLM-powered guardrails work in practice? Is there noticeable latency from the filtering step, or is it negligible compared to the agent response time?