Real LLM Streaming with n8n – Here’s How (with a Little Help from Supabase)

n8n, for all its power in workflow automation, is NOT natively built for streaming HTTP responses. Its sequential, node-by-node execution model is fantastic for many tasks, but it’s a fundamental blocker for true LLM streaming.

The good news? I’ve been wrestling with this and have landed on a robust architectural pattern that brings that smooth streaming experience to n8n-powered UIs, primarily by leveraging the power of Supabase Edge Functions and Realtime subscriptions.