Streaming with Custom Format
Create a custom stream to control the streaming format and structure of tool calls instead of using the built-in AI SDK data stream format (toUIMessageStream()).
fullStream (on StreamTextResult) gives you direct access to all model events. You can transform, filter, and structure these events into your own streaming format. This gives you the benefits of the AI SDK's unified provider interface without prescribing how you consume the stream.
You can:
- Define your own stream chunk format
- Control how steps and tool calls are structured
- Parse the stream manually on the client
- Build custom UI from your stream data
For complete control over both the streaming format and the execution loop, combine this pattern with a manual agent loop.
Implementation
Server
Create a route handler that calls a model and then streams the responses in a custom format:
import { tools } from '@/ai/tools'; // your toolsimport { streamText } from 'ai';
export type StreamEvent = | { type: 'text'; text: string } | { type: 'tool-call'; toolName: string; input: unknown } | { type: 'tool-result'; toolName: string; result: unknown };
const encoder = new TextEncoder();
export async function POST(request: Request) { const { prompt } = await request.json();
const result = streamText({ prompt, model: 'anthropic/claude-sonnet-4.5', tools, });
const formatEvent = (data: StreamEvent) => { return encoder.encode('data: ' + JSON.stringify(data) + '\n\n'); };
const stream = new ReadableStream({ async start(controller) { for await (const part of result.fullStream) { switch (part.type) { case 'text-delta': controller.enqueue(formatEvent({ type: 'text', text: part.text })); break; case 'tool-call': controller.enqueue( formatEvent({ type: 'tool-call', toolName: part.toolName, input: part.input, }), ); break; case 'tool-result': controller.enqueue( formatEvent({ type: 'tool-result', toolName: part.toolName, result: part.output, }), ); break; } } controller.close(); }, });
return new Response(stream, { headers: { 'Content-Type': 'text/event-stream' }, });}The route uses streamText to process the prompt with tools. Each event (text, tool calls, tool results) is encoded as a Server-Sent Event with a data: prefix and sent to the client.
Client
Create a simple interface that parses and displays the stream:
'use client';
import { useState } from 'react';import { StreamEvent } from './api/stream/route';
export default function Home() { const [prompt, setPrompt] = useState(''); const [events, setEvents] = useState<StreamEvent[]>([]); const [isStreaming, setIsStreaming] = useState(false);
const handleSubmit = async () => { setEvents([]); setIsStreaming(true); setPrompt('');
const response = await fetch('/api/stream', { method: 'POST', headers: { 'Content-Type': 'application/json' }, body: JSON.stringify({ prompt }), });
const reader = response.body?.getReader(); const decoder = new TextDecoder();
if (reader) { let buffer = ''; while (true) { const { done, value } = await reader.read(); if (done) break;
buffer += decoder.decode(value, { stream: true }); const lines = buffer.split('\n'); buffer = lines.pop() || '';
for (const line of lines) { if (line.trim()) { const dataStr = line.replace(/^data: /, ''); const event = JSON.parse(dataStr) as StreamEvent; setEvents(prev => [...prev, event]); } } } }
setIsStreaming(false); };
return ( <div> <input value={prompt} onChange={e => setPrompt(e.target.value)} placeholder="Enter a prompt..." /> <button onClick={handleSubmit} disabled={isStreaming}> {isStreaming ? 'Streaming...' : 'Send'} </button>
<pre>{JSON.stringify(events, null, 2)}</pre> </div> );}How it works
The client uses the Fetch API to stream responses from the server. Since the server sends Server-Sent Events (newline-delimited with data: prefix), the client:
- Reads chunks from the stream using
getReader() - Decodes the binary chunks to text
- Splits by newlines to identify complete events
- Removes the
data:prefix and parses the JSON, then appends it to the events list
Events are rendered in order as they arrive, giving you a linear representation of the AI's response.