
nitro-imessage-agent
A durable iMessage AI agent built on:
- Nitro v3 — the API server
- Chat SDK +
chat-adapter-sendblue— message routing over Sendblue - Vercel AI SDK + AI Gateway — LLM replies, swap models with one constant
- Vercel Workflow — durable orchestration with retryable
"use step"units - evlog — structured wide-event logging with first-class AI SDK integration (token usage, tool calls, cost estimation)
A user texts your Sendblue number, Sendblue posts a webhook to this server, the Chat SDK dispatches the message, a workflow runs an agent step (LLM + tools), and the final reply is sent back through Sendblue. Each step is retryable on its own, so a transient LLM error or send hiccup never drops the inbound message.
┌─────────┐ ┌──────────────┐ ┌──────────────┐ ┌────────────┐│ user │───▶│ Sendblue │───▶│ POST /api/ │──▶│ Chat SDK ││iMessage │ │ (cloud) │ │webhooks/... │ │ onMention │└─────────┘ └──────────────┘ └──────────────┘ └─────┬──────┘▲ ││ ▼│ ┌────────────────────────┐│ │ workflow start(...) ││ └───────────┬────────────┘│ ▼│ ┌────────────────────────┐│ │ generateReply(use step)││ │ ├── LLM (AI Gateway) ││ │ ├── tools roundtrip ││ │ └── evlog wide event ││ └───────────┬────────────┘│ ▼│ ┌────────────────────────┐└───────────────────────────────────────────┤ postReply (use step) │Sendblue postMessage │ sendblue.postMessage │└────────────────────────┘
Architecture
The Sendblue cloud holds your dedicated phone line and forwards inbound iMessages to your server as HTTPS webhooks. Outbound replies go back through the same API. There is no gateway listener to keep alive, no Mac in production, and no cron.
The server/api/webhooks/sendblue.post.ts [blocked] route receives every webhook and hands it to chat.webhooks.sendblue(request). The Chat SDK then fires onNewMention (first DM in a thread) or onSubscribedMessage (every following DM) on the bot — handlers registered in server/plugins/imessage.ts [blocked] call start(replyToMessage, [thread.id, message.text]) to queue a workflow.
workflows/reply.ts [blocked] is a thin "use workflow" function that just chains two retryable steps from server/utils/agent-steps.ts [blocked]:
generateReplystep — callsgenerateTextagainst the Vercel AI Gateway. Tools registered in server/tools/ [blocked] are looped withstopWhen: stepCountIs(5)(LLM → tool → LLM until done). The model is wrapped withevlog/ai'sai.wrap()andexperimental_telemetry.integrationscarriescreateEvlogIntegration(ai)so token usage, tool execution timing, and estimated cost are captured into a wide event.postReplystep — callssendblue.postMessage. Independent retryability: a transient send error doesn't re-run the LLM call.
Workflow functions can't import Node-only packages like evlog directly, so the AI SDK calls live in steps (which run as normal Node) — that's why the actual logic is in server/utils/agent-steps.ts and the workflow file just orchestrates.
Local setup (development)
Sendblue is webhook-based, so to receive iMessages on your local machine you expose localhost:3000 through a public tunnel — we use ngrok.
Prerequisites
- Node 20+ (use
corepack enableto get pnpm) - A Sendblue account (sendblue.com). The free tier with a shared line is enough to demo — webhooks work, replies to verified contacts work, and you get 10 contact slots. The AI Agent plan ($100/mo) is required if you want a dedicated number with higher inbound volume; see pricing.
- ngrok installed and authenticated with a free account (
ngrok config add-authtoken <your-token>). Any HTTPS tunnel works —cloudflared tunnel,localtunnel, etc. — ngrok is just what we use here.
Install & run
pnpm installcp .env.example .env# fill in AI_GATEWAY_API_KEY + SENDBLUE_API_KEY + SENDBLUE_API_SECRET +# SENDBLUE_FROM_NUMBER + SENDBLUE_WEBHOOK_SECRETpnpm dev
In a second terminal:
ngrok http 3000
Copy the https://<id>.ngrok-free.app URL ngrok prints, then in the Sendblue dashboard set the inbound webhook URL to:
https://<id>.ngrok-free.app/api/webhooks/sendblue
Test
Text your Sendblue number from any phone. The dev server logs the inbound webhook, the workflow run starts, and a reply lands on your phone. Try "what time is it in Paris?" to validate the getCurrentTime tool path.
Production setup
Sendblue runs in the cloud and just talks HTTP, so production is the same as local minus the tunnel:
-
Deploy this repo to Vercel (or any Node host that supports Nitro):
pnpm dlx vercel. -
Set the same five env vars in your hosting provider's environment settings (Vercel → Project → Settings → Environment Variables).
-
In the Sendblue dashboard, point the inbound webhook URL at your production deployment:
https://<your-app>.vercel.app/api/webhooks/sendblue
That's it — no vercel.json, no cron, no Mac. The Vercel function spins up on each webhook and the Workflow runs durably in the background.
Why Sendblue and not Photon? Photon's self-serve Spectrum dashboard hands out credentials for the new
spectrum-tsSDK, which isn't compatible withchat-adapter-imessage— the latter still uses Photon's older Enterprise SDK and needs negotiated credentials from Photon sales. Sendblue is self-serve, supports webhooks on the free tier, has SMS fallback, and ships US numbers without A2P KYC. Swap the adapter if your needs differ —chat-adapter-imessage(Photon Enterprise),chat-adapter-blooio, or any future Chat SDK iMessage adapter all plug in the same way.
Configuration reference
| Env var | Required | When |
|---|---|---|
AI_GATEWAY_API_KEY | yes | Always. Auto-detected by the AI SDK (no NITRO_ prefix). |
SENDBLUE_API_KEY | yes | Always. Auto-detected by the adapter from process env. |
SENDBLUE_API_SECRET | yes | Always. Auto-detected by the adapter. |
SENDBLUE_FROM_NUMBER | yes | Your provisioned Sendblue line in E.164 format (e.g. +14155551234). |
SENDBLUE_WEBHOOK_SECRET | recommended | If set in the Sendblue dashboard, the adapter validates every incoming webhook against it. Set the same value here. |
Switching the model
Edit one constant in server/utils/agent-steps.ts [blocked]:
const MODEL = 'google/gemini-3-flash'
Any supported AI Gateway slug works. A few useful ones:
google/gemini-3-flashanthropic/claude-sonnet-4.5openai/gpt-4o-minixai/grok-4
The AI SDK reads AI_GATEWAY_API_KEY from the environment automatically, so no provider plumbing is needed.
Extending the agent
Add a tool
Tools live in server/tools/index.ts [blocked]. Each tool is a description + inputSchema (zod) + execute function. The "use step" directive on the execute body makes every tool call a retryable, observable workflow step.
Example (from this repo):
import { z } from 'zod'// eslint-disable-next-line require-awaitasync function getCurrentTime({ timezone }: { timezone: string }): Promise<string> {'use step'return new Intl.DateTimeFormat('en-US', {timeZone: timezone,dateStyle: 'full',timeStyle: 'long',}).format(new Date())}export const tools = {getCurrentTime: {description: 'Get the current date and time in a specific IANA timezone…',inputSchema: z.object({timezone: z.string().describe('IANA timezone identifier'),}),execute: getCurrentTime,},}
To add another tool, write a new step function and register it in the tools map. The agent picks it up automatically — generateText discovers them at call time.
Change the system prompt
Edit SYSTEM_PROMPT in server/utils/agent-steps.ts [blocked].
Add a workflow step
Any function annotated with "use step" becomes a retryable, durable step. Wrap higher-level orchestration in a "use workflow" function and call steps from it. See Workflows and steps for the full mental model. workflows/reply.ts [blocked] is the canonical example: a thin "use workflow" function chaining generateReply and postReply steps from server/utils/agent-steps.ts [blocked].
Workflow functions can't import Node-only modules (evlog, native bindings, etc.). Keep heavy logic in
"use step"files outsideworkflows/and import them into the workflow.
Tweak AI observability
createAILogger(log, { cost }) in server/utils/agent-steps.ts [blocked] controls token-cost estimation. Update COST_MAP with your real Gateway pricing, or set cost to undefined to disable. The wide event under the ai.* namespace already includes inputTokens, outputTokens, toolCalls, tools[] (timing per tool from createEvlogIntegration), msToFirstChunk, tokensPerSecond, and estimatedCost. See evlog AI SDK docs for the full field list.
Add a drain (Axiom, OTLP, Sentry, …)
The Nitro evlog module exposes a evlog:drain hook. Drop a server plugin to forward wide events to your observability backend:
// server/plugins/evlog-drain.tsimport { createAxiomDrain } from 'evlog/axiom'export default defineNitroPlugin((nitroApp) => {nitroApp.hooks.hook('evlog:drain', createAxiomDrain())})
Other adapters: evlog/otlp, evlog/hyperdx, evlog/posthog, evlog/sentry, evlog/better-stack, evlog/datadog. Full list at evlog adapters.
Project layout
workflows/reply.ts # "use workflow" — chains generateReply + postReplyserver/api/index.ts # GET /api — health checkwebhooks/sendblue.post.ts # POST /api/webhooks/sendblue — Sendblue inbound webhookplugins/imessage.ts # Chat SDK handlers, queues the workflow on each DMtools/index.ts # tools passed to generateText (use step)utils/agent-steps.ts # generateReply + postReply steps; evlog AI wiring lives hereutils/bot.ts # Chat instance + Sendblue adapter (singleton)nitro.config.ts # registers `workflow/nitro` and `evlog/nitro/v3`
Observability
Two layers, both opt-in through this repo's defaults:
Workflow runs — durable execution, step retries, replay debugging:
pnpm workflow:web # local dashboard with run history, step retries, live logsnpx workflow inspect runs # CLI
In production on Vercel, runs show up automatically in the Vercel dashboard.
Wide events — every webhook + every workflow step emits a structured wide event via evlog. The AI SDK integration captures token usage, tool execution timing, and cost estimation under the ai.* namespace automatically. By default events go to console (pretty in dev, JSON in prod). Configure a drain (Axiom, OTLP, Sentry, …) to ship them to your backend — see "Add a drain" above.
Scripts
pnpm dev # start the Nitro dev serverpnpm build # build for productionpnpm preview # preview the production buildpnpm lint # eslintpnpm test # vitestpnpm typecheck # tsc --noEmit
References
chat-adapter-sendblue— adapter docs- Sendblue docs — API, webhooks, line provisioning
- Sendblue pricing
- Chat SDK docs
- Vercel Workflow — durable execution model
- Workflows and steps
- Vercel AI SDK
- AI Gateway models
- evlog — wide-event logging
- evlog AI SDK integration
- evlog drain adapters
- Nitro
- ngrok
Contributing
Contributions are welcome — see CONTRIBUTING.md for setup, conventions, and how to add a tool/step. By participating you agree to the Code of Conduct. Security issues: see SECURITY.md.
License
Apache 2.0 — Made by @HugoRCD.

