New Project

Durable iMessage AI Agent

Build a durable iMessage AI agent with Chat SDK.

DeployView Demo

nitro-imessage-agent

A durable iMessage AI agent built on:

  • Nitro v3 — the API server
  • Chat SDK + chat-adapter-sendblue — message routing over Sendblue
  • Vercel AI SDK + AI Gateway — LLM replies, swap models with one constant
  • Vercel Workflow — durable orchestration with retryable "use step" units
  • evlog — structured wide-event logging with first-class AI SDK integration (token usage, tool calls, cost estimation)

A user texts your Sendblue number, Sendblue posts a webhook to this server, the Chat SDK dispatches the message, a workflow runs an agent step (LLM + tools), and the final reply is sent back through Sendblue. Each step is retryable on its own, so a transient LLM error or send hiccup never drops the inbound message.

┌─────────┐ ┌──────────────┐ ┌──────────────┐ ┌────────────┐
│ user │───▶│ Sendblue │───▶│ POST /api/ │──▶│ Chat SDK │
│iMessage │ │ (cloud) │ │webhooks/... │ │ onMention │
└─────────┘ └──────────────┘ └──────────────┘ └─────┬──────┘
▲ │
│ ▼
│ ┌────────────────────────┐
│ │ workflow start(...) │
│ └───────────┬────────────┘
│ ▼
│ ┌────────────────────────┐
│ │ generateReply(use step)│
│ │ ├── LLM (AI Gateway) │
│ │ ├── tools roundtrip │
│ │ └── evlog wide event │
│ └───────────┬────────────┘
│ ▼
│ ┌────────────────────────┐
└───────────────────────────────────────────┤ postReply (use step) │
Sendblue postMessage │ sendblue.postMessage │
└────────────────────────┘

Architecture

The Sendblue cloud holds your dedicated phone line and forwards inbound iMessages to your server as HTTPS webhooks. Outbound replies go back through the same API. There is no gateway listener to keep alive, no Mac in production, and no cron.

The server/api/webhooks/sendblue.post.ts [blocked] route receives every webhook and hands it to chat.webhooks.sendblue(request). The Chat SDK then fires onNewMention (first DM in a thread) or onSubscribedMessage (every following DM) on the bot — handlers registered in server/plugins/imessage.ts [blocked] call start(replyToMessage, [thread.id, message.text]) to queue a workflow.

workflows/reply.ts [blocked] is a thin "use workflow" function that just chains two retryable steps from server/utils/agent-steps.ts [blocked]:

  1. generateReply step — calls generateText against the Vercel AI Gateway. Tools registered in server/tools/ [blocked] are looped with stopWhen: stepCountIs(5) (LLM → tool → LLM until done). The model is wrapped with evlog/ai's ai.wrap() and experimental_telemetry.integrations carries createEvlogIntegration(ai) so token usage, tool execution timing, and estimated cost are captured into a wide event.
  2. postReply step — calls sendblue.postMessage. Independent retryability: a transient send error doesn't re-run the LLM call.

Workflow functions can't import Node-only packages like evlog directly, so the AI SDK calls live in steps (which run as normal Node) — that's why the actual logic is in server/utils/agent-steps.ts and the workflow file just orchestrates.

Local setup (development)

Sendblue is webhook-based, so to receive iMessages on your local machine you expose localhost:3000 through a public tunnel — we use ngrok.

Prerequisites

  • Node 20+ (use corepack enable to get pnpm)
  • A Sendblue account (sendblue.com). The free tier with a shared line is enough to demo — webhooks work, replies to verified contacts work, and you get 10 contact slots. The AI Agent plan ($100/mo) is required if you want a dedicated number with higher inbound volume; see pricing.
  • ngrok installed and authenticated with a free account (ngrok config add-authtoken <your-token>). Any HTTPS tunnel works — cloudflared tunnel, localtunnel, etc. — ngrok is just what we use here.

Install & run

pnpm install
cp .env.example .env
# fill in AI_GATEWAY_API_KEY + SENDBLUE_API_KEY + SENDBLUE_API_SECRET +
# SENDBLUE_FROM_NUMBER + SENDBLUE_WEBHOOK_SECRET
pnpm dev

In a second terminal:

ngrok http 3000

Copy the https://<id>.ngrok-free.app URL ngrok prints, then in the Sendblue dashboard set the inbound webhook URL to:

https://<id>.ngrok-free.app/api/webhooks/sendblue

Test

Text your Sendblue number from any phone. The dev server logs the inbound webhook, the workflow run starts, and a reply lands on your phone. Try "what time is it in Paris?" to validate the getCurrentTime tool path.

Production setup

Sendblue runs in the cloud and just talks HTTP, so production is the same as local minus the tunnel:

  1. Deploy this repo to Vercel (or any Node host that supports Nitro): pnpm dlx vercel.

  2. Set the same five env vars in your hosting provider's environment settings (Vercel → Project → Settings → Environment Variables).

  3. In the Sendblue dashboard, point the inbound webhook URL at your production deployment:

    https://<your-app>.vercel.app/api/webhooks/sendblue

That's it — no vercel.json, no cron, no Mac. The Vercel function spins up on each webhook and the Workflow runs durably in the background.

Why Sendblue and not Photon? Photon's self-serve Spectrum dashboard hands out credentials for the new spectrum-ts SDK, which isn't compatible with chat-adapter-imessage — the latter still uses Photon's older Enterprise SDK and needs negotiated credentials from Photon sales. Sendblue is self-serve, supports webhooks on the free tier, has SMS fallback, and ships US numbers without A2P KYC. Swap the adapter if your needs differ — chat-adapter-imessage (Photon Enterprise), chat-adapter-blooio, or any future Chat SDK iMessage adapter all plug in the same way.

Configuration reference

Env varRequiredWhen
AI_GATEWAY_API_KEYyesAlways. Auto-detected by the AI SDK (no NITRO_ prefix).
SENDBLUE_API_KEYyesAlways. Auto-detected by the adapter from process env.
SENDBLUE_API_SECRETyesAlways. Auto-detected by the adapter.
SENDBLUE_FROM_NUMBERyesYour provisioned Sendblue line in E.164 format (e.g. +14155551234).
SENDBLUE_WEBHOOK_SECRETrecommendedIf set in the Sendblue dashboard, the adapter validates every incoming webhook against it. Set the same value here.

Switching the model

Edit one constant in server/utils/agent-steps.ts [blocked]:

const MODEL = 'google/gemini-3-flash'

Any supported AI Gateway slug works. A few useful ones:

  • google/gemini-3-flash
  • anthropic/claude-sonnet-4.5
  • openai/gpt-4o-mini
  • xai/grok-4

The AI SDK reads AI_GATEWAY_API_KEY from the environment automatically, so no provider plumbing is needed.

Extending the agent

Add a tool

Tools live in server/tools/index.ts [blocked]. Each tool is a description + inputSchema (zod) + execute function. The "use step" directive on the execute body makes every tool call a retryable, observable workflow step.

Example (from this repo):

import { z } from 'zod'
// eslint-disable-next-line require-await
async function getCurrentTime({ timezone }: { timezone: string }): Promise<string> {
'use step'
return new Intl.DateTimeFormat('en-US', {
timeZone: timezone,
dateStyle: 'full',
timeStyle: 'long',
}).format(new Date())
}
export const tools = {
getCurrentTime: {
description: 'Get the current date and time in a specific IANA timezone…',
inputSchema: z.object({
timezone: z.string().describe('IANA timezone identifier'),
}),
execute: getCurrentTime,
},
}

To add another tool, write a new step function and register it in the tools map. The agent picks it up automatically — generateText discovers them at call time.

Change the system prompt

Edit SYSTEM_PROMPT in server/utils/agent-steps.ts [blocked].

Add a workflow step

Any function annotated with "use step" becomes a retryable, durable step. Wrap higher-level orchestration in a "use workflow" function and call steps from it. See Workflows and steps for the full mental model. workflows/reply.ts [blocked] is the canonical example: a thin "use workflow" function chaining generateReply and postReply steps from server/utils/agent-steps.ts [blocked].

Workflow functions can't import Node-only modules (evlog, native bindings, etc.). Keep heavy logic in "use step" files outside workflows/ and import them into the workflow.

Tweak AI observability

createAILogger(log, { cost }) in server/utils/agent-steps.ts [blocked] controls token-cost estimation. Update COST_MAP with your real Gateway pricing, or set cost to undefined to disable. The wide event under the ai.* namespace already includes inputTokens, outputTokens, toolCalls, tools[] (timing per tool from createEvlogIntegration), msToFirstChunk, tokensPerSecond, and estimatedCost. See evlog AI SDK docs for the full field list.

Add a drain (Axiom, OTLP, Sentry, …)

The Nitro evlog module exposes a evlog:drain hook. Drop a server plugin to forward wide events to your observability backend:

// server/plugins/evlog-drain.ts
import { createAxiomDrain } from 'evlog/axiom'
export default defineNitroPlugin((nitroApp) => {
nitroApp.hooks.hook('evlog:drain', createAxiomDrain())
})

Other adapters: evlog/otlp, evlog/hyperdx, evlog/posthog, evlog/sentry, evlog/better-stack, evlog/datadog. Full list at evlog adapters.

Project layout

workflows/
reply.ts # "use workflow" — chains generateReply + postReply
server/
api/
index.ts # GET /api — health check
webhooks/sendblue.post.ts # POST /api/webhooks/sendblue — Sendblue inbound webhook
plugins/imessage.ts # Chat SDK handlers, queues the workflow on each DM
tools/index.ts # tools passed to generateText (use step)
utils/agent-steps.ts # generateReply + postReply steps; evlog AI wiring lives here
utils/bot.ts # Chat instance + Sendblue adapter (singleton)
nitro.config.ts # registers `workflow/nitro` and `evlog/nitro/v3`

Observability

Two layers, both opt-in through this repo's defaults:

Workflow runs — durable execution, step retries, replay debugging:

pnpm workflow:web # local dashboard with run history, step retries, live logs
npx workflow inspect runs # CLI

In production on Vercel, runs show up automatically in the Vercel dashboard.

Wide events — every webhook + every workflow step emits a structured wide event via evlog. The AI SDK integration captures token usage, tool execution timing, and cost estimation under the ai.* namespace automatically. By default events go to console (pretty in dev, JSON in prod). Configure a drain (Axiom, OTLP, Sentry, …) to ship them to your backend — see "Add a drain" above.

Scripts

pnpm dev # start the Nitro dev server
pnpm build # build for production
pnpm preview # preview the production build
pnpm lint # eslint
pnpm test # vitest
pnpm typecheck # tsc --noEmit

References

  • chat-adapter-sendblue — adapter docs
  • Sendblue docs — API, webhooks, line provisioning
  • Sendblue pricing
  • Chat SDK docs
  • Vercel Workflow — durable execution model
  • Workflows and steps
  • Vercel AI SDK
  • AI Gateway models
  • evlog — wide-event logging
  • evlog AI SDK integration
  • evlog drain adapters
  • Nitro
  • ngrok

Contributing

Contributions are welcome — see CONTRIBUTING.md for setup, conventions, and how to add a tool/step. By participating you agree to the Code of Conduct. Security issues: see SECURITY.md.

License

Apache 2.0 — Made by @HugoRCD.

GitHub Repovercel-labs/nitro-imessage-agent-template
Use Cases
AI
Stack
Nitro

Related Templates

Chat SDK Community Agent

Open source AI-powered Slack community management bot with a built-in Next.js admin panel. Uses Chat SDK, AI SDK, and Vercel Workflow.
Chat SDK Community Agent thumbnail

Chat SDK Knowledge Agent

Open source file-system and knowledge based agent template. Build AI agents that stay up to date with your knowledge base.
Chat SDK Knowledge Agent thumbnail
DeployView Demo