Skip to main content
SINT Avatars is the voice-first, visual-first surface for agent interaction. A 3D WebGL avatar with voice input and output, natural-language rewriting, and contextual widget spawning. JARVIS meets Grok.

3D avatar

React Three Fiber, ARKit blendshapes, 56 Mixamo animations. Expressive and performant.

Voice I/O

Web Speech API input, ElevenLabs TTS output. Zero external dependencies for the input path.

18 widget types

Status, code, metric, progress, chart, terminal, table, diff, action, and more — spawned contextually during conversation.

The Conversation Compiler

A small neural rewriter (gpt-4.1-nano or comparable) that sits between the agent and the TTS pipeline. Input: raw agent output (markdown, JSON, technical text) Output: spoken-language version plus emotion, animation, and widget metadata

Example

Agent output:
Deployed to Railway. Build 47 passed. 3 of 5 tests green.
After Conversation Compiler:
  • Spoken: “Deploy’s live. Build passed, three of five tests green.”
  • Emotion: smile
  • Animation: Head-Nod-Yes
  • Widgets: [status: "Deploy" → success], [progress: "Tests" → 60%]
The Compiler makes agents sound like people instead of reading JSON.

Widget system

18 contextual widget types spawn during conversation:
WidgetPurpose
statusSuccess / error / pending indicator
codeSyntax-highlighted code block
metricKPI with trend arrow
progressProgress bar with percentage
listOrdered or unordered list
chartBar, line, or pie chart
terminalTerminal output display
linkClickable link
tableData table with headers
diffCode diff (additions/deletions)
actionActionable button (approve, run, deny)
imageImage with caption
githubLive GitHub repo status
dashboardFull dashboard with KPIs
goalsAgent goals/objectives
tasksTask list with status
agentsActive agents display
activityActivity timeline
Widgets appear next to the avatar’s speech, stay for a configurable duration, and can be interactive (approve/deny buttons route back through the Gateway).

Governance integration

Avatars reacts to policy events from the SINT Protocol gateway. Expression and widget choice depend on the tier and outcome.
EventExpressionAnimationWidget
T0 approvedefault
T1 approvedefaultHead-Nod-YesStatus (3s)
T2 denythinkingThoughtful-Head-ShakeStatus (8s)
T3 escalatesurprisedThinkingApproval action (30s)

Hardware

Browser

WebGL 2.0, modern browser. Works on laptops, tablets, and phones. The avatar scales to the viewport.

Kiosk / desktop

Electron wrapper for dedicated operator kiosks. Voice wake word, always-on microphone, full-screen avatar mode.

Deployment

# Clone avatars repo
git clone https://github.com/sint-ai/sint-avatars.git
cd sint-avatars
pnpm install
pnpm dev
# → http://localhost:3005
Avatars connects to the SINT Protocol gateway and OpenClaw runtime via environment-configured URLs. All URLs through .env.

SINT OS

How Avatars composes with the rest of the stack.

GitHub repo

Source, build instructions, contribution guide.