







One API to call any model, render video, and store assets — at agent scale.
An agent can build an entire app from a prompt, deploy it, and fix bugs autonomously.
But when it needs a product video, a voiceover, or a social media clip – it hits a wall.





Every provider specializes deeply in one capability. There is no single player that connects them into a finished video.
6 steps · 14+ models · 6 providers · No workflow to connect them.
“Enterprise deployments use a median of 14 different models.”
















Powered by our own declarative programming language for video.
talking-head-varg-promo.tsx
Rendered output

Video, image, speech, music, lipsync, captions. All providers, one interface. Built on top of fal.ai inference infra.

Transitions, captions, audio mixing, final MP4 assembly – all in the cloud. No local FFmpeg needed.

Same generation = free. Identical props hit cache instantly. The more you use it, the cheaper it gets.

Apache 2.0. Fork, customize, self-host. Own your workflow, don’t rent it.

Automatic dependency graph across models. Image generates first, animates into video, then lipsync layers on top.

Skills for Claude Code, Cursor, Windsurf. Plus a visual app as playground – but built agentic-first.








Backed by angels & investors from

$1M+ ARR
1K active developers
10K daily workflow renders
5K+ GitHub stars

$10M+ ARR
100+ enterprise customers
Default orchestration layer for AI agents
SOTA LLM model for creative work
Every AI agent that creates media runs through VARG. Every workflow, every model, every render.
$180M+ ARR opportunity