Giving AI agents eyes, hands — and a voice.
Vision is actively developed with an emphasis on Multimodal AI — using vision-capable LLMs to understand and reproduce UI designs with layout analysis, style extraction, and component generation in a single pipeline.
The MCP (Model Context Protocol) module implements the open standard pioneered by Anthropic, letting AI agents call real-world tools through structured JSON schema interfaces via stdio or HTTP/SSE transports.
- Image → Vision LLM → Layout Analysis → Component Generation pipeline
- MCP server exposing web_search, file_read/write, code_exec, api_bridge tools
- Compatible with Antigravity IDE, Claude Code, Cursor, and Gemini CLI
- CSS keyframe / GSAP / Canvas animation reference showcase
- Full Docker containerization for the antigravity-website module