Hi HN!
I built a custom MCP (Model Context Protocol) server that connects Blender to LLMs like ChatGPT, Claude, and any other llm supporting tool calling and mcps, enabling the AI to understand and control 3D scenes using natural language.
You can describe an entire environment like:
> “Create a small village with 5 huts arranged around a central bonfire, add a river flowing on the left, place a wooden bridge across it, and scatter trees randomly.”
And the system parses that, reasons about the scene, and builds it inside Blender — no manual modeling or scripting needed.
What it can do: - Generate multi-object scenes like villages, landscapes, from a single prompt - Understand spatial relations — e.g., “place the bridge over the river” or “add trees behind the huts” - Create camera animations and lighting setups: “orbit around the scene at sunset lighting” - Respond to iterative changes like: “replace all huts with stone houses” or “make the river narrower” - Maintain object hierarchy and labels for later editing
Tech Stack: - Blender Python scripting - Node.js server running MCP - LLM backend (OpenAI / Claude, easily swappable)
Demo: https://blender-mcp-psi.vercel.app/
GitHub: https://github.com/pranav-deshmukh/blender-mcp-demo/
Curious to hear thoughts from folks in 3D tooling, AI-assisted design, or dev interface design. Would you find this useful as a Blender plugin? I’m open to expanding it!
Please try it and give it a star on github
Comments URL: https://news.ycombinator.com/item?id=44622374
Points: 24
# Comments: 7
Connectez-vous pour ajouter un commentaire
Autres messages de ce groupe
Article URL: https://phys.org/news/2025-07-ballistics-explosive-squirting-cucumber.html
Comments URL
Article URL: https://www.science.org/doi/10.1126/sciadv.adu6897


Article URL: https://3quarksdaily.com/3quarksdaily/2025/
Article URL: https://www.romanroads.org/index.html
Comments URL: https://news.yco
