MIT Team Builds a Speech-to-Reality System that Turns Spoken Prompts into Physical Objects Within Minutes


MIT researchers have devloped a system that allows users to speak a request aloud and receive a fabricated object minutes later, demonstrating how natural language, generative AI, and robotics can combine to produce on-demand manufacturing.

According to MIT, the work, presented at the ACM Symposium on Computational Fabrication, shows that the system can assemble simple furniture and decorative items from modular parts without requiring users to know 3D modeling or robotic programming.

Researchers at MIT’s Center for Bits and Atoms, led by graduate student Alexander Htet Kyaw with collaborators Se Hwan Jeon and Miana Smith, built a workflow that begins with speech recognition and a large language model. The model interprets the user’s request — such as asking for a stool — and passes the result to a 3D generative AI system that produces a digital representation of the object. A voxel-based process then breaks that form into discrete components suitable for robotic assembly. After geometric checks for constraints like overhangs, part count, and structural connectivity, the system generates an assembly sequence and robotic path plan to build the finished object.

“These are rapidly advancing areas of research that haven’t been brought together before in a way that you can actually make physical objects just from a simple speech prompt,” Kyaw said. 


https://theaiinsider.tech/2025/12/07/mit-team-builds-a-speech-to-reality-system-that-turns-spoken-prompts-into-physical-objects-within-minutes/

Comments

Popular posts from this blog

AlphaProteo generates novel proteins for biology and health research

A New Tool for ADHD Management