MIT Researchers Develop Faster Way to Train General-Purpose Robots

Researchers at MIT have developed a new technique to train general-purpose robots, inspired by large language models like GPT-4. The method combines diverse data from multiple sources into one system that can teach any robot a wide range of tasks.

This approach can be faster and less expensive than traditional techniques because it requires far fewer task-specific data. Led by Lirui Wang, an electrical engineering and computer science graduate student, the team developed a new architecture called Heterogeneous Pretrained Transformers (HPT) that unifies data from varied modalities and domains.

The researchers drew inspiration from large language models like GPT-4, which are pretrained using an enormous amount of diverse language data and then fine-tuned by feeding them a small amount of task-specific data. Companies involved in the work include Amazon and Toyota Research Institute. The technique has shown promising results, outperforming training from scratch by more than 20 percent in simulation and real-world experiments.

A Novel Approach to Training General-Purpose Robots

Inspired by large language models, researchers at MIT have developed a training technique that pools diverse data to teach robots new skills. This approach has the potential to revolutionize the field of robotics by enabling robots to learn from a vast amount of data and adapt to new tasks and environments with ease.

The Limitations of Current Robotic Policies

Current robotic policies are typically trained using imitation learning, where a human demonstrates actions or teleoperates a robot to generate data. However, this method has several limitations. Firstly, it requires a large amount of task-specific data, which can be time-consuming and expensive to collect. Secondly, robots often fail when their environment or task changes, as they are not able to adapt to new situations.

Drawing Inspiration from Large Language Models

To develop a better approach, the researchers drew inspiration from large language models like GPT-4. These models are pretrained using an enormous amount of diverse language data and then fine-tuned by feeding them a small amount of task-specific data. This pretraining on a vast amount of data helps the models adapt to perform well on a variety of tasks.

The MIT researchers developed a new architecture called Heterogeneous Pretrained Transformers (HPT) that unifies data from varied modalities and domains. They put a machine-learning model known as a transformer into the middle of their architecture, which processes vision and proprioception inputs. A transformer is the same type of model that forms the backbone of large language models.

The researchers align data from vision and proprioception into the same type of input, called a token, which the transformer can process. Each input is represented with the same fixed number of tokens. Then the transformer maps all inputs into one shared space, growing into a huge, pretrained model as it processes and learns from more data.

Enabling Dexterous Motions

One of the biggest challenges of developing HPT was building the massive dataset to pretrain the transformer, which included 52 datasets with more than 200,000 robot trajectories in four categories. The researchers also needed to develop an efficient way to turn raw proprioception signals from an array of sensors into data the transformer could handle.

Results and Future Directions

When they tested HPT, it improved robot performance by more than 20 percent on simulation and real-world tasks, compared with training from scratch each time. Even when the task was very different from the pretraining data, HPT still improved performance.

In the future, the researchers want to study how data diversity could boost the performance of HPT. They also want to enhance HPT so it can process unlabeled data like GPT-4 and other large language models. The ultimate goal is to have a universal robot brain that you could download and use for your robot without any training at all.

This work was funded, in part, by the Amazon Greater Boston Tech Initiative and the Toyota Research Institute.

More information
External Link: Click Here For More
Quantum News

Quantum News

There is so much happening right now in the field of technology, whether AI or the march of robots. Adrian is an expert on how technology can be transformative, especially frontier technologies. But Quantum occupies a special space. Quite literally a special space. A Hilbert space infact, haha! Here I try to provide some of the news that is considered breaking news in the Quantum Computing and Quantum tech space.

Latest Posts by Quantum News:

AWS Quantum Technologies Blog: New QGCA Outperforms Simulated Annealing on Complex Optimization Problems

AWS Quantum Technologies Blog: New QGCA Outperforms Simulated Annealing on Complex Optimization Problems

February 23, 2026
AWS Quantum Technologies has released version 0.11 of the Qiskit-Braket provider on February 20, 2026, significantly enhancing how users access and utilize Amazon Braket’s quantum computing services through the popular Qiskit framework. This update introduces new “BraketEstimator” and “BraketSampler” primitives, mirroring Qiskit routines for improved performance and feature integration with Amazon Braket program sets. Importantly, the provider now fully supports Qiskit 2.0 while maintaining compatibility with versions as far back as v0.34.2, allowing users to “use a richer set of tools for executing quantum programs on Amazon Braket.” The release unlocks flexible compilation features, enabling circuits to be compiled directly for Braket devices using the to_braket function, accepting inputs from Qiskit, Braket, and OpenQASM3.

AWS Quantum Technologies Releases Qiskit-Braket Provider v0.11, Now Compatible with Qiskit 2.0

February 23, 2026
Microsoft Research Details 10,000-Year Data Storage Breakthrough in Nature

Microsoft Research Details 10,000-Year Data Storage Breakthrough in Nature

February 23, 2026