How Do Robots Learn to Ask for Help?


In the developing world of robotics, a groundbreaking collaboration between Princeton University and Google stands out.

Engineers at these prestigious institutions have developed an innovative method that teaches robots a crucial skill:

knowing when they need help and how to ask for it. This development marks a significant advance in robotics and closes the gap between autonomous functioning and human-robot interaction.

The journey towards smarter and more autonomous robots has always been hindered by a significant challenge: the complexity and ambiguity of human language.

Unlike the binary clarity of computer codes, human language is full of nuances and subtleties, making it a maze for robots.

For example, a command as simple as “lift bowl” can become a complex task when multiple bowls are present. Robots equipped to perceive their environment and respond to language often find themselves at a crossroads when faced with such linguistic ambiguities.

Measuring Uncertainty

To overcome this challenge, the Princeton and Google team developed a new approach to measuring the ‘fuzziness’ of human language.

This technique essentially measures the level of ambiguity in language commands and uses this measurement to guide robot actions.

In situations where a command may lead to multiple interpretations, the robot can now gauge the level of uncertainty and decide when to ask for further clarification.

For example, in an environment with multiple bowls, a higher degree of uncertainty will cause the robot to ask which bowl to pick up, thus preventing potential errors or inefficiencies.

This approach not only gives robots the ability to better understand language, but also increases their safety and efficiency in executing tasks.

By integrating large language models (LLMs), such as those behind ChatGPT, researchers have taken an important step in aligning robotic actions more closely with human expectations and needs.

The Role of Large Language Models

The integration of Masters plays a very important role in this new approach. Masters are effective in processing and interpreting human language.

In this context, they are used to evaluate and measure the current uncertainty in language commands given to robots.

But relying on Master’s degrees is not without its challenges. As the research team noted, the outputs from the Master’s can sometimes be unreliable.

Anirudha Majumdar, an assistant professor at Princeton, emphasizes the importance of this balance:

“Blindly following plans created by a Master’s can cause robots to act in an unsafe or unreliable way, and that’s why we need our Master’s-based robots to know what they don’t know.”

This highlights the need for a nuanced approach in which LLMs are used as guidance tools rather than infallible decision-makers.

Practical Application and Tests

The practicality of this method has been tested in various scenarios, demonstrating its versatility and effectiveness.

One such test involved a robotic arm tasked with sorting toy food items into different categories. This simple setup demonstrated the robot’s ability to effectively direct tasks with clear choices.

The complexity increased significantly in another experiment using a robotic arm mounted on a wheeled platform in an office kitchen. Here, the robot faced real-world challenges, such as determining the correct item to place in the microwave when presented with multiple options.

Through these tests, robots successfully demonstrated their ability to use measured uncertainty to make decisions or seek explanations, thus confirming the practical utility of this method.

Future Impacts and Research

Looking ahead, the implications of this research extend far beyond current practices. The team, led by Majumdar and graduate student Allen Ren, is exploring how this approach can be applied to more complex problems in robot perception and artificial intelligence.

This includes scenarios where robots must combine vision and language information to make decisions, further closing the gap between robotic understanding and human interaction.

Ongoing research aims to improve robots’ ability to not only perform tasks with greater accuracy, but also navigate the world with an understanding similar to human cognition. This research could pave the way for robots that are not only more efficient and safe, but also more compatible with the subtle demands of the human environment.


Please enter your comment!
Please enter your name here

Share post:



More like this

Artificial Intelligence Tools That Can Be Used in E-Export

In the "ChatGPT and Artificial Intelligence Tools in E-Export"...

What are SMART goals, why are they needed and how to set them correctly

In the modern world, where everyone strives to achieve...

How and why the United States is developing a lunar economy

The United States is seriously thinking about developing an...

China faces problem of untreatable gonorrhea

In China, there are a growing number of strains...