Skip to Content
Advertisement
Technology

How to Build a Trustworthy Robot

A conversation with a robot researcher about a possible future where robots are like teammates in hospitals, factories, and homes

Paro the seal has large dark eyes, long whiskers, and soft antibacterial fur. It’s a robot, but for the elderly dementia patients who use it, it’s more like a cuddly, attentive nurse. Paro can hold a basic conversation using scripted responses, detect emotional tone through voice and facial recognition in the moment, remind users to drink water and take medication, and report worrisome changes to human caregivers.

Featured Video

But what if Paro could also build a model of its companions over time, learning that, for example, Sally, who is 96, gets anxious in the evenings, that she hasn’t mentioned her daughter in two weeks, and that she responds better to jokes than to reassurance? What if Paro could learn that Sally resists reminders unless they’re delivered after a favorite song? In other words, what if the robot could actively collaborate with the patient?

This is what a team of researchers recently proposed in the journal Science Robotics. Robots should be designed to learn with us, they argue, in our hospitals, homes, and warehouses, adapting to human partners and even other robots over time. They should be able to adjust not only to new partners, but also to new roles and cultural environments. They should be, basically, a bit more human-like.

I spoke with Sharmita Dey, a researcher at the Institute of Robotics and Intelligent Systems in Switzerland about why we need such robots, what it would mean for a robot to trust a human, what risks they might present, and how to protect such robots against bias and bad actors.

Why have we built solitary robots instead of collaborative ones until now?

The solitary robots are the starting point. An individual robot by itself has to have some capabilities. And it looked like solitary robots could solve the problem of interaction with the real world. It’s not that these individual robots have failed, but the more they’re deployed, the more we understand their real-world shortcomings.

Why is it so important to have collaborative robots that learn with us?

Collaborative robots are needed because the ultimate goal of robotics is to ease human effort. We can keep on building very smart individual autonomous robots, but if they’re not able to seamlessly collaborate or seamlessly work with humans—to understand the social context, to anticipate what a human wants to do, and to communicate effectively with a human—then these robots aren’t very useful for us. Most real-world settings, such as rehabilitation, or other forms of human assistance, rely on interaction and collaboration.

What are the biggest design challenges right now?

Teaching the robots to learn norms across cultures and contexts, teaching them to generalize to unseen human behaviors, and balancing the adaptations versus safety constraints. But it’s also very difficult to get data on collaboration or interaction in real-world settings. The collection of such data is quite tedious. For large language models, we get a lot of data from the Internet, but physical interaction is something that’s much more complicated to assess.

Read more: “How Human Is Human?

What kinds of tasks would these robots be most useful for?

Constructive ones, such as rehabilitation, elderly care, and companionship.

What would these robots look like? Would they be human-like in their appearance?

I’m advocating for functional equivalence. The closer they are in physical appearance, the better humans would interact with them.

You write in your paper about the importance of “trust” for collaborative robots—teaching these machines how to trust humans and assess a human’s level of trustworthiness. Can you tell me more about how that would work?

It means building a history of interactions. Robots must remember the successes, the failures, and the user’s preferences. Trust isn’t a static property. It’s an emergent function of interaction history.

Should we worry that humans could begin to attribute feelings and intentions to these robots in the same way some have with chatbots?

Yes, especially in old age where people need companions. Companion robots are already conceived for that. People could start attributing emotions to these robots. But whether this is good or bad is something that’s difficult to comment on at the moment. I’d say that it’s good for people who are lonely. It can be bad though if they attribute emotions to a machine and the expectation doesn’t match reality. I’d say that attributing feelings to robots is neither completely negative nor completely positive.

How do you protect against unforeseen risks? For example, chatbots seem like the linguistic equivalent of the collaborative robot. They learn from interactions with humans, only they lack a physical form. I don’t think most people anticipated that chatbots were going to lead people into delusional thinking, encouraging them to do things like attempt murder or jump off a building.

A lot of people are speaking about the risks of AI, but I’d say that AI is what we create AI to be. So are robots. So the objective function we give to the robots is what the robot would do. Of course, sometimes, there is no control, especially in a reinforcement learning environment. The robot learns from its interactions with the environment. We don’t have control over the experiences it has and learns from. I think we could have a lot of safety constraints there. I’d say human supervision over what the robot is learning or what kind of data it is actually consuming is needed to deploy a robot safely into the physical world.

How do you prevent bias from creeping into reinforcement learning contexts, or shaping the social norms that the robots learn?

It’s not fully under our control, of course, but at the same time, we could supervise the kinds of data that are fed into the robot. We would start out teaching the robot through controlled experiences, and only then allow it to deploy in the wild and retrain the model on that. Basically, humans should still be in the loop. Keeping humans in the loop on training and optimization could prevent a lot of these problems. Of course, it won’t completely mitigate it, but it would help.

You mentioned keeping humans in the loop, but humans are fallible. How do you protect against bad actors using these kinds of collaborative robots for social ills, or teaching the robots to manipulate humans?

Different scholars have different positions on whether to make the code for these robots open source, or accessible to everybody. Keeping them gated could prevent some bad actors from using such models for unintended purposes. I’m a supporter of open-source work, but it’s probably important to have restricted access to some of the models.

If something does go wrong in an interaction between a human and a collaborative robot, how do you know who is to blame?

The first assumption would be that the robot has been unable to follow or anticipate the human needs, because the robot is still learning. At least for the beginning, because humans can socialize and collaborate with new partners with some social common sense—which robots lack.

Enjoying Nautilus? Subscribe to our free newsletter.

Lead image: © FSL - Presentazione Foca Robot

Advertisement

Stay in touch

Sign up for our free newsletter

More from Technology

Explore Technology

AI Music vs. My Parents

My folks were taken in by the latest algorithmic “artist,” and it scares me

May 5, 2026

How Video Calling Worked Almost 100 Years Ago

We’ve come a long way since then

April 7, 2026

I Asked Claude Why It Won’t Stop Flattering Me

An interview with Anthropic’s chatbot about sycophantic AI and how to guard against it

April 3, 2026

Making AI More Human

An interview with Berkeley researcher and author Nina Begus about her new book and proposal to fuse science and the humanities

April 1, 2026

What the US Could Learn From Asia’s Robot Revolution

In Korea and Japan, humanoid machines aren’t rivals but partners

March 19, 2026