Friday, Mar 29, 2024 | Last Update : 04:49 PM IST

  Technology   Teaching robots to say No

Teaching robots to say No

WWW.SCIENCEALERT.COM
Published : Nov 27, 2015, 9:18 pm IST
Updated : Nov 27, 2015, 9:18 pm IST

Engineers are trying to train machines using the same ‘Felicity conditions’ that guide humans towards decisions

ROBO.gif
 ROBO.gif

Engineers are trying to train machines using the same ‘Felicity conditions’ that guide humans towards decisions

At first, the news that software engineers are teaching robots to disobey their human masters does sound slightly troubling: should we really allow the artificial intelligence systems of the future to say no to us But once you think it through, you can see why such a feature might actually end up saving your life.

Consider a robot working on a car production line: most of the time, it would simply follow the instructions it’s been given, but if a human should get in harm’s way, the machine needs to be clever enough to stop what it’s doing. It needs to know to override its default programming to put the human at less risk. It’s this kind of functionality that a team from the Human-Robot Interaction Lab at Tufts University is trying to introduce.Taking a call

The team’s work is based around the same ‘Felicity conditions’ that our human brains apply whenever we're asked to do something. Under these conditions, we subconsciously run through a number of considerations before we perform an action: do I know how to do this Can I physically do it, and do it right now Am I obligated based on my social role to do it And finally, does it violate any normative principle to do it

If robots can be conditioned to ask these same questions, they’ll be able to adapt to unexpected circumstances as they occur.

It’s the last two questions that are the most important to the process. A robot needs to decide if the person giving it instructions has the authority to do so, and it needs to work out if the subsequent actions are dangerous to itself or others. It’s not an easy concept to put into practice, as anyone who’s ever watched 2001: A Space Odyssey will know

Researchers developing self-driving cars are also grappling with a similar moral dilemma: if a driverless car is caught in an emergency situation where it threatens the lives of several people outside, does it intentionally crash itself into a wall to kill its single occupant, or should it plough into a crowd of pedestrians and potentially kill many more

“Surveys suggested people generally agreed that the car should minimise the overall death toll, even if that meant killing the occupant — and perhaps the owner — of the vehicle in question," Pete Dockrill reported for ScienceAlert this month.