Lying Robots

Gizmodo contributor George Dvorsky interviewed the authors of Robot Ethics 2.0: From Autonomous Cars to Artificial Intelligence and they discuss why we might want to consider programming robots to lie to us:

Gizmodo: How can we program a robot to be an effective deceiver?

Bridewell: There are several capacities necessary for recognizing or engaging in deceptive activities, and we focus on three. The first of these is a representational theory of mind, which involves the ability to represent and reason about the beliefs and goals of yourself and others. For example, when buying a car, you might notice that it has high mileage and could be nearly worn out. The salesperson might say, “Sure, this car has high mileage, but that means it’s going to last a long time!” To detect the lie, you need to represent not only your own belief, but also the salesperson’s corresponding (true) belief that high mileage is a bad sign.

Of course, it may be the case that the salesperson really believes what she says. In that case, you would represent her as having a false belief. Since we lack direct access to other people’s beliefs and goals, the distinction between a lie and a false belief can be subtle. However, if we know someone’s motives, we can infer the relatively likelihood that they are lying or expressing a false belief. So, the second capacity a robot would need is to represent “ulterior motives.” The third capacity addresses the question, “Ulterior to what?” These motives need to be contrasted with “standing norms,” which are basic injunctions that guide our behavior and include maxims like “be truthful” or “be polite.” In this context, ulterior motives are goals that can override standing norms and open the door to deceptive speech.

Maybe robots should start reading fiction too.

“The narrative is a bug, not a feature.”

One reason we easily dismiss the astonishing things computers can do is that we know that they don’t carry around a narrative, a play by play, the noise in their head that’s actually (in our view) ‘intelligence.’

It turns out, though, that the narrative is a bug, not a feature. That narrative doesn’t help us perform better, it actually makes us less intelligent. Any athlete or world-class performer (in debate, dance or dungeonmastering) will tell you that they do their best work when they are so engaged that the narrative disappears.

Seth Godin