82 Just ordinAry robots
Interview with Kerstin Dautenhahn (Professor of Articial
Intelligence, University of Hertfordshire, United Kingdom)
“I don’t need a robot to respond to my every smile and frown.”
“Domestic and care robots are getting better and better at human-
like communication. But will they ever really get the hang of it? And
perhaps more importantly, why should we want them to in the rst
place?” asks Professor Kerstin Dautenhahn.
e skills of social robots have advanced a great deal. Ten years
ago, they could only tell people apart who looked directly into the
camera under good lighting conditions. Nowadays, face recognition
software is much more exible.
Speech recognition too has advanced, though it still has serious
limitations. Typically, it has to be trained by the individual user
before it will do a good job, and then only in certain domains, such
as dictation of letters. It can’t do this important thing that we do all
the time: guess what the other person is saying. Listening is much
more than guring out how phonemes combine into words and sen-
tences. It’s also about prediction and anticipation, based on who is
talking and about what. We’ll see more progress, I’m sure, but we
tend to underestimate how hard it is. e AI pioneers in the 1950s,
such as Marvin Minsky, thought they could solve ‘the speech prob-
lem’ in two months. I don’t expect we will ever solve it, because
it covers the whole of AI: cognition, intelligence, personality, the
question of embodiment, theory of mind. Human language cannot
be interpreted on the basis of the words alone. ings such as meta-
phor, humor and irony also enter into it. e fact that highly intelli-
gent people with autism have trouble understanding these illustrates
how complicated they are.
Another skill that robots have become better at is interpreting the
facial expressions that come with strong emotions. However, in daily
life, we don’t experience much ecstatic happiness or deep sadness. We
just smile, or we are somewhat annoyed. For a robotic personal assis-
tant to be useful, it should be able to recognize the subtler emotions.
Which implies that it should also notice our scratching our heads or
rubbing our chins, for human communication is multimodal. It’s not
just the face, not just the posture, not just the words or the voice or
the gaze, it’s all of that and more. Robots have not yet integrated all