A will to survive might take AI to the next level

0

Fiction is full of robots with feelings.

Like that emotional kid David, played by Haley
Joel Osment, in the movie A.I. Or WALL•E, who obviously had feelings for
EVE-uh. Robby the Robot sounded pretty emotional whenever warning Will Robinson
of danger. Not to mention all those emotional train-wreck, wackadoodle robots
on Westworld.

But in real life robots have no more feelings
than a rock submerged in novocaine.

There might be a way, though, to give robots
feelings, say neuroscientists Kingson Man and Antonio Damasio. Simply build the
robot with the ability to sense peril to its own existence. It would then have
to develop feelings to guide the behaviors needed to ensure its own survival.

“Today’s robots lack
feelings,” Man and Damasio write in a new paper (subscription
required) in Nature Machine Intelligence. “They are not designed to represent the internal
state of their operations in a way that would permit them to experience that
state in a mental space.”

So Man and Damasio propose a strategy for
imbuing machines (such as robots or humanlike androids) with the “artificial
equivalent of feeling.” At its core, this proposal calls for machines designed to
observe the biological principle of homeostasis. That’s the idea that life must
regulate itself to remain within a narrow range of suitable conditions — like keeping
temperature and chemical balances within the limits of viability. An
intelligent machine’s awareness of analogous features of its internal state
would amount to the robotic version of feelings.

Such feelings would not only motivate
self-preserving behavior, Man and Damasio believe, but also inspire artificial intelligence
to more closely emulate the real thing.

Typical “intelligent” machines are designed to
perform a specific task, like diagnosing diseases, driving a car, playing Go or
winning at Jeopardy! But intelligence in one arena isn’t the same as the
more general humanlike intelligence that can be deployed to cope with all sorts
of situations, even those never before encountered. Researchers have long
sought the secret recipe for making robots smart in a more general way.

In Man and Damasio’s view, feelings are the
missing ingredient.

Feelings arise from the need to survive. When
humans maintain a robot in a viable state (wires all connected, right amount of
electric current, comfy temperature), the robot has no need to worry about its
own self-preservation. So it has no need for feelings — signals that something
is in need of repair.

Feelings motivate
living things to seek optimum states for survival, helping to ensure that
behaviors maintain the necessary homeostatic balance. An intelligent machine
with a sense of its own vulnerability should similarly act in a way that would
minimize threats to its existence.

To perceive
such threats, though, a robot must be designed to understand its own internal
state.

Man and
Damasio, of the University of Southern California, say the prospects for
building machines with feelings have been enhanced by recent developments in
two key research fields: soft robotics and deep learning. Progress in soft
robotics could provide the raw materials for machines with feelings. Deep
learning methods could enable the sophisticated computation needed to translate
those feelings into existence-sustaining behaviors.

Deep learning
is a modern descendant of the old idea of artificial neural networks — sets of
connected computing elements that mimic the nerve cells at work in a living
brain. Inputs into the neural network modify the strengths of the links between
the artificial neurons, enabling the network to detect patterns in the inputs.

Deep
learning requires multiple neural network layers. Patterns in one layer exposed
to external input are passed on to the next layer and then on to the next,
enabling the machine to discern patterns in the patterns. Deep learning can classify
those patterns into categories, identifying objects (like cats) or determining
whether a CT scan reveals signs of cancer or some other malady.

An
intelligent robot, of course, would need to identify lots of features in its
environment, while also keeping track of its own internal condition. By representing
environmental states computationally, a deep learning machine could merge
different inputs into a coherent assessment of its situation. Such a smart
machine, Man and Damasio note, could “bridge
across sensory modalities” — learning, for instance, how lip movements (visual
modality) correspond to vocal sounds (auditory modality).

Similarly, that robot
could relate external situations to its internal conditions — its feelings, if
it had any. Linking external and internal conditions “provides a crucial piece
of the puzzle of how to intertwine a system’s internal homeostatic states with
its external perceptions and behavior,” Man and Damasio note.

Ability to sense
internal states wouldn’t matter much, though, unless the viability of those states
is vulnerable to assaults from the environment. Robots made of metal do not
worry about mosquito bites, paper cuts or indigestion. But if made from proper
soft materials embedded with electronic sensors, a robot could detect such
dangers — say, a cut through its “skin” threatening its innards — and engage a
program to repair the injury.

A robot
capable of perceiving existential risks might learn to devise novel methods for
its protection, instead of relying on preprogrammed solutions.

“Rather than having to hard-code a
robot for every eventuality or equip it with a limited set of behavioral
policies, a robot concerned with its own survival might creatively solve the
challenges that it encounters,” Man and Damasio suspect. “Basic goals and
values would be organically discovered, rather than being extrinsically
designed.”

Devising novel
self-protection capabilities might also lead to enhanced
thinking skills. Man and Damasio believe advanced human thought may have
developed in that way: Maintaining viable internal states (homeostasis)
required the evolution of better brain power. “We regard high-level
cognition as an outgrowth of resources that originated to solve the ancient biological
problem of homeostasis,” Man and Damasio write.

Protecting
its own existence might therefore be just the motivation a robot needs to
eventually emulate human general intelligence. That motivation is reminiscent
of Isaac Asimov’s famous laws of robotics: Robots must protect humans,
robots must obey humans, robots must protect themselves. In Asimov’s fiction,
self-protection was subordinate to the first two laws. In real-life future
robots, then, some precautions might be needed to protect people from
self-protecting robots.

“Stories about robots
often end poorly for their human creators,” Man and Damasio acknowledge. But
would a supersmart robot (with feelings) really pose Terminator-type dangers?
“We suggest not,” they say, “provided, for example, that in addition to having
access to its own feelings, it would be able to know about the feelings of
others — that is, if it would be endowed with empathy.”

And so Man and Damasio
suggest their own rules for robots: 1. Feel good. 2. Feel empathy.

“Assuming a robot
already capable of genuine feeling, an obligatory link between its feelings and
those of others would result in its ethical and sociable behavior,” the
neuroscientists contend.

That
might just seem a bit optimistic. But if it’s possible, maybe there’s hope for
a better future. If scientists do succeed in instilling empathy in robots,
maybe that would suggest a way for doing it in humans, too.

Click on a tab to select how you'd like to leave your comment

Leave a Reply

Your email address will not be published. Required fields are marked *