Will Machines Someday Rule?
By James Donahue
There
is something interesting occurring in our contemporary world.
--The human race is recklessly polluting the air, water and land and using up the planet’s natural resources
so completely that it appears we are heading for an early extinction.
--While this is going on, the development of intelligent robotic machines is reaching a point where the machines
are beginning to think for themselves.
Of
course this leads to an interesting idea. Is it possible that once robots learn to reason and replace broken parts, repair
each other and even build new and better replicas of themselves that they will continue on to rule this planet long after
it is dead?
This is not as crazy an idea
as it may sound at first glance. Louis Del Monte, physicist and author of the book The Artificial Intelligence Revolution,
believes that the day is soon coming when machines will become self-conscious, have the ability to protect themselves, and
may even view humans as “an unpredictable and dangerous species.”
In a recent interview with Dylan Love for an article in Business Insider, Del Monte warned that most humans
will have become part human and part machine within the next century. We will turn ourselves into cyborgs in our quest for
immortality and our need to survive in a deadly environment. This also may be necessary for us to be capable of traveling
in space and exploring other planets in the universe.
Because we are animals, we cannot leave Earth without taking all of the ingredients needed to sustain life along
with us. Machines don’t need food, water and air to survive. A handy oil can, a few tools and a solar energy receiver
may be all they will need to keep going.
That
is the positive scenario. Del Monte also offers a warning that the machines may regard themselves as superior to us, mostly
because their computer wiring will allow them to think and solve problems faster than the human brain. Also, machines will
lack compassion, which will make them great soldiers with the ability to crush any creature that gets in their way.
They “might view us the same way we view harmful insects,” Del Monte warned. He noted that humans
do dangerous things like start wars, build weapons powerful enough to destroy our world, and make computer viruses. The machines
that operate on computer technology would regard the latter as a serious threat.
Robotic machines already are developing artificial intelligence. For instance, Del Monte notes that a human pacemaker
actually uses sensors and a form of artificial intelligence to regulate the human heart.
He said experiments in Lausanne, Switzerland, have shown that robots designed to cooperate in finding beneficial
resources quickly learned to lie to each other in order to hoard the resources for themselves. This is, in a primitive sense,
an early form of self- preservation.
It
seems that in our rush to develop robots to serve industrial and commercial needs, we may be forgetting the famous three laws
of robotics developed by science fiction writer Isaac Asimov. He outlined the laws in the 1942 short story Runaround. They are:
--A robot may not injure a human being, or, through inaction, allow a human
being to come to harm.
--A robot must obey
the orders given to it by human beings, except where such orders would conflict with the first law.
--A robot must protect its own existence as long as such protection
does not conflict with the first or second law.