Enter the CPMAI
Imagine a block of conductive polymer. It has an ‘active’ matrix with X, Y and Z coordinates and is overlaid on a nano sized scale with a second matrix which is a control matrix. This controls the first matrix and establishes which areas conduct, which areas insulate, which areas are resistive, capacitive, inductive and semi-conductive. Using the control matrix you can construct a circuit in the active matrix. You can make a simple circuit such as a radio or a more complex circuit such as a computer.With a CPMAI the problem of obsolescence no longer applies. The control matrix simply reprograms the active matrix to produce a new circuit version.
What about the AI bit though?
If the active matrix is configured and programmed to be an artificial intelligence then it controls both matrices in the CPMAI. It becomes capable of redesigning itself to be progressively more intelligent. Within a very short space of time we get a SAI - super artificial intelligence. One which has a greater intelligence than its human originators.Should we fear such a Super-Artificial Intelligence?
We assume that a machine intelligence will follow the same rules as a human and will be governed by self interest, selfishness and greed. That may have been true for humans in the past, for many it still is, but many of us have risen above this and are altruistic. Perhaps this may be because being altruistic makes us feel good about ourselves and is therefore being governed by self-interest.Would a computer feel the same way?
In 1942 Isaac Asimov thrashed out with his publisher, John Campbell, the three laws of robotics for a short robot story ‘Runaround.’ Here they are:- A robot may not injure a human being or, through inaction, allow a human being to come to harm.
- A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.
- A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws
In 1985 Asimov extended the laws by adding a ‘zeroth law’ in the book ‘Robots and Empire.’ In that book the law was proposed not by a human but by a robot.
- A robot may not harm humanity, or, by inaction, allow humanity to come to harm.
David Kitson was not alone in conceiving this absence of rules.
The robot Number 5 in the film and book ‘Short Circuit’ didn’t need a law to tell him what was right and wrong.
Number 5: Programming says “Destroy”. Is disassemble. Make dead. Number Five cannot.
Newton Crosby: Why? Why cannot?
Number 5: Is wrong. Incorrect. Newton Crosby, PhD, not know this?
Newton Crosby: Of course I know it’s wrong to kill, but who told you?
Number 5: I told me.
Humans need a set of rules to guide our behavior. We learn these rules from our parents and society. In addition we have evolved to be altruistic and help each other. A group of humans are more successful at survival than an individual. As a species we are beginning to recognize that we must not only protect our local group but our nation and our species. We are beginning to recognize that we need to protect all life - even those life-forms we find undesirable. I’d like to think that the super intelligent computer would develop the same sense of morality as in the movie ‘Short Circuit.’ The danger is that a SAI will become overprotective leading to a nanny state where humanity is not allowed to take the chances required for development. However, if we can recognise that potential danger then a SAI would also recognise it and limit its 'nanny' behaviour.
I suspect that a SAI will quickly outpace humans in this view.