Saturday, October 27, 2018

The future of computers - Conductive polymer matrix for AI (CPMAI)

The problem with existing computers is that they constantly develop. For most users by the time they have saved up to buy the latest computer, technology has advanced and the computer rapidly becomes obsolete with faster more capable machines being available.

Enter the CPMAI

Imagine a block of conductive polymer. It has an ‘active’ matrix with X, Y and Z coordinates and is overlaid on a nano sized scale with a second matrix which is a control matrix. This controls the first matrix and establishes which areas conduct, which areas insulate, which areas are resistive, capacitive, inductive and semi-conductive. Using the control matrix you can construct a circuit in the active matrix. You can make a simple circuit such as a radio or a more complex circuit such as a computer.

 With a CPMAI the problem of obsolescence no longer applies. The control matrix simply reprograms the active matrix to produce a new circuit version.

What about the AI bit though?

If the active matrix is configured and programmed to be an artificial intelligence then it controls both matrices in the CPMAI. It becomes capable of redesigning itself to be progressively more intelligent. Within a very short space of time we get a SAI - super artificial intelligence. One which has a greater intelligence than its human originators.

Should we fear such a Super-Artificial Intelligence?

We assume that a machine intelligence will follow the same rules as a human and will be governed by self interest, selfishness and greed. That may have been true for humans in the past, for many it still is, but many of us have risen above this and are altruistic. Perhaps this may be because being altruistic makes us feel good about ourselves and is therefore being governed by self-interest.

Would a computer feel the same way? 

In 1942 Isaac Asimov thrashed out with his publisher, John Campbell, the three laws of robotics for a short robot story ‘Runaround.’ Here they are:

  1. A robot may not injure a human being or, through inaction, allow a human being to come to harm.
  2. A robot must obey the orders given it by human beings except where such orders would conflict with the First Law.
  3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Laws

In 1985 Asimov extended the laws by adding a ‘zeroth law’ in the book ‘Robots and Empire.’ In that book the law was proposed not by a human but by a robot.
  1. A robot may not harm humanity, or, by inaction, allow humanity to come to harm.
Had Asimov lived, he died in 1992, I have no doubt that a further law would have been conceived which would have replaced all the robot laws. It was left to author David Kitson to conceive this ‘Nihilist’ law in his book ‘Turing Evolved.’
David Kitson was not alone in conceiving this absence of rules.

The robot Number 5 in the film and book ‘Short Circuit’ didn’t need a law to tell him what was right and wrong.

Number 5: Programming says “Destroy”. Is disassemble. Make dead. Number Five cannot.
Newton Crosby: Why? Why cannot?
Number 5: Is wrong. Incorrect. Newton Crosby, PhD, not know this?
Newton Crosby: Of course I know it’s wrong to kill, but who told you?
Number 5: I told me.

Humans need a set of rules to guide our behavior. We learn these rules from our parents and society. In addition we have evolved to be altruistic and help each other. A group of humans are more successful at survival than an individual. As a species we are beginning to recognize that we must not only protect our local group but our nation and our species. We are beginning to recognize that we need to protect all life - even those life-forms we find undesirable. I’d like to think that the super intelligent computer would develop the same sense of morality as in the movie ‘Short Circuit’

I suspect that a SAI will quickly outpace humans in this view.

No comments: