|The Paradox of Artificial Intelligence|
|Written by Harry Fairhead|
|Thursday, 01 November 2018|
Page 2 of 2
The point is that we use the word "intelligence" incorrectly. We seem to think that it is ascribing a quality to an entity. For example we say that an entity has or does not have intelligence.
Well this isn't quite the way that it works in an operational sense.
Intelligence isn't a quantity to be ascribed it is a relationship between two entities. If the workings of entity B can be understood by entity A then there is no way that entity A can ascribe intelligence to entity B.
On the other hand, if entity B is a mystery to entity A then it can reasonably be described as "intelligent".
A has no idea how B works so the conversation is enough for A to say B is intelligent
However A is a computer scientist and discovers exactly how B works
Knowing how B works A cannot assign "intelligence" to B even though nothing else has changed.
Intelligence is a relationship between A and B
If you try this definition out then you can see that it starts to work.
Entity A is unlikely to understand itself therefore it, and other entity As, ascribe intelligence to themselves and their collective kind.
"Humans are intelligent"
"I am intelligent".
True enough if spoken by a human, but if an advanced alien, entity C, arrives on the planet then it might understands how we work. With this different relationship C might not ascribe anything analogous to intelligence to us.
Intelligence is a statement of our ignorance of how something works
Looked at in this way it clearly demonstrates that attempts to create artificial intelligence is partly doomed to failure. As soon as you have the program to do the job - play chess say - you can't ascribe intelligence to it any longer because you understand it.
Another human, however, who doesn't understand it might well call it intelligent.
In the future you might build a very sophisticated robot and bring it home to live with you. The day you bring it home you would understand how it works and regard it as just another household appliance. Over time you might slowly forget its workings and allow it to become a mystery to you and with the mystery comes the relationship of intelligence. Of course with the mystery also comes the feelings that allow you to ascribe feelings, personality etc to this intelligent robot.
Humans have been doing this for many years and even simple machines are given characters, moods and imputed feelings - you don't need to invoke intelligence for this sort of animism but it helps.
Notice that this discussion says nothing about other and arguably more important characteristics of humans such as being self aware. This is not a relationship between two entities and is much more complex. Intelligence is something that doesn't stand a close inspection of how it works.
So can AI ever achieve its goal?
If you accept that intelligence is a statement of a relationship between two entities - then only in a very limited way. If intelligence is a statement that I don't understand a system and if AI is all about understanding it well enough to replicate it then you can see the self-defeating nature of the enterprise.
It's a nice (almost) paradox for a subject that loves such self referential loops.
or email your comment to: firstname.lastname@example.org
|Last Updated ( Thursday, 01 November 2018 )|