poss OT: Exactly what is artificial intelligence? Alex Goodwin (26 Feb 2023 18:09 UTC)
Re: [TML] poss OT: Exactly what is artificial intelligence? Phil Pugliese (26 Feb 2023 20:20 UTC)
Re: [TML] poss OT: Exactly what is artificial intelligence? Thomas Jones-Low (26 Feb 2023 20:22 UTC)
Re: poss OT: Exactly what isartificialintelligence? Sterling Blake (27 Feb 2023 00:40 UTC)
Re: [TML] poss OT: Exactly what is artificial intelligence? NotKnown AtThisAddress (27 Feb 2023 12:45 UTC)

poss OT: Exactly what is artificial intelligence? Alex Goodwin 26 Feb 2023 18:09 UTC

Apologies for the possible off-topicness, but it seemed a bit quiet
recently.

Was talking to FreeTrav over on #traveller (channel he admins on
UnderNet) and he did an-IME characteristically FreeTrav thing - rumble
my implicit assumptions and challenge them.

To paraphrase FT, the problem we have with "AI" now is that we don't
know what it _is_, only what it's _not_.  IOW, as soon as a formerly-AI
problem becomes doable, it's no longer AI.

FT further posited that "the AI question" is the compute/biological
equivalent of Godel's Incompleteness Theorem - we can either never
achieve AI, or never know that we've done so.

Your thoughts on that?

As an aside, an artificial _general_ intelligence (AGI) is a subtype of
AI (whatever _that_ works out to be) with similar breadth and depth of
capability as a human (or, given the list, sophont).  Also known as
"strong AI".

An artificial _super_ intelligence (ASI) , by extension, is an AI with
_greater_ breadth and depth of capability than a human.

An artificial _narrow_ intelligence (ANI), is an AI with _lesser_
breadth and depth of capability than a human.

For a comparison of human capabilities (from
https://www.britannica.com/science/information-theory/Physiology ),
_unconscious_ individual-human processing capacity is on the rough order
of 11 million bits/sec and _conscious_ individual-human processing
capacity is on the rough order of 50-60 bits/sec.  I wouldn't expect a
full order-of-magnitude difference in either capacity between different
people.

Upon further thought (thank you, FT), I'm not sure the human brain has a
meaningful analogue to clock rate.  _Individual_ neurons can't fire
faster than roughly 1 kHz, but there's on the order (IIRC) of 100
billion of them, on average each being connected to another 40,000 neurons.

Given those capacities, I _think_ an individual computer would beat a
baseline human for information processing capacity.  How far am I up the
garden path on this one?

Next step along the chat was what was needed to give a pile of
brute-force compute power the stimulus to "wake up" and become
spontaneously intelligent (at AGI or ASI levels), a la the
(post)-cyberpunk trope.

FT wasn't convinced it was a function of available compute power on its
own - I suggested that was necessary, but insufficient. Some other
unknown, "X", factor or factors was/where also necessary.

What could those X factors be?

FT posed the following criteria for (what I think, from context, is AGI
- sapience/sophontry):

- Ability to learn from one's errors;

- Ability to conceive of need for new information;

- Volition to seek out said new information;

- Volition to tell a pestiferous questioner to sod off;

What else, or has FT come up with a reasonably-minimal set?

Alex
--