Discussion about this post

User's avatar
jabster's avatar

It seems to me that AI is good at the following:

1) Problems requiring many calculations and evaluations of possible rule-defined scenarios. When the rules get fuzzy to nonexistent, then AI has more difficulty.

2) Problems requiring the aggregation of mass quantities of information to produce a result based on specific conditions or rules. Again, same issue with fuzzy-to-nonexistent rules.

3) Creative works based on a library of pre-existing creative works. To give a very specific example of where AI breaks down, if I asked an AI to make a Beatlesesque song, the idea of putting a long "A Day In The Life"-type chord at the end would have not been considered had Lennon and McCartney not done it first. It could not have developed any possible "post-Beatles"-inspired works like what Jeff Lynne did with ELO without the same to begin with.

To sum up, AI is weak at spontaneous, randomish creativity, as well as developing truly novel concepts and ideas. I could see an AI getting better at approximating either, but never truly getting there.

In Japanese martial arts, there are considered 3 levels of mastery called Shuhari:

https://en.wikipedia.org/wiki/Shuhari

1) Obey the rules

2) Break the rules

3) Do your own thing (sometimes described as "Make the rules").

Could AI ever get to that third step of mastery? I'm not sure if it could even master the second step completely.

Expand full comment
dynomight's avatar

I sometimes wonder how much human intelligence will remain useful because it comes coupled to a human body. (Arguably human bodies are more impressive than our brains? They're *very* adaptable, and you can power them with burritos.)

If you assume that AI is better at all "pure intelligence" tasks than humans, but that AI hasn't invented robots that are as good as human bodies, then what follows? Does human intelligence remain vital because it has a high-bandwidth connection to human muscles?

Expand full comment
22 more comments...

No posts