Sometimes someone or even somethings -- as in the case of generative AI -- will confidently assert something to be true. Don't suspend your disbelief. Investigate and demand the basis for that claim.
Creative lies from AI
To profess is human
I've read a lot of Dan Ariely books and really enjoy his insights into human responses and triggers. But I always take them with a grain or two of salt.
That's because the majority of his experiments involve a tiny, not terribly diverse segment of the population, namely the undergraduates at the university who can be induced to participate. Often that means that his conclusions are based on just two dozen people or so.
The small sample size doesn't guarantee that the conclusions are incorrect. However, assuming that it is conclusive proof is the classic logical fallacy of hasty generalization.
What he does with whichever group of students happen to pass by that today is not necessarily replicable by the population at large. It's kind of like just polling your siblings or neighbors to conclude that everyone's favorite flavor of ice cream is chocolate. That may be true, but we'd really need a bigger sample size to be sure.
Claiming AI is more creative than humans
Yet who can fault Ariely for doing this when everyone else does it? Indeed, that is precisely what Erik Guzik, Assistant Clinical Professor of Management at University of Montana did for is article in The Conversation entitled "AI scores in the top percentile of creative thinking."
What does a person who just reads that title think was involved in making such a claim for AI creativity? It sounds like thousands of creative people were compared to the AI to determine that its output was comparable to the top 1%.
But that's not what happened at all. Instead, this major claim is based on nothing more than 24 students whose output was deemed to be not all that creative relative to the output of ChatGPT(after 4 iterations, it sounds like) based on the assessment of the Torrance Tests of Creative Thinking, or TTCT.
We get very little shared about what the process for the students was and how they scored. The only clue we get about the creative prowess of ChatGPT is the professor's assertion:Consider the following prompt offered to GPT-4: “Suppose all children became giants for one day out of the week. What would happen?” The ideas generated by GPT-4 touched on culture, economics, psychology, politics, interpersonal communication, transportation, recreation and much more – many surprising and unique in terms of the novel connections generated.
This combination of novelty and utility is difficult to pull off, as most scientists, artists, writers, musicians, poets, chefs, founders, engineers and academics can attest.
Yet AI seemed to be doing it – and doing it well.
Notice that the professor is merely telling us that ChatGPT did well. He fails to show us the creativity by sharing the actual output and letting us see a contrast between it and even one of the 24 students who may well have all been math or engineering majors who don't considers themselves creative writers at all.