They used the “Torrance Tests of Creative Thinking”, a pseudo-scientific test that measures and evaluates absolutely nothing of any objective measure or value.
Hah, yeah, that was my kneejerk reaction too: I read that as “the metric we use to determine creativity was found to be wildly inaccurate, with ML regularly placing in the 99th percentile”.
Embarrassing, considering how un-creative and original GPT-4 is. It’s an actual struggle to get ChatGPT to think outside of the box. Claude 2 on the other hand is much better at it.
But this goes to show how unimaginative the general population is if this truly is the case.
I have been playing with chat gpt for tabletop character creation. It’s not bad at coming up with new ideas. It is terrible at sticking to the rules of the game.
Add comment