this post was submitted on 16 Aug 2023
26 points (81.0% liked)

Technology

34395 readers
453 users here now

This is the official technology community of Lemmy.ml for all news related to creation and use of technology, and to facilitate civil, meaningful discussion around it.


Ask in DM before posting product reviews or ads. All such posts otherwise are subject to removal.


Rules:

1: All Lemmy rules apply

2: Do not post low effort posts

3: NEVER post naziped*gore stuff

4: Always post article URLs or their archived version URLs as sources, NOT screenshots. Help the blind users.

5: personal rants of Big Tech CEOs like Elon Musk are unwelcome (does not include posts about their companies affecting wide range of people)

6: no advertisement posts unless verified as legitimate and non-exploitative/non-consumerist

7: crypto related posts, unless essential, are disallowed

founded 5 years ago
MODERATORS
all 8 comments
sorted by: hot top controversial new old
[–] [email protected] 28 points 1 year ago (1 children)

They used the "Torrance Tests of Creative Thinking", a pseudo-scientific test that measures and evaluates absolutely nothing of any objective measure or value.

[–] [email protected] 11 points 1 year ago

Hah, yeah, that was my kneejerk reaction too: I read that as “the metric we use to determine creativity was found to be wildly inaccurate, with ML regularly placing in the 99th percentile”.

[–] [email protected] 28 points 1 year ago* (last edited 1 year ago) (1 children)

Embarrassing, considering how un-creative and original GPT-4 is. It's an actual struggle to get ChatGPT to think outside of the box. Claude 2 on the other hand is much better at it.

But this goes to show how unimaginative the general population is if this truly is the case.

[–] [email protected] 10 points 1 year ago (1 children)

I have been playing with chat gpt for tabletop character creation. It's not bad at coming up with new ideas. It is terrible at sticking to the rules of the game.

[–] [email protected] 3 points 1 year ago (1 children)

The context window is still too short for any story. They just forget about old messages and only remember the newest context.

[–] [email protected] 3 points 1 year ago

That makes sense. The further back information would go, the harder it was to recall it. The answer wasn't to think harder, but to fill in the gaps.

[–] [email protected] 6 points 1 year ago

evaluating LLM

ask the researcher if they are testing form or meaning

they don't understand

pull out illustrated diagram explaining what is form and what is meaning

they laugh and say "the model is demonstrating creativity sir"

looks at the test

it's form