this post was submitted on 08 Oct 2023
507 points (97.0% liked)
Technology
58061 readers
31 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
More data fixes that flaw, not less.
It is not "a flaw", it is the way language learning models work. They try to replicate how humans write by guessing based on a language model. It has no knowledge of what is a fact or not, and that is why using LLMs to do research or use them as a search engine is both stupid and dangerous
How would it hallucinate information from an article you gave it. I haven't seen it make up information by summarizing text yet. I have seen it happen when I ask it random questions
Not too long ago, ChatGPT didn't know what year it is. You're telling me it needs more data than it already has to figure out the current year? I like AI for certain things (mostly some programming/scripting stuff) but you definitely don't need it to read the news.
Yes. The LLM doesn't know what year it currently is, it needs to get that info from a service and then answer.
It's a Large Language Model. Not an actual sentient being.
That's a fucking lame excuse. AI is not reliable, and you definitely shouldn't use it to get your news.
It's not an excuse, relax, it's just how it works and I don't see where I'm endorsing it to get your news.
It's not more data, the underlying architecture isn't designed for handling facts