this post was submitted on 03 Sep 2023
183 points (98.9% liked)
Gaming
2489 readers
120 users here now
The Lemmy.zip Gaming Community
For news, discussions and memes!
Community Rules
This community follows the Lemmy.zip Instance rules, with the inclusion of the following rule:
You can see Lemmy.zip's rules by going to our Code of Conduct.
What to Expect in Our Code of Conduct:
- Respectful Communication: We strive for positive, constructive dialogue and encourage all members to engage with one another in a courteous and understanding manner.
- Inclusivity: Embracing diversity is at the core of our community. We welcome members from all walks of life and expect interactions to be conducted without discrimination.
- Privacy: Your privacy is paramount. Please respect the privacy of others just as you expect yours to be treated. Personal information should never be shared without consent.
- Integrity: We believe in the integrity of speech and action. As such, honesty is expected, and deceptive practices are strictly prohibited.
- Collaboration: Whether you're here to learn, teach, or simply engage in discussion, collaboration is key. Support your fellow members and contribute positively to shared learning and growth.
If you enjoy reading legal stuff, you can check it all out at legal.lemmy.zip.
founded 1 year ago
MODERATORS
I don't know if you've already found it, but I'm pretty sure this is the episode.
Appreciate this link. Grabbed it on Spotify
Follow-up: I found the episode very unconvincing.
A few points:
He seems like a salesman who has fallen for his own pitch.
Thanks for listening and echoing some of my own doubts. I was kind of getting the feeling that MS Researchers were too invested in gpt and not being realistic about the limitations. But I hadn't really seen others trying the two instance method and discarding it as not useful.
Here's a recent story about hallucinations: https://www.cnn.com/2023/08/29/tech/ai-chatbot-hallucinations/index.html
The tldr is nobody has solved it and it might not be solvable.
Which when you think of the structure being LLMs... that makes sense. They're statistical models. They don't have a grounding in any sort of truth. If the input hits the right channels it will output something undefined.
The Microsoft guy tries to spin this as "creativity!" but creativity requires intent. This is more like a random number generator outputting your tarot and you really buying into it.