this post was submitted on 26 Jun 2023
114 points (97.5% liked)
Asklemmy
43916 readers
914 users here now
A loosely moderated place to ask open-ended questions
Search asklemmy ๐
If your post meets the following criteria, it's welcome here!
- Open-ended question
- Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
- Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
- Not ad nauseam inducing: please make sure it is a question that would be new to most members
- An actual topic of discussion
Looking for support?
Looking for a community?
- Lemmyverse: community search
- sub.rehab: maps old subreddits to fediverse options, marks official as such
- !lemmy411@lemmy.ca: a community for finding communities
~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~
founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
That's why I used the phrase "faking empathy", I'm fully aware the chatGPT doesn't "understand" the question in any meaningful sense, but that doesn't stop it from giving meaningful answers to the question - that's literally the whole point of it. And to be frank, if you think that a unique question would stump it, I don't think you really understand how LLMs work. I highly doubt that the answer it spit back was just copied verbatim from some response in it's training data (which btw, includes more than just internet scraping). It doesn't just parrot back text as is, it uses existing tangentially related text to form it's responses, so unless you can think of an ethical quandary which is totally unlike any ethical discussion ever posed by humanity before (and continue to do so for millions of users), then it won't have any trouble adapting to your unique questions. It's pretty easy to test this yourself, do what writers currently do with chatGPT - go in and give it an entirely fictional context, with things that don't actually exist in human society, then ask it questions about it. I think you'd be surprised with how well it handles that, even though it's virtually guaranteed there are no verbatim examples to pull from for the conversation