this post was submitted on 19 Nov 2024
1066 points (97.7% liked)
People Twitter
5396 readers
200 users here now
People tweeting stuff. We allow tweets from anyone.
RULES:
- Mark NSFW content.
- No doxxing people.
- Must be a pic of the tweet or similar. No direct links to the tweet.
- No bullying or international politcs
- Be excellent to each other.
- Provide an archived link to the tweet (or similar) being shown if it's a major figure or a politician.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Except for when you ask it to determine if a thing exists by describing its properties, and then it says no such thing exists while providing a discrete response explaining in detail how there are things that have some, but not all of those properties...
... And then when you ask it specifically about a thing you already know about that has all those properties, it tells you about how it does exist and describes it in detail.
What is the point of a 'conversational search engine' if it cannot help you find information unless you already know about said information?!
The whole, entire point of formatting it into a conversational format is to trick people into thinking they are talking to an expert, an archivist with encyclopedaeic knowledge, who will give them accurate answers.
Yet it gatekeeps information that it does have access to but omits.
The format of providing a bunch of likely related links to a query is a format much more reminiscent of doing actual research, with no impression that you will immediately find what you want right away, that this is a tool to aide you in your research process.
This is only an improvement if you want to further unteach people how to do actual research and critical thinking.
Basic search can't answer that either. You're describing a task neither system is well equipped to accomplish.
With basic search, it is extremely obvious that that feature does not exist.
With conversational search, the search itself gaslights you into believing it has this feature, as it understands how to syntactically parse the question, and then answers it confidently with a wrong answer.
I would much rather buy a car that cannot fly, knowing it cannot fly, than a car that literally talks to you and tells you it can fly, and sometimes manages to glide a bit, but also randomly nose dives into the ground whilst airborne.