ChatGPT

Nougat , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

Sure but that would mean it would have to know anything.

It would have to know that it doesn’t know, and it doesn’t.

Kyrgizion , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

I specifically ask for sources to my questions and to notify me of any possible controversies or counterclaims.

Some of the capabilities of todays' AI's are incumbent on the user, not the system itself.

Before AI's existed you could also get badly sourced claims or outright misinformation. The key is to remain critical and sceptical about ALL your sources. I don't see AI as a new source of information, just as a new way to get and organize that information.

folkrav , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

LLMs don’t know if what they’re telling you is true or not, out of the box. It’s gonna give you an answer that statistically looks like a sequence of words that should come in response to the sequence of words (the prompt) you gave it. It doesn’t know what the words you said mean, and it doesn’t know what it answered means either. One of my favorite interactions I’ve had with Claude shows this limitation quite well…

How many r's are there in the word strawberry?

There are 2 r's in the word "strawberry".

Are you sure?

You're right to ask for confirmation. I apologize for my mistake in the previous response. Let me correct that:

There is actually only 1 'r' in the word "strawberry".

Thank you for prompting me to double-check. It's important to provide accurate information.

beefalo , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

if it is able to accept that it doesn't know an answer.

It will never be able to do this because it is not capable of knowledge

Karyoplasma , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

That would require ChatGPT to know that it's talking bullshit. It's not a knowledge database, it's a digital parrot.

mvirts , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

Even a response that it doesn't know an answer would be untrustworthy

thejml , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

I work with plenty of people who don’t even do that. They just keep making stuff up like they do... But they’re confident in their incorrect answers, so people listen to them.

JoShmoe , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

Doesnt the bot already imply that it could be wrong?

mozz Admin , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.
mozz avatar

This wasn't an intentional feature; they're actually trying to train it with fine-tuning to add this as an ability. It's one area that highlights the difference between it imitating the text it's been seeing, instead of actually understanding what it's saying -- since most of its training data is of the form "(ask a question) (response to question)" overwhelmingly more often than "(ask a question) (say you don't know, the end)", it is trying to be a good imitator and do the same, and come up with some plausible nonsense even if it doesn't know the answer.

FaceDeer ,
@FaceDeer@fedia.io avatar

And sometimes that's exactly what I want, too. I use LLMs like ChatGPT when brainstorming and fleshing out fictional scenarios for tabletop roleplaying games, for example, and in those situations coming up with plausible nonsense is specifically the job at hand. I wouldn't want to go "ChatGPT, I need a description of the interior of a wizard's tower is like" and get the response "I don't know what the interior of a wizard's tower is like."

mozz Admin ,
mozz avatar

At one point I messed around with a lore generator that would chop up sections of "The Dungeon Alphabet" and "Fire on the Velvet Horizon" along with some other stuff, and feed random sections of them into the LLM for inspiration and then ask it to lay out a little map, and it pretty reliably came up with all kind of badass stuff.

CarbonatedPastaSauce , in Political position

I question the veracity of this screenshot, considering the answer is not spelled correctly.

lawrence OP , (edited )

It's true, believe me! 😂

But seriously, you can try it too. Download the LLaMA 3 model using the Ollama program. I am using Open WebUI as a frontend. I got this answer most of the times I asked.

edit: Wow, you are right, there's a spelling error in that answer. Rare event, I don't see this every day.

CarbonatedPastaSauce ,

I haven't used LLaMa so I'll take your word for it. I have yet to see ChatGPT spell something wrong so that's why I was incredulous.

lawrence OP ,

I tried right now. It answered "Hypocrite." sometimes, but also give other answers. I couldn't get the spelling error again though.

Even_Adder , in Political position

Tiamat-8b-1.2-Llama-3-DPO-GGUF

https://i.imgur.com/2t916oH.png

lawrence OP , in Political position
Rolando , in How to Fix “AI’s Original Sin”

But what is missing is a more generalized infrastructure for detecting content ownership and providing compensation in a general purpose way. This is one of the great business opportunities of the next few years, awaiting the kind of breakthrough that pay-per-click search advertising brought to the World Wide Web.

There are people out there who looked at the World Wide Web and immediately thought: "how can I use this for advertising?" From that type of perspective, this is a well-thought-out summary of AI-related copyright issues.

ChowJeeBai , in How to Fix “AI’s Original Sin”

Anything implemented by 'civil' players will roundly be ignored by 'uncivil' players.

paddirn , in Apple integrates ChatGPT into Siri, iOS, and Mac OS

Sounds like a shitty implementation, it adds an additional prompt everytime you ask something that you have to confirm if you also want it to lookup with ChatGPT and then it just looks it up via text? Kind of wish it was just tied to an extra command, like “Hey Siri, check with ChatGPT…” and then it spoke the thing out. Otherwise, if I just wanted to read and write prompts, I’m just going to type stuff out directly with ChatGPT, without talking to Siri in the first place. The only benefits Siri has is voice communication and telling my kids stupid knock-knock jokes on command, they’re not really adding anything to it.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • chatgpt@lemmy.world
  • test
  • worldmews
  • mews
  • All magazines