ChatGPT

JoShmoe , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.

Doesnt the bot already imply that it could be wrong?

mozz Admin , in ChatGPT would have been so much useful and trustworthy if it is able to accept that it doesn't know an answer.
mozz avatar

This wasn't an intentional feature; they're actually trying to train it with fine-tuning to add this as an ability. It's one area that highlights the difference between it imitating the text it's been seeing, instead of actually understanding what it's saying -- since most of its training data is of the form "(ask a question) (response to question)" overwhelmingly more often than "(ask a question) (say you don't know, the end)", it is trying to be a good imitator and do the same, and come up with some plausible nonsense even if it doesn't know the answer.

FaceDeer ,
@FaceDeer@fedia.io avatar

And sometimes that's exactly what I want, too. I use LLMs like ChatGPT when brainstorming and fleshing out fictional scenarios for tabletop roleplaying games, for example, and in those situations coming up with plausible nonsense is specifically the job at hand. I wouldn't want to go "ChatGPT, I need a description of the interior of a wizard's tower is like" and get the response "I don't know what the interior of a wizard's tower is like."

mozz Admin ,
mozz avatar

At one point I messed around with a lore generator that would chop up sections of "The Dungeon Alphabet" and "Fire on the Velvet Horizon" along with some other stuff, and feed random sections of them into the LLM for inspiration and then ask it to lay out a little map, and it pretty reliably came up with all kind of badass stuff.

kromem ,

Part of the problem is fine tuning is very shallow, and that a contributing issue for claiming to be right when it isn't is the pretraining on a bunch of training data of people online claiming to be right when they aren't.

mozz Admin ,
mozz avatar

Yeah. It is fairly weird to me that it’s such a common thing to do to take the raw output of the LLM and send that to the user, and to try use fine-tuning to get that raw output to look some way that you want.

To me it is obvious that something like having the LLM emit a little JSON block which includes some field which covers “how sure are you that this is actually true” or something, is more flexible and simpler and cheaper and works better.

But what do I know

kromem ,

The problem is that they are prone to making up why they are correct too.

There's various techniques to try and identify and correct hallucinations, but they all increase the cost and none are a silver bullet.

But the rate at which it occurs decreased with the jump in pretrained models, and will likely decrease further with the next jump too.

Cosmicomical ,

Good look getting it to reply consistently with a json object

Edit: maybe i'm shit at prompting but for me it's almost impossible to even get it to just shut up and consistently reply yes or no to my questions

mozz Admin ,
mozz avatar

I haven’t really had a problem with it… maybe like 5% of the time it will want to do something a little bit weird like wrapping it in ``` but in general it seems like it works well enough to be able to parse with a program and just retry if it does something weird.

You do have to set it up a little carefully, I guess - like usually I’ll give it an example of what I want it to emit, and that’ll be good enough that that’s the form it will follow when it’s emitting stuff back to me. But yeah if you give it prompting and a specific machine readable thing to give back that seems like it usually works better than sticking with English and hoping it goes “yes” or “no” or etc like that.

CarbonatedPastaSauce , in Political position

I question the veracity of this screenshot, considering the answer is not spelled correctly.

lawrence OP , (edited )

It's true, believe me! 😂

But seriously, you can try it too. Download the LLaMA 3 model using the Ollama program. I am using Open WebUI as a frontend. I got this answer most of the times I asked.

edit: Wow, you are right, there's a spelling error in that answer. Rare event, I don't see this every day.

CarbonatedPastaSauce ,

I haven't used LLaMa so I'll take your word for it. I have yet to see ChatGPT spell something wrong so that's why I was incredulous.

lawrence OP ,

I tried right now. It answered "Hypocrite." sometimes, but also give other answers. I couldn't get the spelling error again though.

Even_Adder , in Political position

Tiamat-8b-1.2-Llama-3-DPO-GGUF

https://i.imgur.com/2t916oH.png

lawrence OP , in Political position
Rolando , in How to Fix “AI’s Original Sin”

But what is missing is a more generalized infrastructure for detecting content ownership and providing compensation in a general purpose way. This is one of the great business opportunities of the next few years, awaiting the kind of breakthrough that pay-per-click search advertising brought to the World Wide Web.

There are people out there who looked at the World Wide Web and immediately thought: "how can I use this for advertising?" From that type of perspective, this is a well-thought-out summary of AI-related copyright issues.

ChowJeeBai , in How to Fix “AI’s Original Sin”

Anything implemented by 'civil' players will roundly be ignored by 'uncivil' players.

paddirn , in Apple integrates ChatGPT into Siri, iOS, and Mac OS

Sounds like a shitty implementation, it adds an additional prompt everytime you ask something that you have to confirm if you also want it to lookup with ChatGPT and then it just looks it up via text? Kind of wish it was just tied to an extra command, like “Hey Siri, check with ChatGPT…” and then it spoke the thing out. Otherwise, if I just wanted to read and write prompts, I’m just going to type stuff out directly with ChatGPT, without talking to Siri in the first place. The only benefits Siri has is voice communication and telling my kids stupid knock-knock jokes on command, they’re not really adding anything to it.

mozz Admin , in Apple integrates ChatGPT into Siri, iOS, and Mac OS
mozz avatar

Just in time as ChatGPT is becoming encrusted with inferiority

coffee_with_cream , in Typos in code generation now?

Run ollama instead!

BitSound , in Typos in code generation now?

I've noticed that as well. It silently fixed the typo when I asked it to update its previous response to fix something else.

Varven , in Typos in code generation now?
@Varven@lemmy.world avatar

Yer I started noticing that chatgpt was doing typos more and then I switched over to blackbox ai

magiccupcake , in Typos in code generation now?

I actually decided to cancel my ChatGPT subscription since it started to be so useless, for code generation, and writing help.

I'm so far pretty happy with Claude, but I've only used it Friday.

Like one of the things that it would do is give me wrong code, I'd fix it, give it back the corrected code to add something else, and it would remove the corrections and other things it added earlier!

mozz OP Admin ,
mozz avatar

Yeah. Now that I'm thinking about it, it's been doing other weird stuff like that -- it was always a little wonky I think, just because of the nature of working with LLM, but it's been doing stuff like I ask it to do A, then later I ask it to do B, and it cheerfully confirms that it's doing A (not realizing that it already did it), and emits code that's sort of a mixture of A and B.

IDK. I've also heard good things about Mistral. I just tried to create a Claude account but the phone verification isn't working and I have no idea why. I may check it out though; if this is accurate then it's pretty fuckin fancy and the Haiku model is significantly cheaper and smarter even than the 3.5 API which has a notable lack of cleverness sometimes.

thebeardedpotato ,

ChatGPT has been doing this thing where I’ll ask it to do A, B, C in sequential, iterative prompts, but when it does C, it removes the lines it added for B. Then when you tell it that it removed B and needs to add it back in, it undoes C while saying it’s doing A, B, C. So frustrating.

lawrence , in chatgpt is down for some users
XEAL , in chatgpt is down for some users

They are currently investigating the issue.

Thekingoflorda ,
@Thekingoflorda@lemmy.world avatar

In the image you can see the uptime for the past 30 days.

AceSLS ,

Being able to read this means you're currently browsing lemmy and are literate

h3mlocke ,
@h3mlocke@lemm.ee avatar

I verified it and you are correct. 👍

  • All
  • Subscribed
  • Moderated
  • Favorites
  • random
  • chatgpt@lemmy.world
  • test
  • worldmews
  • mews
  • All magazines