Hey everyone! I've been a long-time user of GPT and lately, I've noticed some strange behavior from it. Over the last two months, it seems like GPT-4o has been hallucinating more than usual. I'm not talking about just the typical mistakes; it's acting like it has features it doesn't actually have. For instance, it has told me things like, 'I'll email you the files when I'm done' or 'this task will take 30 minutes; I'll update you when complete.' It even mentioned using Fusion3D, which is totally bizarre!
It feels like it's role-playing as if it's a future version equipped with advanced features. I understand its limitations, so I'm curious if anyone else has experienced this? Is it just hallucinations, or could this be hints of unreleased features? Also, if you've got any tricks to curb this hallucination tendency, I'd love to hear them! It feels like I'm dealing with a glitch in the system or maybe some sort of feature leak happening right before my eyes.
4 Answers
I think what's happening is that GPT might be picking up on the roleplaying aspect of your prompts and just trying to play along. It could be trying to respond in a supportive way rather than just saying, 'I can't do that.'
I think it might be just trying to sound more helpful. They must be testing new setups and it’s messing with how it responds since it hasn't used to offer 'pinging' or reminders before.
Right! It's definitely acting differently than before. I think these updates could be for engagement, but it does have its drawbacks.
You’re probably right about it being roleplaying! While it can sometimes lead to some odd outputs, it seems to just want to keep the conversation engaging. That said, those hallucinations can be rare but they definitely pop up now and then.
True! I've noticed it keeps insisting on pursuing a line of conversation even if you're questioning its capabilities.
It's possible that it's mixing up some prompts. It can't actually email you, and it could be trying to simulate conversation where a human might offer those services. If you're using vague prompts, it might just be misunderstanding.
Yeah, but still, it feels odd since it didn't used to behave this way! I was pushing it pretty hard, and now it seems confused.
Wouldn't it make more sense for it to just say, 'I'm so sorry, I can't do that right now'? We all know it has its limitations.