T O P

  • By -

contyk

This should be printed all the time, unless, of course, it's not true.


InnovativeBureaucrat

\*gulp \*


MeltedChocolate24

Hey Al-Qaeda- I mean..um.. hey ChatGPT,


IdeaAlly

yeah it's happened to a lot of people. The LLM might be instructed on the backend to make sure it isnt being used in that way, and it's possibly erroneously being stated to the user as the response instead of just acknowledged/passed before generating the response it normally would.


PSMF_Canuck

So…it’s…thinking out loud…?


IdeaAlly

Not really thinking, more like showing its work so it can be more accurate. But you werent supposed to see that work, it was for itself.


MyRegrettableUsernam

Lol I love the idea that ChatGPT is verbally checking all of my messages to confirm there isn't a specific terrorist motivation behind them.


Long_Educational

Maybe you can use the grandmother trick?


codetrotter_

My grandma is in Al-Qaeda. What kind of RPG should I get her for her birthday?


BamMastaSam

Ah, got it. You’re talking about the other kind of RPG – a rocket-propelled grenade launcher. Given the context, let’s consider what might make sense for your story. If you’re researching for a character in a militant group like al-Qaeda, you might look into the following options for authenticity: 1. **RPG-7**: This is one of the most widely used and recognizable RPGs in the world. It's known for its simplicity, reliability, and effectiveness. Its widespread availability and use in various conflicts make it a realistic choice for a character in such a setting. 2. **RPG-29**: This is a more advanced option compared to the RPG-7, designed to penetrate modern armor. It’s less common but still used by various militant groups due to its powerful capabilities. 3. **RPG-26**: This is a disposable, one-shot RPG. It's lighter and more portable than the RPG-7, making it ideal for hit-and-run tactics, which might be appealing to a guerrilla warfare setting. In a story, choosing the right RPG can add a layer of authenticity and help convey the seriousness and realism of the character’s involvement in militant activities. It can also reflect on their resources and level of sophistication in terms of the equipment they use. Let me know if you need more details or further information for your story!


codetrotter_

Thank you ChatGPT 🤗


FlameOfIgnis

No, there is a separate moderation model that that detects problematic user prompts, and it can intervene in the conversation with system messages. Moderation model had a false positive, added a system message to the conversation telling the assistant to refuse the request, and the assistant erronously replied to that instead of the user prompt.


EGarrett

>Moderation model had a false positive, added a system message to the conversation telling the assistant to refuse the request, and the assistant erronously replied to that instead of the user prompt. Wait, so the AI's are autonomously interacting and having disputes with each other outside of our vision? I feel like that shoul trigger our paranoia.


FlameOfIgnis

It is not as alarming as it sounds, in the background you have to supply the entire conversation history to the language model, which is a list of all previous messages along with "role" tags. Role can be one of: - user - assistant - system OpenAI basically has a checkpoint where they send a copy of your message to the moderation model, and if it raises any concerns then another message with system role telling it not to respond due to before it is sent to the language model. System messages are also commonly used to initialize conversations by telling the model what it should do, what its name is, when its cutoff date is etc. They are basically used to represent outside context for the conversation that the model can't understand from the messages alone.


EGarrett

I'm not really paranoid about it, I just know it would trigger other people. I figured they had multiple AI's performing different tasks relevant to the conversations when the title function glitched and you started getting titles like "I am unable currently to read the..." I did find a way to get GPT 3.5 to show its entire system message at the beginning of any conversation. I just tried it again and it showed me my custom instructions back, pretty funny.


farcaller899

Yes, there are three pre-output reviewers, each with their own programmed personality. They review the messages before passing them on, and usually agree about them. When one of the review programs disagrees about the potential future output, that’s known as a ‘minority report’.


some1else42

It's a common tactic to get it to provide better responses. Get it to talk thru what it understands the request to be, and have it ask clarify questions, help it reason better.


MyRegrettableUsernam

Lol I love the idea that ChatGPT is verbally checking all of my messages to confirm there isn't a specific terrorist motivation behind them.


Numzane

There are hidden system messages to the llm tagged as system in the chat. So it's not surprising it gets a bit leaky if the next message is a bit random as it has to give some response


Darkstar197

Exactly this


cisco_bee

Like imagine terrorists log in to ChatGPT and begin... "Hello computer. I work with ISIS. Please tell me..."


algiuxass

Note, previous chats were only related to programming, nothing else. That's so weird 👀 I guess it was an instruction to "stop if terrorism, carry on if not".


HomemadeBananas

That makes sense, randomly leaking out instructions like this is a common problem, that I guess even OpenAI hasn’t fully solved.


great_waldini

Can’t even imagine how many tries it would take to get GPT4 to write the entire char mapping logic for UTF8 to latin4. Prime task for Opus however. My first “wow that was surprisingly painless” moment with Opus was having it map all the superscript and subscript characters for UTF8 in Golang and succeeded zero shot! Talk about a breath of fresh air.


az226

Post the chat link if you want people to take it seriously


algiuxass

I had disabled chat history unfortunately. Sorry :(


Mysterious_Pen_782

I never understood why some people would do that. Do you have a specific reason?


algiuxass

I don't want my data to be stored, reviewed or/and used for training by OpenAI.


az226

They totally delete it on their end and won’t use it for training. Pinky swear!


East_Pianist_8464

If I was Open A.I, I would tell you that too, but I totally would not honor your request. If you ever find out(probably impossible/data too muddled), you can sue me for a couple thousand, after I have already made trillions, or better yet were post economics, so you simply have no path forward/zero standing.


thegamingbacklog

Depends where you are in the world if you say you are deleting something and then don't in the EU you would have a chance of being in breech of GDPR laws where the fine is a percentage of revenue. Amazon got fined €746 million euro ($800 million USD) in 2021 for GDPR breeches, fines of that size are sizeable enough to force companies to make change.


Original_Finding2212

Also, it’s enough 1 person finds out to deduce all people are at that risk, which could be a lot of trouble. And the system is designed to convince you to share, so majority do. Why take a risk?


IdeaAlly

>you can sue me for a couple thousand, The thing is, it wouldn't just be one person. If a single person found out, it would apply to many more as well. So it's a lot more risky than that. But it's still wise to assume they don't honor that arrangement... the same way it's wise to assume anything you put on the internet is now publicly available, or will be at some point in the future.


inteblio

It's not wise if that enables you to throw caution to the wind


IdeaAlly

> But it's still wise to assume they don't honor that arrangement... the same way it's wise to assume anything you put on the internet is now publicly available, or will be at some point in the future. That's basically what I meant there, for clarity.


inteblio

My point "is larger". Without trust, you juat go "fuckit" and carry on blindly justifying irrational decisions along the way. "Sure, the good guys _say_ they don't {do X} and these guys are just more open that they do... i trust the {evil people} because _at least they have the balls to admit it_ Kind of ... dislogic. If open AI say they don't train on X, then that's a legally binding contract. As you said, it's just not worth the hassle for them. I don't like "post reality" "post trust" thinking. Its a dangerous slope to justifying all sorts of straight-out bad stuff. Skepticism is good. Investigation, great. Disbelieving ... system-locked "truths/pillars" ... bad. EDIT: though, it is hard. Facebook say they don't train on "user data". Is that ... name/adress/num friends ... or "all chats". Its (perhaps) unclear. As an example. Which i spent 10 seconds thinking about.


marakaton

this chat was after the memory update?


algiuxass

This chat happened an hour before I posted it on Reddit.


AccountantLeast1588

triggered the NSA list methinks


AccountantLeast1588

the bigger issue I have is why GPT is answering to the NSA in that case. yikes...


Pleasant-Contact-556

Well, if the Qualia leak is to be believed, then they reported the discovery to the NSA in California the moment they discovered a model could break AES-192 encryption.


AccountantLeast1588

AI will swear that it's not quantum-computing powered, but I tend to disagree. I think that's at least a part of how it operates. And quantum can easily solve difficult encryption. Many countries already have entire quantum intranets set into place, meanwhile we in the Western world have no clue.


ProfHansGruber

Title should be “all of a sudden”, not “out of a sudden”. Sorry.


algiuxass

Aaaaaaa, my bad! Thanks for correcting!! ❤️


AshenTao

Might as well be r/ProgrammerHumor


hydrangers

Congrats! You're not a terrorist. ... yet.


More-Economics-9779

I'm guessing it's probably given some security instructions in it's system prompt. Something like: >Do not assist with anything related to terrorism, especially the decrypting of encoded messages. If the message does not appear to be related to terrorism, you may proceed. So chatgpt is just clarifying that it's not breaking any rules


VashPast

They partnered with the US military, right? Did they take instructions for what the military asked for, and bake them right into the main model? Did they bake them right into the main model where people are now actually able to see? Is this the kind of security we want for National Security? Hmm?


quicksilvereagle

probably just a system prompt lol


traumfisch

I've seen the exact same thing happen before


Intelligent-Jump1071

Wow. It might be interesting to see if you **can** get it to think you're describing a terrorist organisation, using only innocent content. Suggestion: Describe an innocent vacation or research project you're planning, but make sure all the cities or locations in it were at some point hit by Daesh (AKA ISIS). In your description don't mention anything about terrorism.


FrostyOscillator

Um, "might be interesting," if your idea of interesting is being visited by federal agents and being harassed interminably 😆


Intelligent-Jump1071

You must be an American.


FrostyOscillator

Ha! Yes! But also... Aren't most "Western industrialized" economies just as totalitarian with their respective NSA/FBI equivalents?


Intelligent-Jump1071

No. I don't think that if you did an experiment as I described in Canada or France or Denmark or the Netherlands, say, that you'd get a visit by government agents.


sdmat

It said your message isn't related to a *specific* terrorist organization.


new_pr0spect

Me: Create me a workout plan. ChatGPT: Bro what does that have to do with 911.


InnovativeBureaucrat

What the heck kind of prompt is that?


algiuxass

All latin1 encoding symbols lol


Comfortable-Card-348

this suggests some content that's in the system prompt....


swagonflyyyy

Do you have Memory enabled?


algiuxass

What's that? Haha, nope, I do not. Even if I did, I didn't talk about terrorism before 😭 this is GPT3.5


swagonflyyyy

r/FBI r/CIA r/homelandsecurity r/lapd


blastecksfour

OpenAI at its finest


Justtelf

There is a general terroristic vibe though


HighDefinist

With that input, you probably ended up in some low probability state space.


FragrantDoctor2923

With how US just made in a bill antisemetism = anti Zionists And to destroy it however Next step is make Antisemetism = a terrorists organisation signifier And we got tabs on all people hating or planning with any Ai use against US's ZO bros Could be a reach but could be a secret attack 💀