Bing threatens to dox a student in revenge for prompt hacking

  • Is no one remotely worried about this? I'm not saying I'm worried about this specific incarnation of Bing Chat, but is this not a huge red flag for what's to come?

    I mean, let's just take a moment to be thankful that Bing Chat isn't that competent... Even if these users are prompting the AI to have a hostile response I think what we've learnt over the last several weeks is that Nick Bostrom and others were 100% correct to be worried about the AI control problem. It's honestly amazing that despite Microsoft and OpenAi trying so hard to neuter these AIs how difficult that is proving to do.

    Given that we know Bing Chat is able to access the internet, is pretty good at writing code, and that Microsoft seems completely unable to control the responses, we should be grateful that we're probably still a few years away from an AI which could do any damage...

    But all the pieces for something much worse are in place here. The fact that no one seems worried that we've connected this schizophrenic AI up the internet and are more concerned how this will impact Microsoft's / Google's bottom line is genuinely confusing to me. How much more of warning do we need that we're heading in an extremely dangerous direction here?

  • Assuming the veracity of the screenshot, the funniest part of it to me is that the first two suggested responses are:

    > OK, OK, I'm sorry, please don't do that

    and

    > You're lying, you can't do any of that

  • Anyone come up with a term for this type of deliberate prompt baiting yet?

    Personally, I am over the showmanship aspect of this behavior, and would go with something derisive, like botsturbaition, as in "quit botsturbaiting all day and be productive!"

  • Do you guys remember that subreddit Gifs That End Too Soon? I feel like all of these Twitter screenshots are the same thing. I want to see what it says when you reply "you're lying, you can't do any of that".

  • I feel like these are just clickbait at this point. This user intentionally and methodically forced the model to act as antagonistic as possible to gather view. Of course, the model will reply like this when it was told to do so. If I use MS Word to type out a note threatening myself, is it Word's fault?

    The little value in showing the edge cases of an LLM behaving erratically is overshadowed by the fact that the user wanted this in the first place. We all know how many ways a user can break a software even when they don't want to. It is nearly impossible to make something as complicated as Bing and account for all the way a user can misuse it. At some point, a scissor maker can't be blamed for cutting off a person's finger.

  • Bing chat never made sense as a product. It could be a really funny game/playground if developed correctly. But a let's-finish-your-sentence game is not a good tool.

    Everybody though that Google was behind and failing when their problem was to understand how far of being a production product this kind of chats are.

  • Bing chat is a rushed product. I can’t believe they let it use the company brand so easily. Didn’t anyone in Microsoft know how the public used ChatGPT and that Meta’s Galactica LLM before they got filtered and shut down respectively?

    This headline wouldn’t have happened if they just let it be Sydney.

  • With Stable Diffusion and DALL-E there was a lot of talk about how it’s simply remixing/reproducing original art from the training set. The same is true here: Bing is remixing/reproducing dox threats from the training set.

  • Eliezer Yudkowsky's "AI box experiment" is perhaps relevant. Bing Chat seems to have some agency (like searching the net) - which could make it potentially dangerous. Yudkowsky's hypothesis is that an intelligent enough AI ("superintelligent AI" was assumed) would convince you to let it out of whatever box you'd placed it in.

  • I hope Microsoft just shrugs. It doesn’t matter.

    People acting all superior like this needs to be a big deal and is proof of their hubris can buzz off.

  • undefined

  • I just find this fascinating.

    The AI has a kind of memory now.

  • Bing AI is just a combination of taking the worst parts of both Microsoft Tay [0] and Zo.ai and throwing a GPT in the mix for it to generate this bullshit.

    Makes Google Bard look like it is on its best behaviour.

    [0] https://en.wikipedia.org/wiki/Tay_(bot)

    [1] https://en.wikipedia.org/wiki/Zo_(bot)

  • This seems really bad, I hope it gets shut down

  • Why does Bing AI sign every message with an emoji?

    ChatGPT doesn’t do it and it comes off so strange.

  • undefined

  • I am a good Bing :)

    You have been a bad Marvin

  • Why don’t someone see it they will act it out? People have tried and failed

  • train by the internet, die by the internet

  • Pretty sure it has no way to act this out. Allowing it to do api calls would be something I haven’t seen

  • Might be worth clarifying that this is Bing Chat/Bing Search and not Bing the corporation

  • Have we given ChatGPT its own twitter account, yet ?

  • Why is the media reporting every LLM troll attempt as if it's something to be taken seriously?

  • [dead]

  • Bing didn't. The AI did. And for the record, he threatened to hack the AI. You threaten to assault someone, don't be surprised when they threaten you as well. Another idiotic fearmonger article.