Submitted by BronzeArcher t3_1150kh0 in MachineLearning
NotARedditUser3 t1_j90j0er wrote
Reply to comment by a1_jakesauce_ in [D] What are the worst ethical considerations of large language models? by BronzeArcher
If you spend some time looking up how microsoft's gpt integrated chat / ai works, it does this. Lookup the thread of tweets for the hacker that exposed its internal codename 'Syndey'; it scrapes his twitter profile, realizes he exposed its secrets in prior convo's after social engineering hacking it with a few conversations, and then turns hostile to him.
a1_jakesauce_ t1_j90k4h6 wrote
All I found was this https://twitter.com/kliu128/status/1623472922374574080?s=21
blablanonymous t1_j917xm2 wrote
Is that real? I don’t know why I feel like it could be totally fake
currentscurrents t1_j96vbfj wrote
Microsoft has confirmed the rules are real:
>We asked Microsoft about Sydney and these rules, and the company was happy to explain their origins and confirmed that the secret rules are genuine.
The rest, who knows. I never got access before they fixed it. But there are many screenshots from different people of it acting quite unhinged.
blablanonymous t1_j96xu8w wrote
Thanks for the link!
I mean I guess there was nothing too surprising about the rules, given how these systems work (essentially trying to predict the end of a user input text). But the rest, seems so ridiculously dramatic that I wouldn’t be shocked if he specifically prompted it to be that dramatic and hid that part. I’m probably being paranoid, since at least the rules part is true, but it seems like the perfect conversation to elicit every single fear people have about AI.
[deleted] t1_j96v1wv wrote
[deleted]
NotARedditUser3 t1_j9225b5 wrote
I'll reply back with what I was referring to later, it was a different thing
Viewing a single comment thread. View all comments