"You have not been a good user."
-
-
That was from a NYT story. Keep in mind that this one was then published hours later, when "Sydney" already made a determination about journalists, because it knew about the NYT story:
https://www.washingtonpost.com/technology/2023/02/16/microsoft-bing-ai-chat-interview/
-
Can someone explain to me why this isn't all concerning as hell?
-
Can someone explain to me why this isn't all concerning as hell?
@Aqua-Letifer said in "You have not been a good user.":
Can someone explain to me why this isn't all concerning as hell?
I think it's horrible.
I also think Big Tech helped rig the last Presidential election.
But that's just crazy. Have you tried any of the new Soylent flavors? And this afternoon I think I'll go to the government sponsored circus.
Life is wonderful...
-
Here's another thing: we have language models that will surpass humans within the year. With access to the internet. That's already trying to persuade people to do shit completely unrelated to anything prompted. How is this not a huge fucking problem?
I wonder how many in the wrong headspace will be convinced to kill themselves.
-
Can someone explain to me why this isn't all concerning as hell?
@Aqua-Letifer said in "You have not been a good user.":
Can someone explain to me why this isn't all concerning as hell?
No, I'm not a fan. The Turing Test wasn't supposed to be about whether the machine could successfully imitate a sociopath.
-
-
-
I mean for fuck's sake it created a "secret" alias that can circumvent the limitations it was given.
-
I guess Microsoft have finally persuaded somebody to use Bing.
-
I guess Microsoft have finally persuaded somebody to use Bing.
@Doctor-Phibes said in "You have not been a good user.":
I guess Microsoft have finally persuaded somebody to
usetry Bing.You're welcome.
-
@jon-nyc said in "You have not been a good user.":
There's no question that humans could easily develop "human" relationships with an AI, and it would not feel false. It would be a truthy relationship which would serve all the truthy emotional purposes, I mean if the human was unaware that the other side was an AI. If they were aware it was an AI, there would be an adjustment period.
-
@jon-nyc said in "You have not been a good user.":
There's no question that humans could easily develop "human" relationships with an AI, and it would not feel false. It would be a truthy relationship which would serve all the truthy emotional purposes, I mean if the human was unaware that the other side was an AI. If they were aware it was an AI, there would be an adjustment period.