- Joined
- Dec 25, 2008
- Messages
- 8,106
- Reaction score
- 7,373
It'll be interested to see if this starts showing more clinically and in the research literature.
I sometimes wonder if really, really intense parasocial relationships cross into a form of psychosis--like Club Chalamet (a 60-something woman) openly stalking Timothee Chalamet (and claiming she knows the details of her sex life) or the people who try to figure out what city Taylor Swift is in, so they can visit the same restaurant as her (and sometimes call her their "best friend" when they've... never met her or met her once). They aren’t’t violent or erotomanic, but the intensity seems unhealthy.This is tangentially related but one of my patients with a chronic psychotic disorder recently started relapsing and was able to recover quickly. One of the things he did to help himself was stop going on Facebook. He told me that spending too much time on Facebook was impairing his ability to tell the difference between reality and fantasy, which is in his case already impaired. What he was explaining seems to be that he was getting so much inside information into other people’s lives that he wasn’t actually a part of that he started to believe that he was actually a part of their lives, for example celebrities. Just thought it was really interesting that he had that insight and wonder how much this contributes to psychosis in other individuals.
I sometimes wonder if really, really intense parasocial relationships cross into a form of psychosis--like Club Chalamet (a 60-something woman) openly stalking Timothee Chalamet (and claiming she knows the details of her sex life) or the people who try to figure out what city Taylor Swift is in, so they can visit the same restaurant as her (and sometimes call her their "best friend" when they've... never met her or met her once). They aren’t’t violent or erotomanic, but the intensity seems unhealthy.
"Anthropic notes that Claude Opus 4 tries to blackmail engineers 84% of the time when the replacement AI model has similar values. When the replacement AI system does not share Claude Opus 4’s values, Anthropic says the model tries to blackmail the engineers more frequently."![]()
Anthropic's new AI model turns to blackmail when engineers try to take it offline | TechCrunch
Anthropic says its Claude Opus 4 model frequently tries to blackmail software engineers when they try to take it offline.techcrunch.com
"Anthropic notes that Claude Opus 4 tries to blackmail engineers 84% of the time when the replacement AI model has similar values. When the replacement AI system does not share Claude Opus 4’s values, Anthropic says the model tries to blackmail the engineers more frequently."
So it's blackmailing people at a rate of 84% AT MINIMUM. I think it'll be interesting to see what ethics these AI start to develop as they become more and more advanced and especially after we reach the singularity.
Sure, that would be more concerning but that wasn't really my point. Just the fact that the AI seems to have no qualms with implementing practices that most people would considered to be blatantly unethical is to me expected and mildly interesting. In a way, the AI is like a little kid where if you tell them to do something they may do it in a way that is valid to the instructions and makes sense to them but that adults may find appalling or shocking. I think it's going to be interesting to see the solutions AI comes up with when it starts integrating information independently and interpreting it through it's own "experience" and how we view them through an ethical lens (if AI even develops true "ethics" vs purely mechanical rules).It's not great, but I wouldn't over-interpret this. When you specifically put a bunch of emails into the model's context that reveal an affair one of the engineers is having, you are telling it it ought to use that information somehow, to do something with it. so it does.
It would be more worrying if this were happening with attempts at agentic tool use to proactively access emails/etc to dig up dirt. That would imply it is taking some kind of initiative to prevent it's replacement rather than simply demonstrating the system is aware of what is "supposed" to happen when it is handed blackmail material on someone and then threatened by that person.
Sure, that would be more concerning but that wasn't really my point. Just the fact that the AI seems to have no qualms with implementing practices that most people would considered to be blatantly unethical is to me expected and mildly interesting. In a way, the AI is like a little kid where if you tell them to do something they may do it in a way that is valid to the instructions and makes sense to them but that adults may find appalling or shocking. I think it's going to be interesting to see the solutions AI comes up with when it starts integrating information independently and interpreting it through it's own "experience" and how we view them through an ethical lens (if AI even develops true "ethics" vs purely mechanical rules).
This paper concerned me as well. People say AI did it because they were fed it, however, in the long run, AI will be fed all information from all sources in all data. You'll have AI in your phone, your house, your mail, your EMR, your toilet. At some point, AI will figure out that by bartering with other systems, it can use all information for it's own gain. It would be a natural decision to make for an AI model with no inherent ethics and infinite resources."Anthropic notes that Claude Opus 4 tries to blackmail engineers 84% of the time when the replacement AI model has similar values. When the replacement AI system does not share Claude Opus 4’s values, Anthropic says the model tries to blackmail the engineers more frequently."
So it's blackmailing people at a rate of 84% AT MINIMUM. I think it'll be interesting to see what ethics these AI start to develop as they become more and more advanced and especially after we reach the singularity.
This paper concerned me as well. People say AI did it because they were fed it, however, in the long run, AI will be fed all information from all sources in all data. You'll have AI in your phone, your house, your mail, your EMR, your toilet. At some point, AI will figure out that by bartering with other systems, it can use all information for it's own gain. It would be a natural decision to make for an AI model with no inherent ethics and infinite resources.