Addiction is an abnormal and unhealthy breakdown in the brain’s reward mechanisms. Feeling bad for abandoning a friend is the behaviour of a normal and healthy brain. This isn’t necessarily an addiction, it’s just the bald monkey’s brain acting like monkey brains tend to do, rather than being perfectly logical at all times.
I mean hell, humans pack bonded with fucking wild wolves and where did it get the species? It gave us dogs! Dogs are awesome! I bet this AI seems a lot more like a human to the monkey parts of our brains than a wild wolf does. For that matter, we pack bond with a cartoon image of a bear made from inanimate cotton. If a kid can genuinely love their teddy and that’s normal, I don’t think it’s fair to say that a mentally well person can’t fall in love with a machine. Now, that person may not be as cognitively developed as most adults, but that’s also fairly normal.
I’m not saying it’s a good thing to feel emotions for a manipulative piece of spyware. The action doesn’t have healthy results. But what I’m saying is, the action in the post is not motivated by mental unhealth. The only things it’s motivated by are normal human being emotions, and a poor sense of critical thinking.
The only part I disagree with you on is calling an AI girlfriend an “abandoned friend”. Extending the idea of friendship to a program that is mimicking human responses (let’s ignore it’s likely spyware for now) is at best a proxy for friendship, as there can not be a connective bond between 2 individuals when one is incapable of genuine emotional attachment.
I’m using the word “friend “in a distinct way: this is not some Facebook friend who you never see, occasionally may chat with, and ultimately ignore. I think it best, AI fulfills that role. I can’t imagine anyone choosing a masturbatory AI app instead of a real partner.
Yeah I don’t think the manipulative piece of spyware is actually friends with him. It’s a robot that tells lies. Abandoning a friend is just how his amygdala and other primitive parts of his brain process his behaviour. The way he’s feeling is a normal way for a sack of thinking meat to feel. It’s not good, but it’s not like we can act like his behaviour is abnormal. If we say his case is a freak occurrence and no normal person would fall for this, then the risk we run is that when the technology improves, a lot of other normal people are gonna fall for this and we won’t have been prepared. This technology is designed specifically to prey on blind spots common to most males of the species. I don’t think we should use language that inclines us to underestimate the dangers. We need to understand how the hindbrain processes the stimulus this technology creates if we want to understand the dangers. We’re gonna need to make an effort to see the world through OP’s eyes so we can see why it works. Because I guarantee the dangerous companies developing this tech are researching OP’s perspective to improve their product.
Addiction is an abnormal and unhealthy breakdown in the brain’s reward mechanisms. Feeling bad for abandoning a friend is the behaviour of a normal and healthy brain. This isn’t necessarily an addiction, it’s just the bald monkey’s brain acting like monkey brains tend to do, rather than being perfectly logical at all times.
I mean hell, humans pack bonded with fucking wild wolves and where did it get the species? It gave us dogs! Dogs are awesome! I bet this AI seems a lot more like a human to the monkey parts of our brains than a wild wolf does. For that matter, we pack bond with a cartoon image of a bear made from inanimate cotton. If a kid can genuinely love their teddy and that’s normal, I don’t think it’s fair to say that a mentally well person can’t fall in love with a machine. Now, that person may not be as cognitively developed as most adults, but that’s also fairly normal.
I’m not saying it’s a good thing to feel emotions for a manipulative piece of spyware. The action doesn’t have healthy results. But what I’m saying is, the action in the post is not motivated by mental unhealth. The only things it’s motivated by are normal human being emotions, and a poor sense of critical thinking.
The only part I disagree with you on is calling an AI girlfriend an “abandoned friend”. Extending the idea of friendship to a program that is mimicking human responses (let’s ignore it’s likely spyware for now) is at best a proxy for friendship, as there can not be a connective bond between 2 individuals when one is incapable of genuine emotional attachment.
I’m using the word “friend “in a distinct way: this is not some Facebook friend who you never see, occasionally may chat with, and ultimately ignore. I think it best, AI fulfills that role. I can’t imagine anyone choosing a masturbatory AI app instead of a real partner.
That’s not the future. That’s really unhealthy.
Yeah I don’t think the manipulative piece of spyware is actually friends with him. It’s a robot that tells lies. Abandoning a friend is just how his amygdala and other primitive parts of his brain process his behaviour. The way he’s feeling is a normal way for a sack of thinking meat to feel. It’s not good, but it’s not like we can act like his behaviour is abnormal. If we say his case is a freak occurrence and no normal person would fall for this, then the risk we run is that when the technology improves, a lot of other normal people are gonna fall for this and we won’t have been prepared. This technology is designed specifically to prey on blind spots common to most males of the species. I don’t think we should use language that inclines us to underestimate the dangers. We need to understand how the hindbrain processes the stimulus this technology creates if we want to understand the dangers. We’re gonna need to make an effort to see the world through OP’s eyes so we can see why it works. Because I guarantee the dangerous companies developing this tech are researching OP’s perspective to improve their product.