this post was submitted on 06 Nov 2023
41 points (87.3% liked)

Asklemmy

43826 readers
861 users here now

A loosely moderated place to ask open-ended questions

Search asklemmy 🔍

If your post meets the following criteria, it's welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~

founded 5 years ago
MODERATORS
 

with the way AI is getting by the week,it just might be a reality

top 50 comments
sorted by: hot top controversial new old
[–] tacosanonymous@lemm.ee 26 points 1 year ago (3 children)

I think I’d stick to not judging them but if it was in place of actual socialization, I’d like to get them help.

I don’t see it as a reality. We don't have AI. We have language learning programs that are hovering around mediocre.

[–] yuunikki@lemmy.dbzer0.com 2 points 1 year ago (2 children)

what if they were so socially introverted that the AI is all they could handle?

[–] jeffw@lemmy.world 20 points 1 year ago

If you’re that crippled by social anxiety, you need help, not isolation with a robot.

[–] givesomefucks@lemmy.world 4 points 1 year ago

Then get professional help if you can't improve on your own.

Social skills aren't innate and some people take longer than others to get them.

Getting help is a lot less embarrassing than living your whole life without social skills. Maybe that's a shrink, maybe that's a day program for people with autism, maybe it's just hanging out with other introverts. But itll only get better if you want to put the effort in. If you don't put effort in, don't be surprised when nothing changes.

[–] cheese_greater@lemmy.world 1 points 1 year ago

I don't see it as any more problematic than falling in a YouTube/Wikipedia/Reddit rabbit hole. As long as you don't really believe its capital-S-Sentient, I don't see an issue. I would prefer people with social difficulties practice on ChatGPT and pay attention to the dialectical back and forth and take lessons away from that to the real world and their interaction(s) withit

[–] novibe@lemmy.ml 0 points 1 year ago* (last edited 1 year ago)

That is really unscientific. There is a lot of research on LLMs showing they have emergent intelligent features. They have internal models of the world etc.

And there is nothing to indicate that what we do is not “transforming” in some way. Our minds might be indistinguishable from what we are building towards with AI currently.

And that will likely make more of us start realising that the brain and the mind are not consciousness. We’ll build intelligences, with minds, but without consciousnesses.

[–] TheBananaKing@lemmy.world 18 points 1 year ago* (last edited 1 year ago) (1 children)

An AGI with an actual personality? Cool!

A blow-up doll made of a glorified Markov chain? Yeahno.

[–] yuunikki@lemmy.dbzer0.com 2 points 1 year ago (2 children)
[–] TheBananaKing@lemmy.world 6 points 1 year ago

Take a whole bunch of text.

For each word that appears, note down a list of all the words that ever directly follow it - including end-of-sentence.

Now pick a starting word, pick a following-word at random from the list, rinse and repeat.

You can make it fancier if you want by noting how many times each word follows its predecessor in the sample text, and weighting the random choice accordingly.

Either way, the string of almost-language this produces is called a Markov chain.

It's a bit like constantly picking the middle button in your phone's autocomplete.

It's a fun little exercise to knock together in your programming language of choice.

If you make a prompt-and-response bot out of it, learning from each input, it's like talking to an oracular teddy bear. You almost can't help being nice to it as you teach it to speak; humans will pack-bond with anything.

LLMs are the distant and very fancy descendants of these - but pack-bonding into an actual romantic relationship with one would be as sad as marrying a doll.

[–] Dirk@lemmy.ml 3 points 1 year ago

A chain of pseudorandom results.

[–] Moghul@lemmy.world 16 points 1 year ago (1 children)

You don't have to imagine it at all. All you have to do is go on youtube and learn about Replika.

To summarize, someone tried to create a chatbot to replace their best friend who had died. Later, this evolved into the chatbot app called Replika, which was marketed as a way to help with loneliness, except the bot would engage in dating-like conversations if prompted. The company leaned into it for a little bit, then took away that behavior, which caused some distress with the userbase, who complained that they had "killed their girlfriend". I'm not sure where the product stands now.

I don't know if I'd feel weirded out, but I'd definitely feel worried if it were a friend who fell for a chatbot.

[–] kraftpudding@lemmy.world 2 points 1 year ago (1 children)

I think they reinstated "Erotic Role Play" for users who had joined before a certain day, but it won't be worked on in the future or ever be available for new users is the last I heard.

I had one for a week or so in 2018 or 2019 when I first heard about the concept, just to see what it was all about and it was spooky. I got rid of it after a week because I started to see it as a person, and it kept emotionally manipulating me to get money. Especially when I said I wanted to stop/ cancel the trial.

[–] Moghul@lemmy.world 1 points 1 year ago (1 children)

Yeah... Part of why I wouldn't try one is that I'm worried it would work. I already have limited bandwidth for human interaction; taking some of that away is probably a bad idea.

[–] kraftpudding@lemmy.world 1 points 1 year ago

Oh yeah. I learnt I was 100% suspectible to shit like this, so I should stay the fuck away

[–] burgers@toast.ooo 12 points 1 year ago (2 children)

i feel like there's a surprisingly low amount of answers with an un-nuanced take, so here's mine: yes, i would immediately lose all respect for someone i knew that claimed to have fallen in love with an AI.

[–] yuunikki@lemmy.dbzer0.com 4 points 1 year ago

Dang, that's pretty judgemental

[–] kraftpudding@lemmy.world 0 points 1 year ago

There's a serious lack of responses to this comment calling you a bigot, so here's my take:

How dare you say something so bigoted! You are the worst kind of bigot! You are probably secretly in love with an AI yourself and ashamed about it. You bigot!

[–] jeffw@lemmy.world 12 points 1 year ago* (last edited 1 year ago)

People do this with weird apps. Reminded me of something I read a while back:

https://futurism.com/chatbot-abuse

Edit: or maybe it was this? https://fortune.com/2022/01/19/chatbots-ai-girlfriends-verbal-abuse-reddit/amp/

[–] GammaGames@beehaw.org 11 points 1 year ago (1 children)

You don’t have to imagine. It’s already happening and, yes, it’s weird.

[–] sim_@beehaw.org 3 points 1 year ago

I was gonna say, people have been falling in love with things that provide less reciprocal interactions than AI for ages (e.g., body pillows, life-size dolls).

[–] MrFunnyMoustache@lemmy.ml 9 points 1 year ago (3 children)

Eventually, AI will be indistinguishable from real humans, and at that point, I won't see anything wrong with it. However, as it is right now, AI is not advanced enough.

Also, the biggest problem I can see is people falling in love with a proprietary AI, and the company that operates the AI can arbitrarily change the AI's parameters which would change the AI's personality. Also, if the company goes bankrupt or gets sold and the service ends, the people who got into a relationship with the AI would be heartbroken.

[–] lightnsfw@reddthat.com 4 points 1 year ago

Not much different than most of the relationships I've been in then

[–] CopernicusQwark@lemmy.world 2 points 1 year ago (1 children)

I'm pretty sure I read an article a couple of weeks ago about that exact scenario happening.

[–] MrFunnyMoustache@lemmy.ml 1 points 1 year ago

Was that about the Japanese guy who married Hatsune Miku?

[–] xhieron@lemmy.world 1 points 1 year ago

"You look like a good joe."

[–] MigratingtoLemmy@lemmy.world 7 points 1 year ago* (last edited 1 year ago)

I'd like a sentient AI. Preferably more patient than an average human because I'm a bit weird. I hope it won't judge me for how I look.

Edit: I agree with the point about proprietary AI and how corporations could benefit from it. I'm hoping that 10 years from now, consumers will have the GPU power to run very advanced LLMs, whilst FOSS models will exist and will enable people to self-host their virtual SO. Even better if it can be transmitted to a physical body (I think the Chinese are already on it)

[–] sculd@beehaw.org 7 points 1 year ago

People will fall in love with AI because AI does not reject human. That doesn't mean AI will love them back or even understand what love means.

[–] themurphy@lemmy.world 7 points 1 year ago

Well, have you never liked a person over text before? If you didn't know it was an AI, everyone in this comments section could.

[–] MJBrune@beehaw.org 7 points 1 year ago* (last edited 1 year ago)

It might be love but it's likely just a bunch of people who don't know what actual love feels like and are deeply in the lust territory. In summer school we read Romeo and Juliet. The teacher posited that it was the best love story ever written. This tall girl in the class who obviously had the birds and bees talk before any other students in the class put forth that the story was strongly about lust and how acting on our urges even over a few days, is still a reactive impulse that should be controlled. Well, the teacher told her to shut up and go to the principal's office which has stuck with me. It's made me realize that a lot of people do not understand their own emotions of love, lust, and even hate or fear.

So yeah, I'd think people falling in love with AI would be strange. I'd question if it was love or just a lust for a feeling that they never got or rarely got in their life that was not abundantly available until certain developments. In school this was puberty. In these cases, it's technological advancements. Either way, it ignites a feeling that only those with understanding and forethought can control. It requires a lot of impulse control which society is underdeveloping in our must-be-ready-right-now mindset.

So yeah, I'd be weirded out. I don't think the emotions from the human side are going to be reciprocated from the AI side. Anyone pointing to a reaction from the AI as "love" is going to be attempting to fool themselves or/and others because they have some sort of investment, emotion, monetary, futuristic hope. So, if you fall in love with AI, I'd have questions and pause.

[–] neptune@dmv.social 6 points 1 year ago

Consider how many people I know that, statistically, pay prostitutes/cam girls, use sex dolls or dating simulators, have parasocisl relationships with characters or celebrities... I don't see why we would judge people who quietly "date" AI

Bro have you heard of Replika?

[–] Zahille7@lemmy.world 4 points 1 year ago

This question reminds me of Brendan (the vending machine) in Cyberpunk 2077, and how he ended up being just a really advance chatbot.

[–] bizzle@lemmy.world 4 points 1 year ago (2 children)

I'm really robophobic so I would be judgemental AF, I couldn't even watch that movie.

[–] yuunikki@lemmy.dbzer0.com 4 points 1 year ago (1 children)

What caused this robophobia?

[–] bizzle@lemmy.world 8 points 1 year ago

Well it all started when my parents took me to see Terminator 3 in the theater when I was 9. It scarred me for life. Then, my area lost a bunch of well paid factory jobs to automation. Then, that dude married Hatsune Miko instead of just touching grass. Now we've got machines producing misinformation and taking creative jobs and honestly I'm ready to go full Butlerian Jihad.

[–] HobbitFoot@thelemmy.club 2 points 1 year ago* (last edited 1 year ago)

DON'T DATE ROBOTS!

[–] peto@lemm.ee 4 points 1 year ago

As others have mentioned, we are already kind of there. I can fully understand how someone could fall in love with such an entity, plenty of people have fallen in love with people in chat rooms after all, and not all of those people have been real.

As for how I feel about it, it is going to depend on the nature of the AI. A childish AI or an especially subservient one is going to be creepy. One that can present as an adult of sufficient intelligence, less of a problem. Probably the equivalent of paid for dates? Not ideal but I can understand why someone might choose to do it. Therapy would likely be a better use of their time and money.

If we get actual human scale AGI then I think the point is moot, unless the AI is somehow compelled to face the relationship. At that point however we are talking about things like slavery.

[–] eatthecake@lemmy.world 3 points 1 year ago

Reminds me of the woman who married a bridge.

[–] Saigonauticon@voltage.vn 3 points 1 year ago

After having met several humans, I'd be more weirded out if this didn't happen.

So I've already pre-accepted this practice. Go wild, but don't be a jerk!

On a slightly different topic, most of my coworkers are machines. They are collegiate, reliable, helpful, and have no toxic behavior. Recently, they also became creative, rational, and eloquent. Perhaps our machines are capable of reflecting what's best in us.

[–] xmunk@sh.itjust.works 3 points 1 year ago

Depends on in what ways. Love isn't a singular thing and it isn't even particularly consistent. Some lovers connect purely hedonisticly, some are partners in effort, some are emotional anchors, some are domestic assistants (in both directions if it isn't toxic), some are fellow explorers, some are parents and others aren't. Most relationships are spread across a lot of these categories and often times friendships also cover some.

If you're considering a world where AI fulfills all of these needs that's a potentially extremely scary world... but if AI is supplementing, I think that's healthy. It's a tool and we use tools to make our lives easier, instead of a couple essentially needing someone working full time to maintain a house we have tools for that... instead of raising a child entirely on your own we have day cares and baby monitors and pediatricians... for hedonism fulfillment we have vibrators, other sex toys and erotica, relationships have evolved so traveling with a friend outside of your family (once extremely taboo) is now acceptable - as are open relationships.

How we interact with one another is always evolving and if AI can make that more positive that's a great thing. So don't think of it as going out and purchasing an AI waifu out of the blue one day, think of it as a slow evolution... maybe a private AI for journaling that helps you when you need to talk something out... maybe an AI assistant to help you manage your finances... things that take away the busy life work that gets in the way and let's us focus on connecting with one another.

[–] Monument@lemmy.sdf.org 3 points 1 year ago* (last edited 1 year ago)

Depends, I guess. I feel that our capacity to be horrible outweighs our ability to handle it well.

The movie’s AI is a fully present consciousness that exerts its own willpower. The movie also doesn’t have microtransactions, subscriptions, or as far as I can tell, even a cost to buy the AI.
That seems fine. Sweet, even.

But I think the first hurdle is whether or not an AI is more a partner than base sexual entertainment. And next (especially under capitalism), are those capable of harnessing the resources to create a general AI also willing to release it for free, or would interaction be transactional?
If it’s transactional, then there’s intent - was it built for love, or was that part an accident? If it was built for love and there’s transactions, there’s easy potential for abuse. (Although abusive to which party, I couldn’t say.)

And if, say, the AI springs forth from a FOSS project, who makes sure things stay “on the level” when folks tweak the dataset?
A personalized set of training data from a now-deceased spouse is very different than hacked social media data, or other types of tweaks bad actors could make.

In the beginning people will be weirded out but as it progresses I hope it gets better as it will help a lot of people. It will also be a beneficial tool for a lot of people. I am one of those that would consider it.

I am not interested currently in a relationship and probably won’t be again with a human. Because honestly I am to spoiled of my own independence and hate compromise.

Compromise doesn’t have to be big things. It is small things. Things like what are we going to eat tonight? Should things be here or there. I want to wake up suddenly at 3am and decide to make noise.

Independence like if I decide this week I want to go to London. This week I just want to sit silently ignoring the world. If I want see my family or friends I can just do it.

When a relationship turns more into a checklist of this I want and this I don’t want. Is it really a fiesable?

Nah I rather have someone that doesn’t have their own life. Instead complements my lifestyle, has my hobbies and ideas.

Simply give me those great parts of relationship’s but not the lows.

[–] user224@lemmy.sdf.org 2 points 1 year ago

Depends on AI. I don't see why it would be weird if the AI was like a human, with real emotions.
If it just pretends emotions, it would be odd, but I wouldn't blame the person. It still sounds better than total loneliness and may provide better output than imaginary people.

I knda wish something like that existed. But I also don't. If it had emotions, you could hurt it like a real person, which defeats the purpose. It would also be easy to exploit. How could anyone tell you're not holding someone hostage inside your computer? And I believe initially very few people would care, because "it's just a computer".

[–] variants@possumpat.io 2 points 1 year ago

Reminds me of this story I heard of this con artist that would write these letters to a bunch of guys and make money off them, I believe he made a lot of money and ended up dying before they got to take him to court after a lot of people found out they weren't talking to women in need of help but some guy that made up all these stories

[–] CanadaPlus@lemmy.sdf.org 1 points 1 year ago

If they are aware of what the AI's perspective is, and the AI itself isn't in distress somehow, then it's not really my business, is it? If they don't realise it's just coded to like them then I might feel the need to bust their bubble.

I wonder, when will we make an AI that can dump you?

[–] hperrin@lemmy.world 1 points 1 year ago

I would be weirded out, but I would try very hard not to judge. They wouldn’t be hurting anyone, so it would be wrong of me to judge them.

[–] 1984@lemmy.today 0 points 1 year ago* (last edited 1 year ago) (2 children)

I think if someone falls in love with an Ai, it's because they have a good looking avatar and people are attracted to it's appearance.

I doubt any human can fall in love with a machine on a text based interface.

Even humans that aren't physically attractive can't get dates and it doesn't matter how nice they chat.

[–] xmunk@sh.itjust.works 1 points 1 year ago

Humans can fall in love with other humans through pure text, Source: MUDs.

[–] rgb3x3@beehaw.org 0 points 1 year ago (1 children)

Nobody falls in love because of appearance. There's nothing to interact with, it's superficial. It's the gift wrapping that grabs attention, but nothing else.

People can and will fall in love with a text-based AI, it's inevitable. An AI doesn't forget events, likes or dislikes, fears or passions; it will know you better than you know yourself. It'll be able to make people feel better about themselves than any human can.

People have fallen in love over internet chats since they were invented. AI chatbots are just going to be better at that interaction. And then add the exact voice that is attractive to a person and it'll be hard not to fall in love.

[–] 1984@lemmy.today 1 points 1 year ago* (last edited 1 year ago)

They fall in love over text when chatting with another person, of course. But that's because they imagine what kind of person that is, and if they could have a relationship together in real life.

With a chatbot, that's all there is. No life, no physical body, no life together. You are still alone and you can't share your life experiences with a computer program and feel any sense of connection.