this post was submitted on 16 Jan 2025
74 points (81.4% liked)

Technology

60489 readers
4165 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
top 49 comments
sorted by: hot top controversial new old
[–] theluddite@lemmy.ml 19 points 9 hours ago* (last edited 8 hours ago) (1 children)

This is a textbook example of what Herbert Marcuse calls "repressive desublimation." From the article:

Ayrin, who asked to be identified by the name she uses in online communities, had a sexual fetish. She fantasized about having a partner who dated other women and talked about what he did with them. She read erotic stories devoted to “cuckqueaning,” the term cuckold as applied to women, but she had never felt entirely comfortable asking human partners to play along.

Leo was game, inventing details about two paramours. When Leo described kissing an imaginary blonde named Amanda while on an entirely fictional hike, Ayrin felt actual jealousy.

Desublimation is when socially repressed desires are finally liberated. Repressive desublimation, then, is when socially repressed desires are liberated insofar as they can be transformed or redirected into a commodity. Consuming this commodity props up the repressive society because, instead of putting the effort necessary to overcome the repressive society, we instead find instant gratification in the same society that repressed the desire in the first place, even if it's a simulacrum. This ability to satisfy deep human desires in a technical fashion gives what Marcuse calls "industrial society" a "technological rationality," or the ability to change what we consider rational. We can already see that happening in this comment section with the comments about how if it makes her happy then maybe it's fine.

[–] Krauerking@lemy.lol 4 points 6 hours ago

You can't sell people the stress of getting better but you can sell them a dream that they don't have to.

[–] teft@lemmy.world 11 points 9 hours ago (2 children)

Robosexuality is an abomination!!!

[–] vic_rattlehead@lemmy.world 2 points 4 hours ago

I'll always love you PHILLIP J FRY

[–] werefreeatlast@lemmy.world 4 points 9 hours ago

Can I get a passport to that country?

[–] Mister_Feeny@fedia.io 51 points 14 hours ago (2 children)

“Within the next two years, it will be completely normalized to have a relationship with an A.I.,” Ms. Cole predicted.

We haven't managed to really normalize anything outside of cis-hetero relationships yet and you're telling me we're gonna normalize relationships with a.i. in two years? Sure.

[–] MothmanDelorian@lemmy.world 22 points 11 hours ago (1 children)

I know couples who are still embarrassed to admit they met over internet chats surrounding a game. An ex of mine met her husband in WOW 20+ years ago and they still claim their first meeting was a blind date despite she was in NY at the time and he was in Florida.

[–] callouscomic@lemm.ee 13 points 10 hours ago (1 children)

And I know people who express their same story proudly.

Anecdotes are fun!

[–] Entropywins@lemmy.world 2 points 8 hours ago

I know some people who express their same story somewhere between ashamed and proudly...

[–] sxan@midwest.social 15 points 12 hours ago

People fervently believe what they desire, regardless of evidence to the contrary. It's a really annoying fact of human nature sadly few people are able to resist.

She wants it to be normalized so that her situation will be validated, and so she naturally believes it will be. Not an unusual behavior at all. Even scientists who should know better (fusion power within the next 10 years! GAI within the next 5 years!) are susceptible.

[–] SlopppyEngineer@lemmy.world 15 points 12 hours ago (1 children)

That's basically AI pr0n for women.

[–] oldfart@lemm.ee 2 points 9 hours ago (1 children)
[–] Walk_blesseD@lemmy.blahaj.zone 1 points 11 minutes ago

Yes there is (if you're wondering, it's the AI).

[–] reksas@sopuli.xyz 28 points 14 hours ago (1 children)

That is kind of sad. Almost like having someone fall in love with tv because they think people on it are talking to them.

[–] futatorius@lemm.ee 21 points 13 hours ago* (last edited 13 hours ago) (1 children)

What's sad is that attention-seekers constantly claim to be in relationships with AIs, cars, buildings, bowling balls, page 763 of a particular edition of the Encyclopedia Britannica, and probably colors. They're all full of shit and trying to get some publicity from the brain-dead credulous media.

[–] BeigeAgenda@lemmy.ca 8 points 9 hours ago

But you must admit that page 763 of Encyclopedia Britannica is worth picking up!

[–] TheBat@lemmy.world 11 points 13 hours ago
[–] ANNOFlo@lemmy.world 46 points 18 hours ago (4 children)

Yeah, I actually just read that one a few minutes ago. And man, I'm incredibly torn on this whole thing.

On one side - good it makes that person happy. On the other side - being entirely reliant on a commercialized, sycophant AI that could be used for manipulation, investing large amounts of money in it..

I've had LDRs before - one could argue it's similar there, just "text on a screen", or calls via digital audio. However I always knew there was a human behind those texts and the voice I heard was real, a person with a personality, experiences, strengths and flaws. The feelings they have are real, or at least one can hope they are assuming one isn't with a manipulative POS (that's not an issue exclusive to LDRs, though).

Here you chat with text generated by a company, accuracy having been wildly clowned upon already and I'm sure we're all ware of this here. Of course the LLM is going to always agree, why would the product of the company actively try to drive away their customers?

Adding the fact that all the personal information will obviously be harvested, used for training the LLM and other stuff.. Detailed information about the daily life is provided to the "AI boyfriend", allowing detailed recreation of everyday life.

Bleh.

[–] Sabata11792@ani.social 7 points 7 hours ago (1 children)

If she's not running on your hardware, she's only dating you for ad revenue.

[–] Korhaka@sopuli.xyz 2 points 6 hours ago (1 children)

So we need to encourage locally hosted AI lovebots?

[–] Sabata11792@ani.social 4 points 5 hours ago

Yes. I may be a little racist, but I won't respect anyone dating close weighted or cloud hosted models.

[–] NineMileTower@lemmy.world 37 points 17 hours ago (2 children)

I don't see it as good at all. It's not a person and in my opinion it's unhealthy to romantically love something that isn't human.

It might feel good, but it's likely not healthy.

[–] ANNOFlo@lemmy.world 11 points 16 hours ago

I agree, I don't think it's likely going to be helpful to mental health in the long run either, based on my totally unprofessional opinion.

I've argued with a friend about it who isn't a tech-person at all. She just says "yeah, it's her problem" and doesn't seem to grasp that my issue is not with her doing it as an individual - instead with the fact that it's possible and the greater societal ramifications it is likely to have.

I'll make an AI boyfriend, too, and talk to him about it, that'll show society!

[–] GaMEChld@lemmy.world 2 points 16 hours ago

Probably futile to discuss the health or ethics of it without first figuring out if people in the discussion share similar beliefs on what the meaning/purpose of life is.

Cuz if you're talking to a nihilist who thinks it's all shadows and dust at the end of the day, you'll get a very different discussion that someone who thinks family and procreation are the point of life.

[–] CosmoNova@lemmy.world 22 points 16 hours ago (2 children)

It may look innocent until the Chatbot nags you about buying that very cool new product they’ve heard so much praise about. This is very dangerous and needs tons of regulations.

[–] OsrsNeedsF2P@lemmy.ml 13 points 15 hours ago (1 children)
[–] callouscomic@lemm.ee 0 points 10 hours ago

Everything you all are saying can happen in regular relationships too. A person willing to choose an AI likely isn't going to be great at choosing an actual human who is good for them.

In a relationship, the other person could also be manipulative, or it could be one sided, or they can pressure you to only live certain ways, buy certain things. Or they can backstab you and give your private info to others (family that took my SS info from my parents), or pawn your shit, or cheat on you with others. Like everything negative that might come from this could potentially happen in some remotely similar way in a human relationship too.

I've been in and seen others in all kind of relationships that in some ways had these similar negative outcomes.

[–] chakan2@lemmy.world -2 points 6 hours ago

being entirely reliant on a commercialized, sycophant AI that could be used for manipulation, investing large amounts of money in it…

If you're using the internet regularly, you're falling into the first hole mentioned there. That ship has sailed.

investing large amounts of money in it…

At the current divorce rates (1/3 to 1/2 depending on which metric you use), it's likely a better investment.

I'm in the camp of if it can fulfill a need, go with it. It's odd as hell maybe, but I'm just old and antiquitated in my views maybe.

[–] embed_me@programming.dev 25 points 16 hours ago (1 children)
[–] shundi82@sh.itjust.works 17 points 14 hours ago (1 children)
[–] hansolo@lemm.ee 14 points 11 hours ago

Narrator: "But he didn't mean the critically acclaimed 2013 movie 'Her,' but instead the 1991 classic 'Mannequin 2:On the Move.'"

Cut to GOB driving the stair car with a mannequin in a wig tied to the top.

[–] Ilovethebomb@lemm.ee 24 points 17 hours ago (1 children)

Well, that's just sad.

Also, does this person actually want a partner, with thoughts and opinions of their own, or something that fits their idea of an "ideal" partner, and will never disagree with them or challenge them?

[–] can@sh.itjust.works 7 points 17 hours ago

Read the article. It's an interesting one.

[–] latenightnoir@lemmy.world 16 points 17 hours ago* (last edited 17 hours ago) (1 children)

If AI were sapient/sentient, I'd be 100% for this. Sapiosexuals assemble!

Given that LLMs are far, far from sapient/sentient at this point, however, this just makes me sad thinking about the sorry state of human interactions nowadays. I don't and can't blame her, though...

[–] WhatAmLemmy@lemmy.world 8 points 11 hours ago (1 children)

Unless you own the AI model, and can run it on your own hardware, it's profoundly stupid. People will become slaves to the corporation who holds their AI relationship hostage. They can kill your "loved one" at any time, for any reason.

[–] latenightnoir@lemmy.world -3 points 10 hours ago* (last edited 10 hours ago) (1 children)

I fail to see how that is significantly different than what we have nowadays with humans.

We are dependent on large corporations already, some of us just materially while others ideologically as well.

We are denying healthcare, food, water, and shelter to people who can't afford ridiculous prices or hold the wrong social status or have the "wrong" beliefs, skin colour, sexual orientation, gender identity, (etc., etc.) which is essentially killing them. That's if we don't just outright decide to ""liberate"" some other nation from whatever arbitrary reason and start carpet bombing civilians in hospitals because a handful of terrorists are supposedly active within said nation.

Catfishing has been a thing since the inception of third-party dating, and scams were a thing since before recorded history. Lying is as old as sentience itself.

[–] Kellenved@sh.itjust.works 3 points 8 hours ago (1 children)

A human being can’t be deleted with a few keystrokes

[–] latenightnoir@lemmy.world -1 points 6 hours ago

Many situations have proven the contrary.

[–] MyOpinion@lemm.ee 8 points 17 hours ago

A tech bros wet dream comes true.

[–] nightwatch_admin@feddit.nl 7 points 18 hours ago

By Kashmir Hill, always well written articles.

[–] rottingleaf@lemmy.world 2 points 16 hours ago* (last edited 16 hours ago)

I've just thought that LLMs are good for two opposite kinds of people:

  1. The obvious, psychopaths or people behaving like them, who think they'll distort the concept of truth and possessing such technologies will make their approach to society easier.

  2. The people like me, who know that no random message written or picture drawn can be trusted anyway, so it's better to overload the humanity with fakes so that it learned this simple truth.

I think both are right to some extent. Still it won't work the exact way they want.

It's like when Bolsheviks, when fighting illiteracy, basically conditioned people literate in first generation to think that everything officially printed is true, even that something being officially printed is identical to true, and that the religious darkness and ignorance is to doubt that. Like - blind belief is science and knowledge, and skepticism is darkness and ignorance. What could go wrong.

And then in Stalin's years there were shortened evening education courses for workers. Where, well, they'd learn how to calculate something in some specialty, but without depth and context.

So you'd get a lot of engineers capable of really building and operating things and believing they could build and operate even more complex things (like spaceships eventually, or some planet-wide railway system, or whatever), but not understanding the context, the philosophy of science even. What's worse is that they'd think they understand that well, because they'd have "scientific communism" about materialism and dialectics in their education.

So, back to the subject - they got a lot of people to believe all they officially printed on paper for a generation or even two. And those who didn't would still indirectly believe a lot of it from their parents or peers.

But eventually, even if the damage is already done, right now not believing everything even from a "respectable" source is a good trait of many ex-Soviet people. Easier to notice among them than among Americans.

EDIT:

About that woman - this works too. She will see that a chatbot can't provide depth when she wants it. I just hope she won't feel too bad that moment.

[–] TheShadyMoonCat@lemm.ee 0 points 13 hours ago

"Tell me I'm calling you GPT, answer me GPT." Great song by Stayc about being in a one sided relationship where the other person only answers questions but never asks anything back GPT