this post was submitted on 28 Jan 2025
2 points (100.0% liked)

Technology

38083 readers
391 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 3 years ago
MODERATORS
 

The original post is from r/Romania, but I thought it would be interesting to share this here too.

top 21 comments
sorted by: hot top controversial new old
[–] gerryflap@feddit.nl 1 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

Although censorship is obviously bad, I'm kinda intrigued by the way it's yapping against itself. Trying to weigh the very important goal of providing useful information against its "programming" telling it not to upset Winnie the Pooh. It's like a person mumbling "oh god oh fuck what do I do" to themselves when faced with a complex situation.

[–] Kacarott@aussie.zone 0 points 3 weeks ago (1 children)

I know right, while reading it I kept thinking "I can totally see how people might start to believe these models are sentient", it was fascinating, the way it was "thinking"

[–] TanyaJLaird@beehaw.org 0 points 3 weeks ago (1 children)

It reminds me one of Asimov's robots trying to reason a way around the Three Laws.

[–] crapwittyname@lemm.ee 1 points 3 weeks ago

"That Thou Art Mindful of Him" is the robot story of Asimov's that scared me the most, because of this exact reasoning happening. I remember closing the book, staring into space and thinking 'shit....we are all gonna die'

[–] xia@lemmy.sdf.org 1 points 3 weeks ago

Knowing how it works is so much better than guessing around OpenAI's censoring-out-the-censorship approach. I wonder if these kind of things can be teased out, enumerated, and then run as a specialization pass to nullify.

[–] melp@beehaw.org 1 points 3 weeks ago

Bing's Copilot and DuckDuckGos ChatGPT are the same way with Israel's genocide.

[–] Sauerkraut@discuss.tchncs.de 0 points 3 weeks ago* (last edited 3 weeks ago) (3 children)

I don't understand how we have such an obsession with Tiananmen square but no one talks about the Athens Polytech massacre where Greek tanks crushed 40 college students to death. The Chinese tanks stopped for the man in the photo! So we just ignore the atrocities of other capitalist nations and hyperfixate on the failings of any country that tries to move away from capitalism???

[–] Kazumara@discuss.tchncs.de 1 points 3 weeks ago

The Chinese tanks stopped for the man in the photo!

What a line dude.

The military shot at the crowd and ran over people in the square the day before. Hundreds died. Stopping for this guy doesn't mean much.

[–] TheOakTree@lemm.ee 1 points 3 weeks ago* (last edited 3 weeks ago)

I think the argument here is that ChatGPT will tell you about Kent State, Athens Polytech, and Tianenmen square. Deepseek won't report on Tianenmen, but it likely reports on Kent State and Athens Polytech (I have no evidence). If a Greek AI refused to talk about the Athens Polytech incident, it would also raise concerns, no?

ChatGPT hesitates to talk about the Palestinian situation, so we still criticize ChatGPT for pandering to American imperialism.

[–] williams_482@startrek.website 0 points 3 weeks ago (1 children)

Greece is not a major world power, and the event in question (which was awful!) happened in 1974 under a government which is no longer in power. Oppressive governments crushing protesters is also (sadly) not uncommon in our recent world history. There are many other examples out there for you to dig up.

Tiananmen Square is gets such emphasis because it was carried out by the government of one of the most powerful countries in the world (1), which is both still very much in power (2) and which takes active efforts to hide that event from it's own citizens (3). These in tandem are three very good reasons why it's important to keep talking about it.

[–] Sauerkraut@discuss.tchncs.de 1 points 2 weeks ago (1 children)

Hmm. Well, all I can say is that the US has commited countless atrocities against other nations and even our own citizens. Last I checked, China didn't infect their ethnic minorities with Syphilis and force the doctors not to treat it under a threat of death, but the US government did that to black Americans.

[–] williams_482@startrek.website 2 points 2 weeks ago

You have no idea if China did that. If they had, they would have taken great efforts to cover it up, and could very well have succeeded. It's a small wonder we know any of the terrible things they did, such as the genocide they are actively engaging in right now.

[–] Moonrise2473@feddit.it 0 points 3 weeks ago (1 children)
[–] jarfil@beehaw.org 1 points 3 weeks ago

Nah, just being "helpful and harmless"... when "harm" = "anything against the CCP".

[–] drspod@lemmy.ml 0 points 3 weeks ago (1 children)

I thought that guardrails were implemented just through the initial prompt that would say something like "You are an AI assistant blah blah don't say any of these things..." but by the sounds of it, DeepSeek has the guardrails literally trained into the net?

This must be the result of the reinforcement learning that they do. I haven't read the paper yet, but I bet this extra reinforcement learning step was initially conceived to add these kind of censorship guardrails rather than making it "more inclined to use chain of thought" which is the way they've advertised it (at least in the articles I've read).

[–] iii@mander.xyz 0 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

Most commercial models have that, sadly. At training time they're presented with both positive and negative responses to prompts.

If you have access to the trained model weights and biases, it's possible to undo through a method called abliteration (1)

The silver lining is that a it makes explicit what different societies want to censor.

[–] SnotFlickerman@lemmy.blahaj.zone 0 points 3 weeks ago* (last edited 3 weeks ago) (1 children)

Hi I noticed you added a footnote. Did you know that footnotes are actually able to be used like this?[^1]

[^1]: Here's my footnote

Code for it looks like this :able to be used like this?[^1]

[^1]: Here's my footnote

[–] abfarid@startrek.website 0 points 3 weeks ago (1 children)

Do you mean that the app should render them in a special way? My Voyager isn't doing anything.

[–] SnotFlickerman@lemmy.blahaj.zone 0 points 3 weeks ago (1 children)

I actually mostly interact with Lemmy via a web interface on the desktop, so I'm unfamiliar with how much support for the more obscure tagging options there is in each app.

It's rendered in a special way on the web, at least.

[–] codexarcanum@lemmy.dbzer0.com 0 points 3 weeks ago (1 children)

That's just markdown syntax I think. Clients vary a lot in which markdown they support though.

markdown syntax

yeah I always forget the actual name of it I just memorized some of them early on in using Lemmy.