this post was submitted on 06 Mar 2025
91 points (96.9% liked)

Linux

9071 readers
105 users here now

Welcome to c/linux!

Welcome to our thriving Linux community! Whether you're a seasoned Linux enthusiast or just starting your journey, we're excited to have you here. Explore, learn, and collaborate with like-minded individuals who share a passion for open-source software and the endless possibilities it offers. Together, let's dive into the world of Linux and embrace the power of freedom, customization, and innovation. Enjoy your stay and feel free to join the vibrant discussions that await you!

Rules:

  1. Stay on topic: Posts and discussions should be related to Linux, open source software, and related technologies.

  2. Be respectful: Treat fellow community members with respect and courtesy.

  3. Quality over quantity: Share informative and thought-provoking content.

  4. No spam or self-promotion: Avoid excessive self-promotion or spamming.

  5. No NSFW adult content

  6. Follow general lemmy guidelines.

founded 2 years ago
MODERATORS
top 28 comments
sorted by: hot top controversial new old
[–] TheGrandNagus@lemmy.world 46 points 2 days ago

Fully open and accessible: Fully open-source release of model weights, training hyperparameters, datasets, and code, fostering innovation and collaboration within the AI community.

That's actually pretty good. Seems to be open source as the OSI defines it, rather than the much more common "this model is open source, but the dataset is a secret".

[–] rando@sh.itjust.works 1 points 1 day ago

Got it up and running on a Debian distrobox… now I need to figure out how to train it. Will be my first steps into this type of thing – so prob will take me a bit to figure out how it all works

[–] sxan@midwest.social 5 points 2 days ago (2 children)

I need to catch up on training. I need an LLM that I can train on all my ebooks and digitized music, and can answer questions "what's that book where the girl goes to the thing and does that deed?"

[–] catloaf@lemm.ee 3 points 2 days ago (1 children)

Existing implementations can probably do that already.

[–] sxan@midwest.social 1 points 2 days ago (1 children)

I'm sure; I just don't know how. I need to set aside some time and educate myself.

Frankly, this generation of AI I find rather dull. It won't directly lead a AGI, although I'm sure it'll be a component, but I think that'll be another 10-20 years before the next breakthrough. I personally don't think it's as interesting as the symbolic, knowledge-based systems of the mid-80's; at least those were reasoning systems. LLMs look impressive to lay people (including myself - I understand the general concepts, but have no experience with the programming or training, so I'm just another lay user), but there's no reasoning or understanding behind it, and if what it produces is truthful or accurate, it's largely on accident. So I've had trouble getting excited about it.

[–] catloaf@lemm.ee 1 points 2 days ago (1 children)

I mean literally just go to chatgpt or whatever and ask it "what's that movie with Morgan Freeman playing God" and it'll give a few guesses. For common info, it's usually pretty good.

[–] sxan@midwest.social 2 points 2 days ago

I don't want ChatGPT to remember that I was searching for Cafe Flesh by description.

[–] just_another_person@lemmy.world 1 points 2 days ago (1 children)
[–] sxan@midwest.social 2 points 2 days ago

Well, it's either that or finding a Lemmy community for "what's that book" and posting a long description of how you remember the book.

I did this recently about a TV show I once saw an episode or two on broadcast TV in the early 90's. Only I completely misremembered it; I thought one of the main characters was a little girl when, in fact, it was a grizzled old woodworker. I mean, really, Brain? What the actual fuck? Luckily, from the context and time period, someone recognized it, but I no longer trust my own memory (if I ever really did).

"So, I read this book once - I think there were two or three in the series - and it was about some little furry girls who find a magic potion in a cave, but the potion is poison that can kill everyone in the world, and they had to take it to a special magic hole and throw it in. And they're being chased the entire time by ghosts on ghost horses. I think there are trees in it." - me, trying to describe the LOTR. Thanks, Brain; you're a big help.

[–] HappyFrog@lemmy.blahaj.zone 6 points 2 days ago (2 children)

I see all these graphs about how much better this LLM is than another, but do those graphs actually translate to real world usefulness?

[–] oldfart@lemm.ee 1 points 15 hours ago

I have yet to see a 3B model that's not dumb.

[–] just_another_person@lemmy.world 7 points 2 days ago (3 children)

I think more of the issue is what constitutes actual open source. This is actually open source, and it performs well. If you're familiar with the space, then it's a big deal.

[–] null_dot@lemmy.dbzer0.com 4 points 2 days ago

I'm not familiar with the space but realised this was a big deal.

I feel like I need to shower after interacting with any of the other LMs.

Something fully open source will hopefully be embraced by the community and be used for some interesting, useful, and value producing things instead of just attracting venture capital.

[–] HappyFrog@lemmy.blahaj.zone 3 points 2 days ago* (last edited 2 days ago)

I see, thank you.

Damn, they even chose a dataset with a open license.

[–] possiblylinux127@lemmy.zip 1 points 2 days ago* (last edited 2 days ago) (1 children)

Is it really or is it just a binary release like everything else?

Edit: It is actually Foss

[–] just_another_person@lemmy.world 3 points 2 days ago (1 children)

Everything is explained and linked in the project, so...

[–] possiblylinux127@lemmy.zip 2 points 2 days ago

Yeah I noticed that after writing this. Really cool stuff

[–] penquin@lemm.ee 2 points 2 days ago

Thank god for this. Setting up deepseek to utilize my AMD GPU through llama was near impossible

[–] brokenlcd@feddit.it 3 points 2 days ago (2 children)

The problem is... How do we run it if rocm is still a mess for most of their gpus? Cpu time?

[–] vividspecter@lemm.ee 1 points 2 days ago* (last edited 2 days ago)

Is it still a mess? I thought it was reasonably well supported on Linux with GPUs from the past few years.

Well it's not necessarily geared towards consumer devices. As mentioned in the writeup, it's not trained on consumer gear.

[–] GaMEChld@lemmy.world 2 points 2 days ago* (last edited 2 days ago) (1 children)

Smart people, I beg of thee, explain! What can it do?

Edit: looks to be another text based one, not image generation right?

[–] just_another_person@lemmy.world 4 points 2 days ago (1 children)

It's language only, hence, LM

[–] GaMEChld@lemmy.world 2 points 2 days ago (1 children)

To be fair, I didn't know if that language included programming language, and thus maybe still consider image based AI to be included in LLM. Is there a different designation for the type of AI that does image generation?