this post was submitted on 29 Jan 2025
1268 points (99.2% liked)
Microblog Memes
6328 readers
4325 users here now
A place to share screenshots of Microblog posts, whether from Mastodon, tumblr, ~~Twitter~~ X, KBin, Threads or elsewhere.
Created as an evolution of White People Twitter and other tweet-capture subreddits.
Rules:
- Please put at least one word relevant to the post in the post title.
- Be nice.
- No advertising, brand promotion or guerilla marketing.
- Posters are encouraged to link to the toot or tweet etc in the description of posts.
Related communities:
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Yeah it went from hey the monopoly justifies the cost. To Oh shit they did it for how much? Real fast.
I suspect china is fudging the training timeline tho...
I'm more prone to believe OpenAI is just a clunky POS. DeepSeek released a model that's operating on theories kicking around the LLM community for years. Now Alibaba is claiming they've got a better model, too.
Altman insisting he needed $1T in new physical infrastructure to get to the next iteration of his product should have been a red flag for everyone.
They're trying to brute force a solution to a problem that more elegate coding accomplishes better.
I had suspicious before, but I knew they were screwed when Qwen 2.5 came out. 32Bs and 72Bs nipping at their heels… O3 was a joke in comparison.
And they probably aren’t fudging anything. Base Deepseek isn’t like crazy or anything, and the way they finetuned it to R1 is public. Researchers are trying to replicate it now.
Also, the thing the Chinese govt did probably do is give Deepseek training data.
For all the memes about the NSA, the US govt isn’t really in that position, as whatever the US govt has pales in comparison to Microsoft or Google.