pull down to refresh
153 sats \ 3 replies \ @elvismercury 25 Jan \ parent \ on: Deepseek represents extreme efficiency gain in AI training or is this a psyops? tech
We'll know a lot more when the replication efforts bear fruit. There's some amount of accounting sleight-of-hand, but most sources I follow find the methodology credible, and you can run the thing on modest amounts of consumer hardware.
I mostly meant leap frogging and 10x-ing will be common. I've found this deepseek story very interesting, but the parts that interest me are pretty sparse (ie ai sanctions futility, "cracked" STEM kids leaving wall street for ai, theoretical best case energy and compute requirements for AGI) so I'm going try to compress it into a scatterbrained post this weekend.
reply
Ah. Well, I look fwd to the post.
A thing I find awe-inspiring is the idea (which I endorse) that, if all progress on training new frontier models stopped, we'd be exploiting the results of what we've already got for another decade. Makes my brain hurt to really think about it.