R. David Edelman
banner
rdave.bsky.social
R. David Edelman
@rdave.bsky.social
Guidance counselor to toddling technologies; wrangler of adolescent algorithms; couples therapist for isolated isotopes. TwiX: @R_D
15/. We’re 2+ years into availability of GenAI. Maybe six months into broad deployment in industry. Surprises will happen. It just so happens that with AI, they matter immensely for the U.S., China, and the world. END.
January 27, 2025 at 2:38 PM
14/ So as a few stocks wipe a couple trillion (!) off of the markets, it’s a good time for investors and analysts to take a deep breath and remember: we’re at the top of the first inning, at best, with this tech. www.nature.com/articles/d41...
China’s cheap, open AI model DeepSeek thrills scientists
DeepSeek-R1 performs reasoning tasks at the same level as OpenAI’s o1 — and is open for researchers to examine.
www.nature.com
January 27, 2025 at 2:38 PM
13/ But we’ll see; new hardware matters; selection of performance benchmarks matter; usage trends for the AI-implementing industry matter; all of that. We’ll know more in a month.
January 27, 2025 at 2:38 PM
12/ Second, to me, @DeepSeek S1 does not provide much more evidence of the commoditization thesis than @Meta’s LLaMA does — *if you think open source will beat out commercial AI, THAT’S the macro thesis.* Not that a cheaper training is possible.
January 27, 2025 at 2:38 PM
11/ On AI/Chip markets — quickly as they open: first, the last Admin’s chip restrictions *may* not have conferred some desired "3yr advantage"; or even a 6mo one…but none at all at a market level. (i.e. still slowing the PLA; not AI companies.) www.wired.com/story/new-us...
New US Rule Aims to Block China’s Access to AI Chips and Models by Restricting the World
The US government has announced a radical plan to control exports of cutting-edge AI technology to most nations.
www.wired.com
January 27, 2025 at 2:38 PM
10/ (Watch this space; energy is the medium-term pinch-point of training cost and AI availability; this is the subject of the @brookings.edu piece.)
January 27, 2025 at 2:38 PM
9/ And if DeepSeek R1 has done that for a SOTA LLM, it’s a big deal. But that innovation is unlikely to stay proprietary long, and hardly un-replicable. U.S. companies would LOVE to have a SOTA model they can train for $5M! They’ll spend money to get it!
January 27, 2025 at 2:38 PM
8/ Efficiency innovations aren’t theirs alone. Remember when @MIT grad student @jefrankle cracked the code on the Lottery Ticket Hypothesis? AI is immature science; you can have a breakthrough and boom, drop training costs hugely without gutting performance. arxiv.org/abs/1803.03635
The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
Neural network pruning techniques can reduce the parameter counts of trained networks by over 90%, decreasing storage requirements and improving computational performance of inference without compromi...
arxiv.org
January 27, 2025 at 2:38 PM
7/ Several of us have been making the point that we haven’t seen U.S. AI development under massive constraint yet; that companies haven’t had e.g. chips as a principal limiting reagent to their innovations. The Chinese have. www.nature.com/articles/d41...
China’s cheap, open AI model DeepSeek thrills scientists
DeepSeek-R1 performs reasoning tasks at the same level as OpenAI’s o1 — and is open for researchers to examine.
www.nature.com
January 27, 2025 at 2:38 PM
6/ On INPUTS: A @brookings.edu piece coming out in a few days will have my full thoughts, but until then…worth noting that Chinese companies have been trying to do more with less; fewer chips, maybe less capital.
January 27, 2025 at 2:38 PM
5/ So two key takeaways from DeepSeek R1. First, a tech reality check: it’s early days, and different inputs — money, energy, social constraints — yield different outputs. Second, a market reality check: the chances that GenAI (even AGI?) gets commoditized is non-zero.
January 27, 2025 at 2:38 PM
4/ IF it turns S1 is both state-of-the-art-ish in performance AND tightly censored, that’s an important data point in the US/China AI race, and will shake some priors on both sides. See e.g. www.nytimes.com/2025/01/23/t...
How Chinese A.I. Start-Up DeepSeek Is Competing With Silicon Valley Giants
The company built a cheaper, competitive chatbot with fewer high-end computer chips than U.S. behemoths like Google and OpenAI, showing the limits of chip export control.
www.nytimes.com
January 27, 2025 at 2:38 PM
3/ Like all Chinese models available to the public, it’s a censorship machine! Some say that is irrelevant for the underlying tech; AI companies tell me otherwise, that nerfing a model to comply with the party has real performance impacts. alecmuffett.com/article/110996
x.com
x.com
January 27, 2025 at 2:38 PM
2/ Some experts are privately noting (in their words) like many Chinese models on the rankings, that it “teaches to the test” — and that we should look askance at claims about training costs. Maybe. BUT…
January 27, 2025 at 2:38 PM
1/ The DeepSeek Model isn’t new overnight; it’s been the “talk of Davos” late last week — but it seems the broader Internet (and investing-net) has woken up that there’s something interesting here. techcrunch.com/2025/01/20/d...
DeepSeek claims its 'reasoning' model beats OpenAI's o1 on certain benchmarks | TechCrunch
DeepSeek has released an open version of its 'reasoning' AI model, DeepSeek-R1, that it claims performs as well as OpenAI's o1 on certain benchmarks.
techcrunch.com
January 27, 2025 at 2:38 PM