Lak Lakshmanan
banner
lakluster.bsky.social
Lak Lakshmanan
@lakluster.bsky.social
🌥️ Personal observations, not investment advice.
📚 O'Reilly: BigQuery, ML Design Patterns, Data Science 👨‍🏫Coursera 🌪️Ex: @googlecloud @NOAA
https://www.vlakshman.com/
Cameras used to be considered by painters as shortcuts and traitorous to the art ... They opened up new uses
July 6, 2025 at 4:05 PM
But the big tech firms that have been responsible for the bulk of recent layoffs are highly profitable. Not sure how much the tax treatment was a contributing factor. (as opposed to making space for AI data center spend, for example)
July 5, 2025 at 3:19 PM
Putting it behind a paywall is a choice that the author made. I write on medium and my articles are free and can even be read in and incognito window. Try this: medium.com/data-science...
How to Choose the Architecture for Your GenAI Application
A framework to select the simplest, fastest, cheapest architecture that will balance LLMs’ creativity and risk
medium.com
July 5, 2025 at 2:58 PM
Thanks to all the AI engineers who shared what was working and what wasn't. And to Glen Yu , David Cardozo, Andrew Stein, Sarah Grey for their prompt and helpful reviews.
May 31, 2025 at 1:59 AM
The patterns' code and brief descriptions are on GitHub:

github.com/lakshmanok/g...

Looking forward to hearing from you! What did you find helpful? Where can we do better? This is early access, so you still have a chance to make it better for future readers.
GitHub - lakshmanok/generative-ai-design-patterns: A catalog of design patterns when building generative AI applications
A catalog of design patterns when building generative AI applications - lakshmanok/generative-ai-design-patterns
github.com
May 31, 2025 at 1:59 AM
Looks like an O'Reilly book cover
April 24, 2025 at 2:52 PM
Cute story. The guy who's stuck tells the story tongue in cheek. Too many haters in the comments.
April 18, 2025 at 6:54 PM
The graph shows change from the same 30 days a year earlier. So it's YoY
April 6, 2025 at 10:07 PM
I was not able to replicate this, at least on Gemini or Claude
g.co/gemini/share...

claude.ai/share/00cb35...
April 3, 2025 at 2:30 AM
Very much what Kahneman called "system 1" and "system 2" in "Thinking Fast and Slow". System 1 is intuitive, system 2 is deliberate, and there are lots of ways that system 1 gets tripped up
February 23, 2025 at 1:06 PM
Reposted by Lak Lakshmanan
Hold on your outrage. They just show the map per the official edicts. Indian and Chinese users do not even see the same border.
February 11, 2025 at 3:22 AM
There are technical challenges to making self-play work for things other than games. This is legitimately advancing the state of the art

bsky.app/profile/lakl...
Additional context is that much of the cost of training foundational models like GPT-4, Claude Opus, and Gemini Pro is for the compute hardware (GPus/TPUs) and human raters/annotators. So DeepSeek managed to cut both these costs by getting pure Reinforcement Learning to work. To do that, (1/3)
January 24, 2025 at 5:10 PM
Reposted by Lak Lakshmanan
They open sourced their entire training regimen. Deepseek trained their model from scratch using the same technique that AlphaGo used ("self-play") and that requires both less compute and less labeled data. They are legit.
January 24, 2025 at 4:25 PM
DeepSeek is now ranked just below the leading, recently deployed versions of Gemini and ChatGPT on LMArena. So how? Secondly, distillation requires access to so many input-output pairs that OpenAI/Google/Anthropic would immediately recognize and shut then down. (3/3)
January 24, 2025 at 5:07 PM
They focused on easily verifiable problems. Which also makes sense given that the DeepSeek team is China's version of Jane Street. The people claiming the Chinese copied are thinking that they must have distilled their model from GPT-4 or Gemini but there are two problems with this theory: (2/3)
January 24, 2025 at 5:07 PM
Additional context is that much of the cost of training foundational models like GPT-4, Claude Opus, and Gemini Pro is for the compute hardware (GPus/TPUs) and human raters/annotators. So DeepSeek managed to cut both these costs by getting pure Reinforcement Learning to work. To do that, (1/3)
January 24, 2025 at 5:07 PM
They open sourced their entire training regimen. Deepseek trained their model from scratch using the same technique that AlphaGo used ("self-play") and that requires both less compute and less labeled data. They are legit.
January 24, 2025 at 4:25 PM