You can also find me at threads: @sung.kim.mw
Is that true?
Is that true?
A minimal, secure Python interpreter written in Rust for use by AI.
github.com/pydantic/monty
A minimal, secure Python interpreter written in Rust for use by AI.
github.com/pydantic/monty
LiteBox is a sandboxing library OS. Example use cases include:
- Running unmodified Linux programs on Windows
- Sandboxing Linux applications on Linux
github.com/microsoft/li...
LiteBox is a sandboxing library OS. Example use cases include:
- Running unmodified Linux programs on Windows
- Sandboxing Linux applications on Linux
github.com/microsoft/li...
Somehow, decades of watching anime have done absolutely nothing for my Japanese language skills.
Somehow, decades of watching anime have done absolutely nothing for my Japanese language skills.
Instead of "claude --dangerously-skip-permissions", you can do same thing in Copilot CLI by "copilot --yolo".
Instead of "claude --dangerously-skip-permissions", you can do same thing in Copilot CLI by "copilot --yolo".
They observe that reinforcement learning does not maximize this likelihood, and instead optimizes only a lower-order approximation. Inspired by this observation, they introduce Maximum Likelihood Reinforcement Learning (MaxRL),
They observe that reinforcement learning does not maximize this likelihood, and instead optimizes only a lower-order approximation. Inspired by this observation, they introduce Maximum Likelihood Reinforcement Learning (MaxRL),
This essay describes how he think we can do backprop, why interp is key, the relevance to alignment, and how we should do it right.
Blog: www.goodfire.ai/blog/intenti...
This essay describes how he think we can do backprop, why interp is key, the relevance to alignment, and how we should do it right.
Blog: www.goodfire.ai/blog/intenti...
OpenScholar is an open-source model for synthesizing scientific research—with citations as accurate as human experts. 🧵
OpenScholar is an open-source model for synthesizing scientific research—with citations as accurate as human experts. 🧵
They propose Divergence Proximal Policy Optimization (DPPO) with the benefits, such as:
- super stable and smooth training
They propose Divergence Proximal Policy Optimization (DPPO) with the benefits, such as:
- super stable and smooth training
On a serious note, Bitcoin hits highs then cycles into crypto winter. It is expected to go down to $30K or so. It should stay there for a while. Accumulate Bitcoin then for a new high.
On a serious note, Bitcoin hits highs then cycles into crypto winter. It is expected to go down to $30K or so. It should stay there for a while. Accumulate Bitcoin then for a new high.
techcrunch.com/2026/02/05/f...
techcrunch.com/2026/02/05/f...
You’d think they know exactly what the other is doing at any given time.
You’d think they know exactly what the other is doing at any given time.
"We pair OpenAI Forward Deployed Engineers (FDEs) with your teams, working side by side to help you develop the best practices to build and run agents in production."
openai.com/index/introd...
"We pair OpenAI Forward Deployed Engineers (FDEs) with your teams, working side by side to help you develop the best practices to build and run agents in production."
openai.com/index/introd...
voyage-4-nano is ideal for local development and prototyping while providing an easy path to production.
- Shared Embedding Space: voyage-4-nano shares an embedding space as its larger siblings
voyage-4-nano is ideal for local development and prototyping while providing an easy path to production.
- Shared Embedding Space: voyage-4-nano shares an embedding space as its larger siblings
“AI will be everywhere” → sell SaaS stocks
“The AI bubble is imploding” → sell AI hardware stocks
Both can’t be true at the same time, yet U.S. hyperscalers are guiding toward $700B in CapEx for 2026.
“AI will be everywhere” → sell SaaS stocks
“The AI bubble is imploding” → sell AI hardware stocks
Both can’t be true at the same time, yet U.S. hyperscalers are guiding toward $700B in CapEx for 2026.
"Learning to Reason in 13 Parameters"
Paper: arxiv.org/abs/2602.04118
"Learning to Reason in 13 Parameters"
Paper: arxiv.org/abs/2602.04118