hardmaru
hardmaru.bsky.social
hardmaru
@hardmaru.bsky.social
I work at Sakana AI 🐟🐠🐡 → @sakanaai.bsky.social

https://sakana.ai/careers
Excited to announce Sakana AI’s Series B! 🐟
sakana.ai/series-b

From day one, Sakana AI has done things differently. Our research has always focused on developing efficient AI technology sustainably, driven by the belief that resource constraints—not limitless compute—are key to true innovation.
November 17, 2025 at 12:03 AM
Proud to release ShinkaEvolve, our open-source framework that evolves programs for scientific discovery with very good sample-efficiency! 🐙🧠

Paper: arxiv.org/abs/2509.19349
Blog: sakana.ai/shinka-evolve/
GitHub Project: github.com/SakanaAI/Shi...
September 25, 2025 at 6:01 AM
Just received my copy of “What Is Intelligence?” by @blaiseaguera.bsky.social 🧠🪱

Thanks for sending it to Japan! 🗼

whatisintelligence.antikythera.org
September 11, 2025 at 5:49 AM
Why Greatness Cannot Be Planned

Both the English and Japanese editions now found a home in the Sakana AI library ✨ @sakanaai.bsky.social
August 26, 2025 at 8:58 AM
In “The Vertigo Years: Europe 1900-1914”, Blom describes how turn-of-the-century technology changed the way people thought about art and human nature and how it contributed to a nervous breakdown across the west.
August 15, 2025 at 8:01 AM
Andrew Ng’s piece on 🇺🇸 vs 🇨🇳 competition in AI worth reading:

Full article: www.deeplearning.ai/the-batch/is...
August 1, 2025 at 1:42 AM
ICML’s Statement about subversive hidden LLM prompts

We live in a weird timeline…

icml.cc/Conferences/...
July 23, 2025 at 12:50 PM
Not Even Bronze: Evaluating LLMs on 2025 International Math Olympiad 🥉

matharena.ai/imo/

Nice blog post from the team behind MathArena: Evaluating LLMs on Uncontaminated Math Competitions (arxiv.org/abs/2505.23281) providing independent analysis, debunking some claims about LLM performance on IMO.
July 20, 2025 at 2:35 PM
Hikaru Utada and Yuval Noah Harari talk about The Evolution of AI and Creativity

youtu.be/xw-9mwZxl-0

Gotta admit, this was not on my bingo card!
July 15, 2025 at 4:04 AM
Google’s Gemini 2.5 paper has 3295 authors

arxiv.org/abs/2507.06261
July 13, 2025 at 1:21 PM
The AB-MCTS (Adaptive Branching Monte Carlo Tree Search) algorithm we are introducing is a step towards this. AB-MCTS is a method for inference-time scaling that enables frontier AIs from cheap, low-cost providers like OpenAI, Google, DeepSeek to cooperate and efficiently perform trial-and-error.
July 1, 2025 at 2:26 AM
Just as humanity’s greatest achievements arise from the collaboration of diverse minds, we believe the same principle applies to AI. Like a team of human experts tackling difficult problems, AIs should also collaborate by bringing their unique strengths to the table.
July 1, 2025 at 2:25 AM
This is because no matter how advanced each model becomes, each model retains its own individuality from its unique training data and methods. We see these biases and variations not as limitations, but as precious resources for creating collective intelligence, further enhancing their performance.
July 1, 2025 at 2:24 AM
Frontier AI models like ChatGPT, Gemini, Grok, DeepSeek, are evolving at a breathtaking pace. While each model requires enormous resources to develop, fierce market competition has made them become low-cost commodities, effectively ensuring that there will never be a single “winner” in AI.
July 1, 2025 at 2:23 AM
Inference-Time Scaling and Collective Intelligence for Frontier AI

sakana.ai/ab-mcts/

We developed AB-MCTS, a new inference-time scaling algorithm that enables multiple frontier AI models to cooperate, achieving promising initial results on the ARC-AGI-2 benchmark.
July 1, 2025 at 2:21 AM
Haha, nice.
June 27, 2025 at 8:25 AM
It feels like the agent layer is almost as important as the foundation model layer for hard tasks.
June 18, 2025 at 12:55 AM
Nvidia’s Jensen Huang says he disagrees with almost everything Anthropic CEO Dario Amodei says

fortune.com/2025/06/11/n...

I agree with Jensen. If you want AI development to be done safely and responsibly, you do it in the open. Don’t do it in a dark room and tell me it’s “safe”.
June 15, 2025 at 2:23 AM
Working with Hokkoku Bank, Sakana AI aims to help transform the regional banking industry in Japan, to serve as a model case for other regional banks in the future.
June 10, 2025 at 9:13 AM
Our team traveled to Ishikawa Prefecture in Japan today, where we had the honor meeting with Shuji Tsuemura, the President of Hokkoku Bank, to announce our partnership deal with the Bank where Sakana AI will provide bank-specific AI solutions to the regional bank.
June 10, 2025 at 9:13 AM
It turns out that Logistic Regression is still a very strong baseline for detecting fraudulent Japanese financial statements, matching frontier models like Claude3.7, R1, o4-mini. Much more room for future improvement!

GitHub: github.com/SakanaAI/EDI...
HuggingFace: huggingface.co/datasets/Sak...
June 9, 2025 at 2:04 AM
I like the comparison chart between AlphaEvolve and the Darwin Gödel Machine, and the analogy of the two approaches with two different kinds of chefs 🍽️
June 4, 2025 at 10:03 AM
AI that can improve itself: A deep dive into self-improving AI and the Darwin-Gödel Machine.

richardcsuwandi.github.io/blog/2025/dgm/

Excellent blog post by Richard Suwandi reviewing the Darwin Gödel Machine (DGM) and future implications.
June 4, 2025 at 10:03 AM
Default Alive or Default Dead

“If the company is default alive, we can talk about ambitious new things they could do. If it’s default dead, we probably need to talk about how to save it.”

I keep coming back to this 2015 Paul Graham essay about the importance for (AI) startups to be “Default Alive”
June 3, 2025 at 4:10 AM
“The growth trajectory is continuing, especially on the enterprise side. For a top Japanese bank like MUFG, they want to work with the top company that can move the needle, develop new things, not just copying or reimplementing existing ideas.”

No paywall link to Bloomberg article: archive.is/2qJ4M
May 19, 2025 at 3:10 PM