Graphcore Research
gcresearchteam.bsky.social
Graphcore Research
@gcresearchteam.bsky.social
The 🦋 account of the Graphcore Research team.

Our mission is to contribute to the advancement of AI research and understand the computational requirements of intelligence.
Pinned
September’s Papers of the Month is here, and this month is all about LLMs! 🧠

This month, we cover:

➡️ FlowRL
➡️ Soft Tokens, Hard Truths
➡️ Set Block Decoding is a Language Model Inference Accelerator
➡️ Turning Recurring LLM Reasoning into Concise Behaviors

🧵
September’s Papers of the Month is here, and this month is all about LLMs! 🧠

This month, we cover:

➡️ FlowRL
➡️ Soft Tokens, Hard Truths
➡️ Set Block Decoding is a Language Model Inference Accelerator
➡️ Turning Recurring LLM Reasoning into Concise Behaviors

🧵
October 9, 2025 at 8:48 AM
Summer may be over, but Papers of the Month certainly isn’t!

For August’s edition, we covered the following papers:

➡️ ADMIRE-BayesOpt
➡️ Guiding Diffusion Models with RL for Stable Molecule Generation
➡️ Graph-R1



🧵
September 10, 2025 at 3:24 PM
July's Papers of the Month are here!

🧠 Subliminal Learning: Language Models Transmit Behavioral Traits via Hidden Signals in Data

💽 Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive Token-Level Computation

📊 DataRater: Meta-Learned Dataset Curation

🧵 ⬇️
August 6, 2025 at 10:40 AM
Your boss emails you a point in 128-billion-dimensional space. It's Llama 8B in bfloat16. They want it compressed.

What should you do 🤔... quantise to NF4? 🧵
June 12, 2025 at 11:19 AM
As we hurtle into the summer, it’s time for May’s Papers of the Month! This month, we cover Parallel Scaling Laws for Language Models, Alpha Evolve, Soft Thinking and Spurious Rewards! 🧵
June 4, 2025 at 1:21 PM
Our latest work uses theory from the '50s to figure out how to design weight quantisation formats for LLM inference.

It's called Optimal Formats for Weight Quantisation and has just hit arXiv.

1/6
May 22, 2025 at 12:25 PM
It's time for April's Papers of the Month! This month, we cover:
➡️ Motion Prompting: Controlling Video Generation with Motion Trajectories
➡️ Inference-Time Scaling for Generalist Reward Modeling
➡️ M1: Towards Scalable Test-Time Compute with Mamba Reasoning Models!
🧵
May 8, 2025 at 10:58 AM
Spring is here and so is Papers of the Month! In this March edition, we cover Transformers without Normalisation, Compute Optimal Scaling of Skills, Overtrained Language Models Are Harder to Fine-Tune, and Multi-Domain Distribution Learning for De Novo Drug Design! 🧵
April 7, 2025 at 8:26 AM
February might have been the shortest month, but it wasn’t short of papers! In this edition of Papers of the Month, we cover Distillation Scaling Laws, Matryoshka Quantisation, ParetoQ, and Scaling Test-Time Compute with Latent Reasoning! 🧵
March 7, 2025 at 1:21 PM
New year, new Papers of the Month! To kick off 2025, we cover:

Titans, Evolving Deeper LLM Thinking, Transformer-Squared and the recent DeepSeek technical reports! 🧵

graphcore-research.github.io/papers-of-th...
January Papers: More Like “Reas-anuary Papers”
New year, new Papers of the Month! Kicking off 2025, it’s apparent that reasoning and test-time compute are the hot topics on the block, with much research investigating how to best use these new meth...
graphcore-research.github.io
February 4, 2025 at 11:04 AM
Each month our team writes up summaries and analysis of our favourite ML papers. For December we cover:

The Byte Latent Transformer, Large Concept Models, Memory Layers & Phi-4 — all grouped under the title "Spend Your FLOPs Wisely". Here's our take (🧵)

graphcore-research.github.io/papers-of-th...
December Papers: Spend Your FLOPs Wisely
Welcome to Papers of the Month — Graphcore Research’s effort to bring you our pick of the most interesting ML papers. In December we noted a collection of papers which took innovative approaches to al...
graphcore-research.github.io
January 9, 2025 at 11:00 AM