I’m combining biophysical modeling with machine learning to study genome folding.
check the preprint: 🔗 www.biorxiv.org/content/10.1...
check the preprint: 🔗 www.biorxiv.org/content/10.1...
By reformulating loop extrusion as a differentiable process, we get:
✅ Interpretability
✅ Scalability
✅ Perturbation prediction
✅ Integration with ML
10/10
By reformulating loop extrusion as a differentiable process, we get:
✅ Interpretability
✅ Scalability
✅ Perturbation prediction
✅ Integration with ML
10/10
- 280× fewer than C.Origami
- 750× fewer than Orca
Yet we achieve competitive accuracy!
Why? Because we hard-coded the biophysics instead of making the model learn it.
9/10
- 280× fewer than C.Origami
- 750× fewer than Orca
Yet we achieve competitive accuracy!
Why? Because we hard-coded the biophysics instead of making the model learn it.
9/10
We embedded dLEM as a biophysical layer in a neural network that learns from:
- DNA sequence
- Chromatin accessibility (DNase/ATAC)
Result: Sequence → structure prediction that stays mechanistically grounded
8/10
We embedded dLEM as a biophysical layer in a neural network that learns from:
- DNA sequence
- Chromatin accessibility (DNase/ATAC)
Result: Sequence → structure prediction that stays mechanistically grounded
8/10
→ We systematically reduced CTCF coefficients in our linear model
→ Optimal prediction at α=0.33, matching the 33% of CTCF peaks that remain after auxin treatment
The model reveals residual CTCF is still functionally active!
7/10
→ We systematically reduced CTCF coefficients in our linear model
→ Optimal prediction at α=0.33, matching the 33% of CTCF peaks that remain after auxin treatment
The model reveals residual CTCF is still functionally active!
7/10
Example: WAPL depletion
→ We modified just the detachment rate
→ Successfully predicted emergence of new loops & extended stripes
Example: WAPL depletion
→ We modified just the detachment rate
→ Successfully predicted emergence of new loops & extended stripes
😌 CTCF shows expected directional asymmetry (blocks L or R depending on orientation)
😲 BUT: Transcription machinery (H3K4me3, H3K36me3, H3K4me1,...) ALSO modulates cohesin dynamics!
CTCF isn't the whole story 👀
5/10
😌 CTCF shows expected directional asymmetry (blocks L or R depending on orientation)
😲 BUT: Transcription machinery (H3K4me3, H3K36me3, H3K4me1,...) ALSO modulates cohesin dynamics!
CTCF isn't the whole story 👀
5/10
📥 ENCODER: Compress complex 2D Hi-C maps into simple 1D velocity profiles
📤 DECODER: Reconstruct contact maps from these profiles
It's dimensionality reduction, but with built-in biophysics!
4/10
📥 ENCODER: Compress complex 2D Hi-C maps into simple 1D velocity profiles
📤 DECODER: Reconstruct contact maps from these profiles
It's dimensionality reduction, but with built-in biophysics!
4/10
dLEM captures this with two velocity profiles:
→ L: leftward cohesin speed
→ R: rightward cohesin speed
Obstacles (like CTCF) = dips in velocity 📉
3/10
dLEM captures this with two velocity profiles:
→ L: leftward cohesin speed
→ R: rightward cohesin speed
Obstacles (like CTCF) = dips in velocity 📉
3/10
❌ Polymer simulations = mechanistic, but not scalable
❌ Deep learning = predictive, but black-box
❌ Summary metrics = compressed, but mechanism-agnostic
We needed something that's mechanistic, interpretable, AND scalable.
✅ Enter dLEM!
2/10
❌ Polymer simulations = mechanistic, but not scalable
❌ Deep learning = predictive, but black-box
❌ Summary metrics = compressed, but mechanism-agnostic
We needed something that's mechanistic, interpretable, AND scalable.
✅ Enter dLEM!
2/10