Lorenzo Pantolini
banner
lorenzopantolini.bsky.social
Lorenzo Pantolini
@lorenzopantolini.bsky.social
Postdoc at Biozentrum, University of Basel. Applying AI to structural biology, specializing in pLMs and remote homology detection.
I’m presenting this work at the EMBO Computational Structural Biology Workshop in Heidelberg #EMBOComp3D this week, and @workshopmlsb.bsky.social in Copenhagen over the weekend. Let’s connect!
December 1, 2025 at 10:47 AM
Huge thanks to my co-authors: @lauraengist.bsky.social, @ievapudz.bsky.social @martinsteinegger.bsky.social, @torstenschwede.bsky.social, especially to @ninjani.bsky.social! Couldn't have done this without the whole team, including the Swiss-Model development team and the rest of the Schwede group.
December 1, 2025 at 10:47 AM
Try it out yourself! github.com/PickyBinders/tea. A web-service for search is coming soon at alphabet.scicore.unibas.ch.
GitHub - PickyBinders/tea: The Embedded Alphabet (TEA)
The Embedded Alphabet (TEA). Contribute to PickyBinders/tea development by creating an account on GitHub.
github.com
December 1, 2025 at 10:37 AM
Ultimately, TEA brings deep learning representation to protein sequence bioinformatics algorithms, such as profiles, phylogenetic trees, motif finding, multiple sequence alignments, and more, all while maintaining the speed and low resource consumption of amino acid sequences. (6/n)
December 1, 2025 at 10:37 AM
We used TEA to connect >1.5 million singletons in AFDB Clusters, proteins which slipped past structure-based clustering approaches due to disordered or repetitive regions or simply because of low confidence structure predictions. (5/n)
December 1, 2025 at 10:36 AM
TEA sequences come with a built-in confidence metric in the form of Shannon entropy, which we saw correlates with pLDDT, and can be used to filter out uncertain predictions. (4/n)
December 1, 2025 at 10:36 AM
Running MMseqs2 with TEA gives extremely fast and highly sensitive results, similar to structural searches, even on unseen folds! Check out our ablations to see how we ended up with the final architecture. (3/n)
December 1, 2025 at 10:35 AM
By using a contrastive objective, we trained an alphabet enriched with structural information, without the need for the actual structure. This approach ensures that remote homologs expressed with TEA maintain high sequence identity. (2/n)
December 1, 2025 at 10:29 AM