Andrés Moreno
andresmorenob.bsky.social
Andrés Moreno
@andresmorenob.bsky.social
Systems Engineering and Computer Science Professor at Pontificia Universidad Javeriana in Bogotá, Colombia. Interested in data, AI,and education. Opinions expressed are my own and do not reflect the views of my employer.
Reposted by Andrés Moreno
🚨 New FREE self-paced course!

We are excited to launch the #EpiTrainingKit #Africa: Introduction to Infectious Disease Modelling for Public Health.

🎯 Tailored for the African context
🌍 With a gender perspective
🆓 Open-access and online

#EpiTKit #Epiverse #PublicHealth
August 4, 2025 at 6:30 PM
Reposted by Andrés Moreno
#EpiverseTRACE is now on Bluesky & LinkedIn! 🎉

We’re expanding to be more inclusive & diverse, reaching a wider audience in public health & data science.

Want to know more about what we do?⁉️🤔

🧵a thread!
March 11, 2025 at 10:29 AM
Reposted by Andrés Moreno
These thoughts after working on suspenseful story generation (pre o1) arxiv.org/abs/2402.17119. GPT can be beat at suspense generation. Interestingly GPT can be improved by guiding it with the theory of-mind part of story b planning.

~3 pages is the longest form we’ve tried.
Creating Suspenseful Stories: Iterative Planning with Large Language Models
Automated story generation has been one of the long-standing challenges in NLP. Among all dimensions of stories, suspense is very common in human-written stories but relatively under-explored in AI-ge...
arxiv.org
December 30, 2024 at 1:21 AM
Reposted by Andrés Moreno
Something I don't understand is: why can't LLMs write novel-length fiction yet?

They've got the context length for it. And new models seem capable of the multi-hop reasoning required for plot. So why hasn't anyone demoed a model that can write long interesting stories?

I do have a theory ... +
December 30, 2024 at 12:17 AM
Reposted by Andrés Moreno
Great blog post (by a 15-author team!) on their release of ModernBERT, the continuing relevance of encoder-only models, and how they relate to, say, GPT-4/llama. Accessible enough that I might use this as an undergrad reading.
Finally, a Replacement for BERT: Introducing ModernBERT
We’re on a journey to advance and democratize artificial intelligence through open source and open science.
huggingface.co
December 19, 2024 at 7:11 PM
Reposted by Andrés Moreno
With students writing my theory exam today, I figured it's a good time to share a link to my open textbook with all you current (and future!) theoreticians.
This term was the first time I used it in class, and students loved it. Big plans for future editions, so stay tuned!
taylorjsmith.xyz/tocopen/
December 16, 2024 at 5:35 PM
Great tutorial on language models!
December 11, 2024 at 8:04 AM
Reposted by Andrés Moreno
Check out this BEAUTIFUL interactive blog about cameras and lenses

ciechanow.ski/cameras-and-...
November 27, 2024 at 2:54 AM
Reposted by Andrés Moreno
A timely paper exploring ways academics can pretrain larger models than they think, e.g. by trading time against GPU count.

Since the title is misleading, let me also say: US academics do not need $100k for this. They used 2,000 GPU hours in this paper; NSF will give you that. #MLSky
$100K or 100 Days: Trade-offs when Pre-Training with Academic Resources
Pre-training is notoriously compute-intensive and academic researchers are notoriously under-resourced. It is, therefore, commonly assumed that academics can't pre-train models. In this paper, we seek...
arxiv.org
November 23, 2024 at 1:50 PM
Reposted by Andrés Moreno
A poem for my last day working at the writing center for the semester (by Joseph Fasano)
November 23, 2024 at 2:09 AM