Projects: dylancastillo.co/projects
The only thing I’m sure about is that I’m rarely able to ask something without getting an overengineered solution and a random new README in my codebase.
The only thing I’m sure about is that I’m rarely able to ask something without getting an overengineered solution and a random new README in my codebase.
Once a project is big enough, they definitely make you more productive vs vanilla, htmx, etc
But I'm happy that I didn't switch earlier, writing frontend code without AI tools must be horrible.
Once a project is big enough, they definitely make you more productive vs vanilla, htmx, etc
But I'm happy that I didn't switch earlier, writing frontend code without AI tools must be horrible.
It's a quick test designed to assess your estimation skills: estimator.dylancastillo.co/
This is inspired by @codinghorror's great posts: blog.codinghorror.com/how-good-an...
archive.is/qDc0v
It's a quick test designed to assess your estimation skills: estimator.dylancastillo.co/
This is inspired by @codinghorror's great posts: blog.codinghorror.com/how-good-an...
archive.is/qDc0v
Last year:
💵 I worked on 9 projects with 7 clients. Doubled revenue, costs are up by 155%.
💻 Coded 322 days. Wrote 14 blog posts.
🧠 Struggled with focus. Nearly burned out.
📸 Debi tirar mas fotos.
dylancastillo.co/posts/2024-...
Last year:
💵 I worked on 9 projects with 7 clients. Doubled revenue, costs are up by 155%.
💻 Coded 322 days. Wrote 14 blog posts.
🧠 Struggled with focus. Nearly burned out.
📸 Debi tirar mas fotos.
dylancastillo.co/posts/2024-...
After a bit of digging, I realized that it was just due to people misspelling "DeepSeek."
There are now people out there who think that China's top AI is a 💩 that makes charts.
After a bit of digging, I realized that it was just due to people misspelling "DeepSeek."
There are now people out there who think that China's top AI is a 💩 that makes charts.
Sounds easy, but happens to everyone.
Here's OpenAI breaking the CoT reasoning of an LLM judge.
Sounds easy, but happens to everyone.
Here's OpenAI breaking the CoT reasoning of an LLM judge.
1. Using constrained decoding seems to lower performance in reasoning tasks.
2. The Generative AI SDK can break your model's reasoning.
Just re-ran Let Me Speak Freely benchmarks with Gemini and got some interesting news
1. Using constrained decoding seems to lower performance in reasoning tasks.
2. The Generative AI SDK can break your model's reasoning.
Just re-ran Let Me Speak Freely benchmarks with Gemini and got some interesting news
I replicated @willkurt.bsky.social / @dottxtai.bsky.social rebuttal of Let Me Speak Freely? (LMSF) using gpt-4o-mini
The rebuttal correctly highlights many flaws with the original study, but ironically, LMSF's conclusion still holds
I replicated @willkurt.bsky.social / @dottxtai.bsky.social rebuttal of Let Me Speak Freely? (LMSF) using gpt-4o-mini
The rebuttal correctly highlights many flaws with the original study, but ironically, LMSF's conclusion still holds
h/t @alpindale.bsky.social and @danielvanstrien.bsky.social for helping with that goal
h/t @alpindale.bsky.social and @danielvanstrien.bsky.social for helping with that goal
1. Silently fixes their spelling mistakes on their iPhone
2. Calculate the fastest way home on Uber
3. Pick the right music for the ride on Spotify
4. Keeps their credit card safe to buy their subscriptions
The same 99% will happen here too, but if AI researchers continue to get perma-banned for making available the datasets needed to filter it, it’s going to make this platform unusable.
1. Silently fixes their spelling mistakes on their iPhone
2. Calculate the fastest way home on Uber
3. Pick the right music for the ride on Spotify
4. Keeps their credit card safe to buy their subscriptions
You can be skeptical about its real-life utility (which I am), but it's hard to bet against such powerful motivators.
You can be skeptical about its real-life utility (which I am), but it's hard to bet against such powerful motivators.