https://youtube.com/@deeplearningexplained?si=yzsA4kpGN_8VDEY0
1. Secret is pasted straight into file, no environment management.
2. The code is for a deprecated API.
What a vibe.
1. Secret is pasted straight into file, no environment management.
2. The code is for a deprecated API.
What a vibe.
It’s like a flash of every paper, formula and code seen that just come flooding all at once in its correct form.
It’s like a flash of every paper, formula and code seen that just come flooding all at once in its correct form.
Linear attention has existed since 2020, yet large-scale models rarely use it. Why?
minimax-01 finally makes linear attention work at scale. Deep dive here: 📌 youtu.be/iRuvGU-Sk3c
Linear attention has existed since 2020, yet large-scale models rarely use it. Why?
minimax-01 finally makes linear attention work at scale. Deep dive here: 📌 youtu.be/iRuvGU-Sk3c
Sorry for the month break, was a bit overwhelmed with lots of things at work.
I'll try to move around the schedule a bit so that more people in different time zones can attend.
📸 PS: I gave a talk at a conference in February!
Sorry for the month break, was a bit overwhelmed with lots of things at work.
I'll try to move around the schedule a bit so that more people in different time zones can attend.
📸 PS: I gave a talk at a conference in February!
What's an agent, what's a workflow, what's an agentic system, etc.
I made this tutorial on the topic packed with information from the latest research from HuggingFace.
Check it out over here:
youtu.be/UMYKjT9exb4
Enjoy! 🌹
What's an agent, what's a workflow, what's an agentic system, etc.
I made this tutorial on the topic packed with information from the latest research from HuggingFace.
Check it out over here:
youtu.be/UMYKjT9exb4
Enjoy! 🌹
Then we patch the whole thing left and right to keep the illusion of general intelligence with massive injection of capital.
Literally yeeting the AI field in a local minima and digging.
Then we patch the whole thing left and right to keep the illusion of general intelligence with massive injection of capital.
Literally yeeting the AI field in a local minima and digging.
Crazy how we went from the darling of AI to a company researchers loathe.
not a good vibe.
Crazy how we went from the darling of AI to a company researchers loathe.
not a good vibe.
They are hinting big breakthrough, but man look at the landscape.
Every competitors around is stacked with billions and PhD.
Whatever they are trying to win, won’t be achieved by secrecy.
They are hinting big breakthrough, but man look at the landscape.
Every competitors around is stacked with billions and PhD.
Whatever they are trying to win, won’t be achieved by secrecy.
1. find a log that i initially planned to turn into a table leg
2. make it into a puppet that can walk and talk
3. have the puppet, through a series of adventures, turn into a real boy and realize the true value of friendship
1. find a log that i initially planned to turn into a table leg
2. make it into a puppet that can walk and talk
3. have the puppet, through a series of adventures, turn into a real boy and realize the true value of friendship
The argument feels pretty reasonable. Here is my take: (1/6)
#MLSky #NeuroAI 🧠📈
The argument feels pretty reasonable. Here is my take: (1/6)
#MLSky #NeuroAI 🧠📈
arxiv.org/abs/2412.20292
Our closed-form theory needs no training, is mechanistically interpretable & accurately predicts diffusion model outputs with high median r^2~0.9
arxiv.org/abs/2412.20292
Our closed-form theory needs no training, is mechanistically interpretable & accurately predicts diffusion model outputs with high median r^2~0.9
Table of contents:
Table of contents:
We’re going to see more and more heavily gated web communities.
Source: news.ycombinator.com/item?id=4254...
We’re going to see more and more heavily gated web communities.