Nighttime: hacking on generative ai
Host of How AI Is Built
Unless we have a massive political change that money will just go to waste.
For innovation to happen, we don't need money first, but deregulation.
We cannot work on breakthrough technologies, when we are at constant fear of being sued.
Unless we have a massive political change that money will just go to waste.
For innovation to happen, we don't need money first, but deregulation.
We cannot work on breakthrough technologies, when we are at constant fear of being sued.
LLMs suck at long context.
This paper shows what I have seen in most deployments.
With longer contexts, performance degrades.
LLMs suck at long context.
This paper shows what I have seen in most deployments.
With longer contexts, performance degrades.
Throw everything in a vector database and hope something good comes out.
Throw all ranking signals into one big ML model and hope it makes something good out of it.
You don’t want to create this witch’s cauldron.
New episode on @howaiisbuilt.fm
Throw everything in a vector database and hope something good comes out.
Throw all ranking signals into one big ML model and hope it makes something good out of it.
You don’t want to create this witch’s cauldron.
New episode on @howaiisbuilt.fm
The reality is that it's easy to build, it's easy to get up and running, but it's really hard to get right.
And if you don't have a good setup, it's near impossible to debug.
One of the reasons it's really hard is chunking.
The reality is that it's easy to build, it's easy to get up and running, but it's really hard to get right.
And if you don't have a good setup, it's near impossible to debug.
One of the reasons it's really hard is chunking.
Can you show the Amazon people how to use a VLM to do the handwriting recognition.
That's atrocious.
Can you show the Amazon people how to use a VLM to do the handwriting recognition.
That's atrocious.
This lets you build fast, cheap models that do one thing well or that are more capable because they have (nearly) identical capabilities distilled into a smaller number of parameters.
This lets you build fast, cheap models that do one thing well or that are more capable because they have (nearly) identical capabilities distilled into a smaller number of parameters.
Meta has created BLT, a new AI model that works with raw bytes instead of tokens.
Current AI models split text into tokens (fixed chunks of letters) before processing it.
>>
Meta has created BLT, a new AI model that works with raw bytes instead of tokens.
Current AI models split text into tokens (fixed chunks of letters) before processing it.
>>
Different questions need different approaches.
➡️ 𝗤𝘂𝗲𝗿𝘆-𝗕𝗮𝘀𝗲𝗱 𝗙𝗹𝗲𝘅𝗶𝗯𝗶𝗹𝗶𝘁𝘆:
- Structured data? Use SQL
- Context-rich query? Use vector search
- Date-specific? Apply filters first
Different questions need different approaches.
➡️ 𝗤𝘂𝗲𝗿𝘆-𝗕𝗮𝘀𝗲𝗱 𝗙𝗹𝗲𝘅𝗶𝗯𝗶𝗹𝗶𝘁𝘆:
- Structured data? Use SQL
- Context-rich query? Use vector search
- Date-specific? Apply filters first
I always tend to forget the last step (the check for hitting the length of the dataset) on the first implementation.
I always tend to forget the last step (the check for hitting the length of the dataset) on the first implementation.
Coding a project forces you to apply concepts directly. It’s a richer learning experience than just reading technical articles.
You discover gaps, solve real problems, and solidify your understanding.
Coding a project forces you to apply concepts directly. It’s a richer learning experience than just reading technical articles.
You discover gaps, solve real problems, and solidify your understanding.
Data drift happens when the real world changes but your model doesn't.
- Input drift: The data coming in changes (like cameras getting better resolution)
- Label drift: What you're predicting changes (like what counts as "spam" evolving)
>>
Data drift happens when the real world changes but your model doesn't.
- Input drift: The data coming in changes (like cameras getting better resolution)
- Label drift: What you're predicting changes (like what counts as "spam" evolving)
>>