Gated DeltaNet hybrids (Qwen3-Next, Kimi Linear), text diffusion, code world models, and small reasoning transformers.
🔗 magazine.sebastianraschka.com/p/beyond-sta...
Gated DeltaNet hybrids (Qwen3-Next, Kimi Linear), text diffusion, code world models, and small reasoning transformers.
🔗 magazine.sebastianraschka.com/p/beyond-sta...
Link to the full article: magazine.sebastianraschka.com/p/the-big-ll...
Link to the full article: magazine.sebastianraschka.com/p/the-big-ll...
This year has been brutal for tech workers:
- September: 19,300 people impacted
- October: 5,100 more
- Tomorrow: 30,000 from Amazon alone.
#amazon #layoffs #jobmarket
This year has been brutal for tech workers:
- September: 19,300 people impacted
- October: 5,100 more
- Tomorrow: 30,000 from Amazon alone.
#amazon #layoffs #jobmarket
They’re subscribing to the world’s most advanced 𝗺𝗼𝗱𝗲𝗹 𝗼𝗳 𝗰𝗼𝗻𝗱𝗶𝘁𝗶𝗼𝗻𝗮𝗹 𝗽𝗿𝗼𝗯𝗮𝗯𝗶𝗹𝗶𝘁𝘆 — one that has learned how words, ideas, and meanings relate to each other in ways that 𝗳𝗲𝗲𝗹 𝗵𝘂𝗺𝗮𝗻.
#ai #probability
They’re subscribing to the world’s most advanced 𝗺𝗼𝗱𝗲𝗹 𝗼𝗳 𝗰𝗼𝗻𝗱𝗶𝘁𝗶𝗼𝗻𝗮𝗹 𝗽𝗿𝗼𝗯𝗮𝗯𝗶𝗹𝗶𝘁𝘆 — one that has learned how words, ideas, and meanings relate to each other in ways that 𝗳𝗲𝗲𝗹 𝗵𝘂𝗺𝗮𝗻.
#ai #probability
🔗 github.com/rasbt/LLMs-f...
🔗 github.com/rasbt/LLMs-f...
🔗 github.com/rasbt/LLMs-f...
🔗 github.com/rasbt/LLMs-f...
System:
• 𝗙𝗔𝗜𝗦𝗦 for vector search with 𝘀𝗲𝗻𝘁𝗲𝗻𝗰𝗲-𝘁𝗿𝗮𝗻𝘀𝗳𝗼𝗿𝗺𝗲𝗿𝘀/𝗮𝗹𝗹-𝗠𝗶𝗻𝗶𝗟𝗠-𝗟𝟲-𝘃𝟮
• 𝗙𝗮𝘀𝘁𝗔𝗣𝗜 backend with async
• 𝗚𝗿𝗼𝗾 𝗔𝗣𝗜 (𝗟𝗹𝗮𝗺𝗮 𝟯.𝟯 𝟳𝟬𝗕) for ultra-fast inference
• 𝗘𝗹𝗲𝘃𝗲𝗻𝗟𝗮𝗯𝘀 𝗧𝗧𝗦 (with 𝗴𝗧𝗧𝗦 fallback) for realistic audio generation
GitHub: github.com/mcrao/Build-...
System:
• 𝗙𝗔𝗜𝗦𝗦 for vector search with 𝘀𝗲𝗻𝘁𝗲𝗻𝗰𝗲-𝘁𝗿𝗮𝗻𝘀𝗳𝗼𝗿𝗺𝗲𝗿𝘀/𝗮𝗹𝗹-𝗠𝗶𝗻𝗶𝗟𝗠-𝗟𝟲-𝘃𝟮
• 𝗙𝗮𝘀𝘁𝗔𝗣𝗜 backend with async
• 𝗚𝗿𝗼𝗾 𝗔𝗣𝗜 (𝗟𝗹𝗮𝗺𝗮 𝟯.𝟯 𝟳𝟬𝗕) for ultra-fast inference
• 𝗘𝗹𝗲𝘃𝗲𝗻𝗟𝗮𝗯𝘀 𝗧𝗧𝗦 (with 𝗴𝗧𝗧𝗦 fallback) for realistic audio generation
GitHub: github.com/mcrao/Build-...
sebastianraschka.com/blog/2021/dl...
sebastianraschka.com/blog/2021/dl...
Here are my thoughts:
In my experiment, I was dealing with embedding vectors less than 100,000+.
#pinecone #chroma #weaviate #qdrant #vectordb
To build production-ready RAG you need:
📄 Ingestion (PDF, OCR)
✂️ Chunking (Fixed, Semantic, Recursive, LLM-based)
🔎 Embeddings (Pinecone, Chroma, pgvector)
🧪 Eval (RAGAS)
🌐 Deploy (Supabase + pgvector + Lovable)
Demo 👉 diet-whisper.lovable.app
Code 👉 github.com/mcrao/RAG/tr...
Here are my thoughts:
In my experiment, I was dealing with embedding vectors less than 100,000+.
#pinecone #chroma #weaviate #qdrant #vectordb
To build production-ready RAG you need:
📄 Ingestion (PDF, OCR)
✂️ Chunking (Fixed, Semantic, Recursive, LLM-based)
🔎 Embeddings (Pinecone, Chroma, pgvector)
🧪 Eval (RAGAS)
🌐 Deploy (Supabase + pgvector + Lovable)
Demo 👉 diet-whisper.lovable.app
Code 👉 github.com/mcrao/RAG/tr...
To build production-ready RAG you need:
📄 Ingestion (PDF, OCR)
✂️ Chunking (Fixed, Semantic, Recursive, LLM-based)
🔎 Embeddings (Pinecone, Chroma, pgvector)
🧪 Eval (RAGAS)
🌐 Deploy (Supabase + pgvector + Lovable)
Demo 👉 diet-whisper.lovable.app
Code 👉 github.com/mcrao/RAG/tr...
Here's the result of my training runs:
• RQ-VAE to compress item embeddings into tokens
• SASRec to predict the next item (i.e., 4-tokens) exactly
• Qwen3-8B that can return recs and natural language!
eugeneyan.com/writing/sema...
Here's the result of my training runs:
• RQ-VAE to compress item embeddings into tokens
• SASRec to predict the next item (i.e., 4-tokens) exactly
• Qwen3-8B that can return recs and natural language!
eugeneyan.com/writing/sema...
🐧🦩 Who needs legs?!
simonwillison.net/2025/Sep/12/...
🐧🦩 Who needs legs?!
simonwillison.net/2025/Sep/12/...
👉 Try the demo with your audio or ours, share your feedback, and help us shape the future of decoding animal communication: huggingface.co/blog/EarthSp...
👉 Try the demo with your audio or ours, share your feedback, and help us shape the future of decoding animal communication: huggingface.co/blog/EarthSp...
• How it differs from basic Q&A
• What dimensions & metrics to eval on
• How to build llm-evaluators
• How to build eval datasets
• Benchmarks: narratives, technical docs, multi-docs
eugeneyan.com/writing/qa-e...
• How it differs from basic Q&A
• What dimensions & metrics to eval on
• How to build llm-evaluators
• How to build eval datasets
• Benchmarks: narratives, technical docs, multi-docs
eugeneyan.com/writing/qa-e...
Why? Because I think 1B & 3B models are great for experimentation, and I wanted to share a clean, readable implementation for learning and research: huggingface.co/rasbt/llama-...
Why? Because I think 1B & 3B models are great for experimentation, and I wanted to share a clean, readable implementation for learning and research: huggingface.co/rasbt/llama-...
If you are new to reinforcement learning, this article has a generous intro section (PPO, GRPO, etc)
Also, I cover 15 recent articles focused on RL & Reasoning.
🔗 magazine.sebastianraschka.com/p/the-state-...
If you are new to reinforcement learning, this article has a generous intro section (PPO, GRPO, etc)
Also, I cover 15 recent articles focused on RL & Reasoning.
🔗 magazine.sebastianraschka.com/p/the-state-...
Here, I
1. Discuss the advantages & disadvantages of reasoning models
2. Of course, describe and discuss DeepSeek R1
3. Describe the 4 main ways to building & improving reasoning models
Here, I
1. Discuss the advantages & disadvantages of reasoning models
2. Of course, describe and discuss DeepSeek R1
3. Describe the 4 main ways to building & improving reasoning models