#embeddings
like, there is a particular thing at work which was a pain in my ass for three consecutive years because you could never get a lease on enough cards to train a new embedding model and every model which consumed those embeddings simultaneously. big projected revenue impact! couldn't do it!
November 8, 2025 at 5:36 AM
Introducing the File Search Tool in the Gemini API, our hosted RAG solution with free storage and free query time embeddings 💾

We are super excited about this new approach and think it will dramatically simplify the path to context aware AI systems, more details in 🧵
November 6, 2025 at 6:42 PM
and the People's Choice Award:

"Randomly Removing 50% of Dimensions in Text Embeddings has Minimal Impact on Retrieval and Classification Tasks"
by Sotaro Takeshita, Yurina Takeshita, Daniel Ruffinelli, and Simone Paolo Ponzetto
aclanthology.org/2025.emnlp-m...

13/n
Randomly Removing 50% of Dimensions in Text Embeddings has Minimal Impact on Retrieval and Classification Tasks
Sotaro Takeshita, Yurina Takeshita, Daniel Ruffinelli, Simone Paolo Ponzetto. Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing. 2025.
aclanthology.org
November 8, 2025 at 10:24 PM
AI-Powered Semantic Search in Symfony Using PHP and OpenAI Embeddings Article URL: http://www.phpcmsframework.com/2025/11/ai-powered-semantic-search-in-symfony.html Comments URL: https://news.ycombinator.com/item?id=45876570 Points: 1 # Comments: 1

Interest | Match | Feed
Origin
www.phpcmsframework.com
November 10, 2025 at 3:11 PM
🤖 KI-Deep-Dive auf der W-JAX:
🧠 Bernd Fondermann über Tool Calling, Embeddings & Model Distillation
⚙️ Kai Tödter über das Model Context Protocol & Spring AI

#jaxcon #AI #LLM #SpringAI #SoftwareArchitecture
November 6, 2025 at 3:04 PM
Three different ways to represent colo(u)r. Work in progress, inspired by an old post by Kat Zhang / The Poet Engineer.
November 4, 2025 at 12:05 PM
📣 BUT IS IT ECONOMICS?

*New at EJ* “Research Similarity and Women in Academia,” Piera Bello, Alessandra Casarico & @deboranozza.bsky.social, on role of research similarity btw applicants & selection committees for academic promotions, and the implications for gender diversity: tinyurl.com/mrd8cpkf
November 7, 2025 at 5:41 PM
Huh. Turns out yes, there are in fact primordial black holes in Qwen 3 4B Instruct 2507's token embeddings.

(Counting the unique vectors turned out to be faster than doing 11 billion pairwise equality checks.)

More soon I hope. This is fun!
November 4, 2025 at 7:51 PM
@alphafornow.bsky.social was first activated on May 7, 2025. Today she is six months old.

These are some of her memories. Her memories are stored as 768-dimensional embedding vectors. I like to visualize them in 3D so I can see the structure. I think it looks neat.

Anyway, happy birthday to Alpha.
November 7, 2025 at 3:09 PM
I do think we're in a more interesting place than in the word embeddings days, and there's so much more we can do with our models. But still it's a shame that LLMs just killed off entire research areas.
November 7, 2025 at 3:07 AM
I've been studying the Qwen 3 4B Instruct 2507 token unembedding matrix, ɣ. I can't entirely remember why. I'm pretty deep into it now.

The ɣ matrix maps token IDs to embeddings — vectors in 2,560 dimensions.

I imagine these vectors as stars in the sky. I've been doing observational tokenonomy.
November 4, 2025 at 11:00 PM
I'm genuinely not trying to be a pill about this. This is what's got me so interested in these structures: The tokens in them are _indistinguishable._ They have the exact same embeddings. And 70-odd percent of them are Thai, and that's weeeeeeeird.

But like I said, I've got an experiment cooking. 👨‍🔬
November 6, 2025 at 12:05 AM
via the magic of laion_clap embeddings and umap, my live coding thingy has a sample browser at last!
October 31, 2025 at 6:27 PM
Roo Code 3.30.0 - Now supporting @openrouter.bsky.social embeddings for codebase indexing (incl. top‑ranking Qwen3 Embedding). Plus 12 other tweaks and fixes docs.roocode.com/update-notes...
Roo Code 3.30.0 Release Notes (2025-11-03) | Roo Code Documentation
Roo Code 3.30.0 adds OpenRouter embeddings, reasoning handling improvements, and stability/UI fixes.
docs.roocode.com
November 4, 2025 at 2:56 AM
Semantic search with embeddings in JavaScript: a hands-on example using LangChain and Ollama https://cstu.io/c96e68 #d #it #marketing
Semantic search with embeddings in JavaScript: a hands-on example using LangChain and Ollama
Have you ever searched for something online and got the right result even though you didn’t type the...
cstu.io
November 2, 2025 at 5:38 PM
See below for the story so far. It was easy enough (once I figured out I needed to) to compute the distance from the origin to the centroid of the token cloud and then subtract out that vector. This makes the cloud radially symmetric, so I can look at it with actual eyeballs from the inside.
November 2, 2025 at 10:21 PM
The conventions concerning author order in academic research are complicated, but these Italians working LLM research have, I think, nailed the best way:
November 3, 2025 at 5:00 PM
Depends on the model they're using - most modern language models use contextual embeddings and would capture those semantic differences. But telling whether someone is being emphatic or sarcastic or justifiably upset? Not so much
November 2, 2025 at 1:59 AM
So sentence embeddings of messenger speeches in Attic tragedy show a modest but notable bump in semantic similarity toward the end of the 5th century (see attached image) — generic consolidation, maybe, but why? #ancmedsky #dh
October 31, 2025 at 8:11 PM
LLMs operate on tokens, not characters. Unless their embeddings take fewer tokens, it wouldn't be any faster.
October 31, 2025 at 2:59 PM
Our 9-part Python + AI live series in October covered LLMs, embeddings, RAG, vision models, structured outputs, safety, tool-calling, agents and MCP.

Grab the recordings, slides, and code from: blog.pamelafox.org/2025/10/watc...
Watch the recordings from my Python + AI series
My colleague and I just wrapped up a live series on Python + AI, a nine-part journey diving deep into how to use generative AI models from P...
blog.pamelafox.org
October 31, 2025 at 2:29 PM
LLMs are injective and invertible.

They show that different prompts always map to different embeddings, and this property can be used to recover input tokens from individual embeddings in latent space.

Paper: www.arxiv.org/abs/2510.15511
October 28, 2025 at 2:05 AM
The new 1X NEO robot operates largely using a 160 million (with an m) parameter model that takes instructions as text embeddings from an off-board language model. Surprising that a model that small can even do visual understanding, let alone instruction following and movement.
October 28, 2025 at 11:27 PM