Yoav Goldberg
@yoavgo.bsky.social
Pinned
Yoav Goldberg
@yoavgo.bsky.social
· Dec 16
my current belief is that while thinking of DL in the lens of NLP was expanding, thinking of LLMs with the lens of NLP is mostly limiting.
חשבתי שההתעלמות מסודאן היא סתם סוג-של הזנחה, אבל הנה אני קורא ספר שמתאמץ מאד להראות ש:
- מה שהיה בדארפור זה לא ג'נוסייד אלא רק תוצאת לוואי של מלחמת אזרחים אכזרית
- "Darfur is where the language of genocide had become an instrument"
- הערבים אינם settlers באפריקה, הם הגיעו ממקומות שונים
- המערב אשם
- מה שהיה בדארפור זה לא ג'נוסייד אלא רק תוצאת לוואי של מלחמת אזרחים אכזרית
- "Darfur is where the language of genocide had become an instrument"
- הערבים אינם settlers באפריקה, הם הגיעו ממקומות שונים
- המערב אשם
November 8, 2025 at 9:58 AM
חשבתי שההתעלמות מסודאן היא סתם סוג-של הזנחה, אבל הנה אני קורא ספר שמתאמץ מאד להראות ש:
- מה שהיה בדארפור זה לא ג'נוסייד אלא רק תוצאת לוואי של מלחמת אזרחים אכזרית
- "Darfur is where the language of genocide had become an instrument"
- הערבים אינם settlers באפריקה, הם הגיעו ממקומות שונים
- המערב אשם
- מה שהיה בדארפור זה לא ג'נוסייד אלא רק תוצאת לוואי של מלחמת אזרחים אכזרית
- "Darfur is where the language of genocide had become an instrument"
- הערבים אינם settlers באפריקה, הם הגיעו ממקומות שונים
- המערב אשם
how many of you are aware of ai-2027? how many of you read it? curious what ya'll think.
November 2, 2025 at 9:36 PM
how many of you are aware of ai-2027? how many of you read it? curious what ya'll think.
"in a future in which AI assistants are being used by world leaders and influential figures, aren't you afraid of rogue AI using this access to control the world?"
well maybe, but I am more worried about near future in which these influentials act on random AI advice.
well maybe, but I am more worried about near future in which these influentials act on random AI advice.
October 29, 2025 at 12:15 PM
"in a future in which AI assistants are being used by world leaders and influential figures, aren't you afraid of rogue AI using this access to control the world?"
well maybe, but I am more worried about near future in which these influentials act on random AI advice.
well maybe, but I am more worried about near future in which these influentials act on random AI advice.
LISP code does not have significantly more parentheses than other typical languages. change my mind.
October 17, 2025 at 11:41 AM
LISP code does not have significantly more parentheses than other typical languages. change my mind.
this "feature" in code editors where you type an opening character and it immediately inserts the closing one for you as well -- why?? at the very very best case, you will have to skip over this character with an arrow. so why?
September 11, 2025 at 6:18 PM
this "feature" in code editors where you type an opening character and it immediately inserts the closing one for you as well -- why?? at the very very best case, you will have to skip over this character with an arrow. so why?
can someone explain "serverless backends" to me? it seems that they run functions on demand. but if these functions cannot access any persistent state, why not run them on the client? the only reason I see is to hide DBs/APIs tokens/secrets from the client, but is that really all there is to it?
September 6, 2025 at 8:55 AM
can someone explain "serverless backends" to me? it seems that they run functions on demand. but if these functions cannot access any persistent state, why not run them on the client? the only reason I see is to hide DBs/APIs tokens/secrets from the client, but is that really all there is to it?
a trivia fact about this paper is that we submitted it to arxiv weeks ago, and it was hanging there in limbo for quite a while. apparently because we submitted to "AI" while they moved it to "HCI".
When reading AI reasoning text (aka CoT), we (humans) form a narrative about the underlying computation process, which we take as a transparent explanation of model behavior. But what if our narratives are wrong? We measure that and find it usually is.
Now on arXiv: arxiv.org/abs/2508.16599
Now on arXiv: arxiv.org/abs/2508.16599
Humans Perceive Wrong Narratives from AI Reasoning Texts
A new generation of AI models generates step-by-step reasoning text before producing an answer. This text appears to offer a human-readable window into their computation process, and is increasingly r...
arxiv.org
August 27, 2025 at 9:45 PM
a trivia fact about this paper is that we submitted it to arxiv weeks ago, and it was hanging there in limbo for quite a while. apparently because we submitted to "AI" while they moved it to "HCI".
When reading AI reasoning text (aka CoT), we (humans) form a narrative about the underlying computation process, which we take as a transparent explanation of model behavior. But what if our narratives are wrong? We measure that and find it usually is.
Now on arXiv: arxiv.org/abs/2508.16599
Now on arXiv: arxiv.org/abs/2508.16599
Humans Perceive Wrong Narratives from AI Reasoning Texts
A new generation of AI models generates step-by-step reasoning text before producing an answer. This text appears to offer a human-readable window into their computation process, and is increasingly r...
arxiv.org
August 27, 2025 at 9:30 PM
When reading AI reasoning text (aka CoT), we (humans) form a narrative about the underlying computation process, which we take as a transparent explanation of model behavior. But what if our narratives are wrong? We measure that and find it usually is.
Now on arXiv: arxiv.org/abs/2508.16599
Now on arXiv: arxiv.org/abs/2508.16599
if you REALLY want to understand DL, you should start by honing your Category Theory skills, as almost everything in DL at its core can be mapped to a functor or an endofunctor.
July 4, 2025 at 10:39 AM
if you REALLY want to understand DL, you should start by honing your Category Theory skills, as almost everything in DL at its core can be mapped to a functor or an endofunctor.
taking it a step further, I'd say in many cases using the algebra jargon is harmful to understanding, and its better to just describe whats really going on. ie, "we add an L2 penalty term" --> want the sum of squares to be small. "project to vocab space" --> compute similarity to each vocab item.
i'll elaborate: a common computation pattern in DL happens to coincide with a known operator in linear algebra (matmul), and so we conveniently borrow linalg notation and terminology (matrices, vectors, ranks, norms). but this is just jargon. the algebric properties arent needed.
"Modern ML is built on Linear Algebra".
lol no its not.
lol no its not.
July 4, 2025 at 9:14 AM
taking it a step further, I'd say in many cases using the algebra jargon is harmful to understanding, and its better to just describe whats really going on. ie, "we add an L2 penalty term" --> want the sum of squares to be small. "project to vocab space" --> compute similarity to each vocab item.
i'll elaborate: a common computation pattern in DL happens to coincide with a known operator in linear algebra (matmul), and so we conveniently borrow linalg notation and terminology (matrices, vectors, ranks, norms). but this is just jargon. the algebric properties arent needed.
"Modern ML is built on Linear Algebra".
lol no its not.
lol no its not.
July 3, 2025 at 7:38 PM
i'll elaborate: a common computation pattern in DL happens to coincide with a known operator in linear algebra (matmul), and so we conveniently borrow linalg notation and terminology (matrices, vectors, ranks, norms). but this is just jargon. the algebric properties arent needed.
"Modern ML is built on Linear Algebra".
lol no its not.
lol no its not.
July 3, 2025 at 6:41 PM
"Modern ML is built on Linear Algebra".
lol no its not.
lol no its not.
why is "MCP" implemented as server exposing a set of endpoints, rather than as some JSON schema for defining tool descriptions and allowing these JSON files to be accessed over http? what is the purpose/benefit of the middleman server?
July 3, 2025 at 10:01 AM
why is "MCP" implemented as server exposing a set of endpoints, rather than as some JSON schema for defining tool descriptions and allowing these JSON files to be accessed over http? what is the purpose/benefit of the middleman server?
you know what, nah, we don't want to close it. it will be just 80% closed.
and now, we will proceed to peacefully close the strait of Hormuz. you know, for the environment.
June 22, 2025 at 1:32 PM
you know what, nah, we don't want to close it. it will be just 80% closed.
and now, we will proceed to peacefully close the strait of Hormuz. you know, for the environment.
June 22, 2025 at 1:21 PM
and now, we will proceed to peacefully close the strait of Hormuz. you know, for the environment.
today, during a peaceful flight over an iranian mountain, a US airplane dropped a mostly peaceful bunker buster bomb, who flew peacefully until it hit the mountain and mostly peaceful facility underneath it. there was a brief period of violent detonation on impact, then peace again.
June 22, 2025 at 12:01 PM
today, during a peaceful flight over an iranian mountain, a US airplane dropped a mostly peaceful bunker buster bomb, who flew peacefully until it hit the mountain and mostly peaceful facility underneath it. there was a brief period of violent detonation on impact, then peace again.
אחד מלקחי ליל המקלטים אמש הוא שאין לי סבלנות לקרוא ספרות מקצועית, אבל לקרוא ספרות קלה זה די סבבה. מצד שני הספר שהיה לי בנייד, הוא כזה שהתחלתי לקרוא והפסקתי והיתה סיבה שהפסקתי, הוא מייגע ומעפן. בקיצור שילחו המלצות לספרים. עברית או אנגלית, אבל באנגלית ככהנ יהיה לי יותר קל להתארגן.
June 14, 2025 at 4:16 PM
אחד מלקחי ליל המקלטים אמש הוא שאין לי סבלנות לקרוא ספרות מקצועית, אבל לקרוא ספרות קלה זה די סבבה. מצד שני הספר שהיה לי בנייד, הוא כזה שהתחלתי לקרוא והפסקתי והיתה סיבה שהפסקתי, הוא מייגע ומעפן. בקיצור שילחו המלצות לספרים. עברית או אנגלית, אבל באנגלית ככהנ יהיה לי יותר קל להתארגן.
there are lots of "AI agents are just software around LLMs" comments recently. this is true by definition, but also misses the point. while there is no "magic" in software, there is engineering and best practices. and we dont yet have them for systems that involve LLMs.
June 14, 2025 at 12:30 PM
there are lots of "AI agents are just software around LLMs" comments recently. this is true by definition, but also misses the point. while there is no "magic" in software, there is engineering and best practices. and we dont yet have them for systems that involve LLMs.
a good gag from the other social network:
June 8, 2025 at 9:35 PM
a good gag from the other social network:
this "class 9" is such a cool idea for an LLM course!
(from ufal.mff.cuni.cz/courses/npfl... via @zdenekkasner.bsky.social )
(from ufal.mff.cuni.cz/courses/npfl... via @zdenekkasner.bsky.social )
June 8, 2025 at 8:45 PM
this "class 9" is such a cool idea for an LLM course!
(from ufal.mff.cuni.cz/courses/npfl... via @zdenekkasner.bsky.social )
(from ufal.mff.cuni.cz/courses/npfl... via @zdenekkasner.bsky.social )
June 8, 2025 at 1:36 PM
Reposted by Yoav Goldberg
A prompt posted by Twitter user eigenrobot
June 4, 2025 at 8:30 PM
A prompt posted by Twitter user eigenrobot
what's an "eigenprompt"?
June 4, 2025 at 7:57 PM
what's an "eigenprompt"?
i am looking for academic (or similarly "serious") courses or textbook-y texts on LLMs with a major component (lets say >66%) that is NOT about the technicalities of Transformers, optimization, RL, tokenization, etc, but instead focus more on data, eval, interpretability, etc. pointers?
June 3, 2025 at 5:41 PM
i am looking for academic (or similarly "serious") courses or textbook-y texts on LLMs with a major component (lets say >66%) that is NOT about the technicalities of Transformers, optimization, RL, tokenization, etc, but instead focus more on data, eval, interpretability, etc. pointers?