Rishu Kumar
rishuk.bsky.social
Rishu Kumar
@rishuk.bsky.social
Narrator: He definitely let that slide. :D
March 13, 2025 at 4:07 PM
Few obvious questions: Is the upper limit of ~30% hallucination acceptable? And how much work it adds to a lawyer’s work to ensure there is no hallucinations at all in the generated document?

(Haven’t read the entire paper yet)
March 4, 2025 at 8:44 PM
Reposted by Rishu Kumar
Honest answer: I don't read anything from the enormous stream of papers unless they are *critical* to my current work or I am deeply interested in the topic.
February 28, 2025 at 4:51 PM
Okay, automated cross-posting has its quirks. This was the original tweet: x.com/OpenAI/statu...
x.com
x.com
February 19, 2025 at 12:31 AM
Next year we will reach NeurIPS level at this rate.
February 17, 2025 at 1:57 AM
If we start calling them anything else, it doesn’t anthropomorphise the models enough. :)
February 15, 2025 at 10:56 PM
I don’t think any of the “foundation model” orgs will ever talk about contamination, because that makes scores look bad. At this point a lot of these models seem like clowns in circus, only there to keep the public (investors?) at large excited about the circus that is “AI research”.
January 28, 2025 at 9:19 PM