Ujan
banner
ujaan.bsky.social
Ujan
@ujaan.bsky.social
Learning NLP. PhD student @UNSW Sydney
Reposted by Ujan
The problem with most machine-based random number generators is that they’re not TRULY random, so if you need genuine randomness it is sometimes necessary to link your code to an external random process like a physical noise source or the current rate of US tariffs on a given country.
April 9, 2025 at 7:15 PM
Reposted by Ujan
This is horrifying. She co-wrote an OpEd at her school, then was grabbed by masked people, put into an unmarked black van, is now detained in an ICE facility over 1000 miles away.
youtu.be/PuFIs7OkzYY
Trying to limit the topics I post about, but this is my country and this one shook me. =(
Surveillance shows Tufts graduate student detained
YouTube video by WCVB Channel 5 Boston
youtu.be
March 26, 2025 at 11:05 PM
Reposted by Ujan
Major weaknesses raised by Reviewer 2
February 14, 2025 at 7:14 PM
Reposted by Ujan
any international/intergovernmental report on “AI safety” that fails to acknowledge/ground recommendation on the growing unprecedented corporate power and monopoly is not a friend of the public, civil society, and definitely of marginalised and impacted groups
February 3, 2025 at 12:23 PM
Reposted by Ujan
In fact, I have coauthored a whole paper arguing that the networked nature of modern production can amplify the harm of a small or moderate shock...

way beyond what standard models predict.

bengolub.net/snff-2/

4/
February 2, 2025 at 12:59 PM
Reposted by Ujan
It turns out that = trust me bro
January 29, 2025 at 11:51 AM
Reposted by Ujan
One of my grand interpretability goals is to improve human scientific understanding by analyzing scientific discovery models, but this is the most convincing case yet that we CAN learn from model interpretation: Chess grandmasters learned new play concepts from AlphaZero's internal representations.
Bridging the Human-AI Knowledge Gap: Concept Discovery and Transfer in AlphaZero
Artificial Intelligence (AI) systems have made remarkable progress, attaining super-human performance across various domains. This presents us with an opportunity to further human knowledge and improv...
arxiv.org
January 27, 2025 at 9:43 PM
Reposted by Ujan
I recently talked to someone doing a PhD in multivariate calculus.

Or, as they would call it, a grad student
January 26, 2025 at 9:54 AM
Reposted by Ujan
When they introduced group projects in school they did not tell us that life would be a continuous string of group projects until we retire
January 25, 2025 at 10:37 PM
Reposted by Ujan
After several rounds, to the granting agency that sent me this email, with the following subject line:
"Your application has moved to the next stage"

Well done.

That is some Grade 'A' level trolling.

You got me.
January 24, 2025 at 4:04 PM
Reposted by Ujan
“I don’t know when this will happen but probably in 10 years” is researcher-speak for “I have no clue”
January 24, 2025 at 1:47 PM
Reposted by Ujan
I hate what deep learning has done to the word "inference". Now we are forced to write sentences in our papers like "approximate inference at inference time"—I'm all for nonsense Dr Seuss language, but not in a scientific paper!
January 23, 2025 at 12:02 PM
Reposted by Ujan
Every time I go to a public library I begin to believe again that we can create amazing things that are just there for all to enjoy and use for free.
January 3, 2025 at 8:42 PM
Reposted by Ujan
If you have ever wondered what colonization was like—the violence, the dispossession, the massacres, the genocide, the displacement, the famine, the dehumanization, the racism—you are watching it in real time in Palestine. It is among the most depraved of all possible acts.
December 25, 2024 at 9:34 AM
Reposted by Ujan
With the new OpenAI O3 moving performance from 5% up to 25% on FrontierMath it’s time to push open-source models upwards!

We're super happy to release FineMath, the best open math dataset yet. A strong baseline to start training your own models

Find it in the trending section of HuggingFace ;)
December 23, 2024 at 1:20 PM
Reposted by Ujan
"a blue ball is on a wooden table, perfectly motionless. There is no motion in the scene. The camera does not move or pan in any way."
December 17, 2024 at 1:27 PM
Reposted by Ujan
ethernet? you mean the wifi cable
December 9, 2024 at 5:01 PM
Reposted by Ujan
AI influencers "this <big company name / Stanford / Berkeley / MIT>" person/paper says ..... Yes, and? there is always so much context.
But social media incentivizes non-nuanced bite-sized panic/optimism inducing takes. How do we increase the context with which information is disemminated in media?
December 5, 2024 at 6:17 AM
Reposted by Ujan
Two stages of writing:

1) This shouldn't take too long
2) Oh no
December 4, 2024 at 7:36 PM
Reposted by Ujan
We just updated the OLMo repo at github.com/allenai/OLMo!
There are now several training configs that together reproduce the training runs that lead to the final OLMo 2 models.
In particular, all the training data is available, tokenized and shuffled exactly as we trained on it!
GitHub - allenai/OLMo: Modeling, training, eval, and inference code for OLMo
Modeling, training, eval, and inference code for OLMo - allenai/OLMo
github.com
December 2, 2024 at 8:13 PM
Reposted by Ujan
Let's cycle through the memes for this one until it stops... 😇😅🙏
December 2, 2024 at 1:23 AM
Reposted by Ujan
🎉Good news, everyone! 🎉
I will recruit graduate students on the algorithmic and theoretical aspects of Reinforcement Learning.
You will join Adage, @mila-quebec.bsky.social and @polymtl.bsky.social.
More info on why and how you should apply:
academic.sologen.net/2024/11/22/g...
Deadline: Dec 1st
November 29, 2024 at 3:23 AM
Reposted by Ujan
Space: *amorphous blob*

Astronomers: ah yes the choo choo train nebula
November 24, 2024 at 8:54 AM
Reposted by Ujan
Ok, last two papers for this week!

A final game-theoretic RLHF method and a different take on RLHF altogether inspired by prospect theory.

1. 🧲 Magnetic Preference Optimization (MPO).

2. Kahneman-Tversky Optimization (KTO).

🧵 1/3.
Last week, I shared some papers in the intersection of agent/model evaluation and social choice theory.

The last was a position paper on RLHF/alignment.

This week I will share papers (in pairs) on the topic of "game-theoretic or social choice meet meet alignment/RLHF".

🧵 1/3.
November 22, 2024 at 12:43 PM
Reposted by Ujan
I have become a fan of the game-theoretic approaches to RLHF, so here are two more papers in that category! (with one more tomorrow 😅)

1. Self-Play Preference Optimization (SPO).

2. Direct Nash Optimization (DNO).

🧵 1/3.
Last week, I shared some papers in the intersection of agent/model evaluation and social choice theory.

The last was a position paper on RLHF/alignment.

This week I will share papers (in pairs) on the topic of "game-theoretic or social choice meet meet alignment/RLHF".

🧵 1/3.
November 21, 2024 at 12:30 PM