Ramesh Manuvinakurike
banner
rameshddrr.bsky.social
Ramesh Manuvinakurike
@rameshddrr.bsky.social
AI researcher/scientist. PhD from USC (LA). Passionate about learning ...
Finished reading this really nice survey paper. Love reading survey papers as a method to revise and fill in gaps in my understanding of a topic ..

arxiv.org/abs/2406.16838
From Decoding to Meta-Generation: Inference-time Algorithms for Large Language Models
One of the most striking findings in modern research on large language models (LLMs) is that scaling up compute during training leads to better results. However, less attention has been given to the b...
arxiv.org
December 31, 2024 at 7:53 PM
Couldn't present my own poster at neurips workshop because I couldn't go to the poster hall with my baby ... apparently some workshops are 14 years + .. why 14+ ? No such issue with main conference .. @neuripsconf.bsky.social
December 14, 2024 at 9:59 PM
@neuripsconf.bsky.social kudos ... a conference at this scale and such high quality .. truly mindblowing ...
December 13, 2024 at 6:15 PM
Reposted by Ramesh Manuvinakurike
First slide deck for NeurIPS is done -- an overview of how I view post-training for applications.
A higher level summary on the key decisions along the way of scoping a problem, choosing a base model, optimization algorithm, etc. (+some thoughts on OpenAI's RL Finetuning).

https://buff.ly/3ZpY5IR
December 9, 2024 at 7:04 PM
Reposted by Ramesh Manuvinakurike
The extraordinary recent takeover of ML/AI by #NLP is well-known but insufficiently reflected on.

Look at the @neuripsconf.bsky.social tutorials in 2024!

neurips.cc/virtual/2024...

14 tutorials; 6 have "LLM" in the title; 4 more cover foundation models, with large NLP coverage. That's > 70% 😲
NeurIPS 2024 TutorialsNeurIPS 2024
neurips.cc
December 9, 2024 at 7:29 PM
@neuripsconf.bsky.social has some awesome tutorials list ... If only time turner was available ... Looking forward to these !
December 9, 2024 at 6:21 PM
Reposted by Ramesh Manuvinakurike
For anyone interested in fine-tuning or aligning LLMs, I’m running this free and open course called smol course. It’s not a big deal, it’s just smol.

🧵>>
December 3, 2024 at 9:21 AM
When I speak to non-tech people, I get equally scared and encouraged. One one hand they truly underestimate the power of AI and on the other they don't care much about it ...
December 5, 2024 at 4:54 AM
In-context learning can be very difficult to understand. This thread from a long time back has some interesting points ...

www.reddit.com/r/MachineLea...
From the MachineLearning community on Reddit
Explore this post and more from the MachineLearning community
www.reddit.com
December 4, 2024 at 5:25 PM
Dspy + Gradio + Huggingface = Magic !!
December 4, 2024 at 5:16 PM
Listening to this awesome talk from @cgpotts.bsky.social .. so in love with the message here ..

As I'm building systems the most common questions (and review comments) I get asked is about the LL(M)M I'm using and not the systems and the problems they're solving ..

youtu.be/vRTcE19M-KE?...
youtu.be
December 4, 2024 at 5:11 AM
I wonder if we can create "crying mindlessly" mode in the LLMs. You know, a mode where a baby cries a lot and stops immediately when shown something completely random and uninteresting ...
December 3, 2024 at 10:10 PM
One abilities we humans have is to select role models depending on our likings. I aspire to be as nice and humble as my primary school friend from 7th grade and not rude and arrogant as a certain billionaire. Our ability select the samples we train our policy astonishes me !!!
December 1, 2024 at 10:56 PM
One of the first papers I read during my PhD was "Referring as a Collaborative Process". I tried to test out the ChatGPT and Claude on some of the problems mentioned.

For instance, something as simple as self - repairs in the same utterance confuses the model .. hmm ..
November 30, 2024 at 11:03 PM