Riyadh
riyadhrazzaq.bsky.social
Riyadh
@riyadhrazzaq.bsky.social
NLP Master's @UPV/EHU
Reposted by Riyadh
#Latxa txatbota probatarako erabilgarri jarri dugu! latxa.hitz.eus

Jaso ditugun eskaerei erantzunez zuen eskura jarri dugu Latxaren bertsio ahaltsuena, chatGPT-tik gertu dabilena, baina euskara txukunagoa sortuz.
Gradio
Click to try out the app!
latxa.hitz.eus
October 31, 2025 at 6:57 AM
A fun read on Deep RL from Alex Irpan.
www.alexirpan.com/2018/02/14/r...
July 29, 2025 at 3:11 PM
Reposted by Riyadh
In summary, the infra layer is eating the model layer, the model layer is eating the application layer, the application layer might soon eat the people layer. AI food chain.
March 23, 2025 at 3:06 PM
Reposted by Riyadh
« appending "Wait" multiple times to the model's generation » is our current most likely path to AGI :)

See the fresh arxiv.org/abs/2501.19393 by Niklas Muennighoff et al.
February 3, 2025 at 2:31 PM
Reposted by Riyadh
Cohere's Maya: Open-source (code, dataset, and model) Multimodal Multilingual LLM

- Paper: Maya: An Instruction Finetuned Multilingual Multimodal Model ( arxiv.org/abs/2412.07112 )
- Model and Dataset: huggingface.co/maya-multimo...
- Repo: github.com/nahidalam/maya
December 11, 2024 at 8:42 AM
Reposted by Riyadh
Have you ever wondered how to train an autoregressive generative transformer on text and raw pixels, without a pretrained visual tokenizer (e.g. VQ-VAE)?

We have been pondering this during summer and developed a new model: JetFormer 🌊🤖

arxiv.org/abs/2411.19722

A thread 👇

1/
December 2, 2024 at 4:41 PM
I wrote a long note on LLaMA 3 based speech models.
Paper Notes / Multimodal LLMs on top of LLaMA 3 | Md Abdur Razzaq Riyadh
notes on paper read in Nov, 24.
riyadhrazzaq.github.io
December 1, 2024 at 8:36 PM
Reposted by Riyadh
get it while it’s hot #ai

go.bsky.app/Pxcnfu6
November 18, 2024 at 1:36 PM