debuinu.com
@debuinu.com
Reposted
I asked ChatGPT to tell me about the history of the Super Mario Bros any% world record.

I think my job is safe for a while...
January 18, 2025 at 12:54 AM
Cool paper where they use a conv/natten-like scanning method of attention to dramatically speed up FLUX, especially with large sizes.
arxiv.org/abs/2412.16112
CLEAR: Conv-Like Linearization Revs Pre-Trained Diffusion Transformers Up
Diffusion Transformers (DiT) have become a leading architecture in image generation. However, the quadratic complexity of attention mechanisms, which are responsible for modeling token-wise relationsh...
arxiv.org
December 23, 2024 at 9:44 PM
*slaps 8b llama3 llm*
this baby can fit loads of generative neural images for diffusion model conditioning in just a little finetuning
arxiv.org/abs/2412.14164
MetaMorph: Multimodal Understanding and Generation via Instruction Tuning
In this work, we propose Visual-Predictive Instruction Tuning (VPiT) - a simple and effective extension to visual instruction tuning that enables a pretrained LLM to quickly morph into an unified auto...
arxiv.org
December 20, 2024 at 4:47 AM
Hello world!
November 23, 2024 at 5:41 PM