On an unrelated note, can we train latent diffusion models for text? 🤔 4/4
On an unrelated note, can we train latent diffusion models for text? 🤔 4/4
This leads to much more stable training and better overall results. 3/4
This leads to much more stable training and better overall results. 3/4
The wall is about diminishing returns in scaling LLM pre-training. From GPT-1 to GPT-4, we've benefited from scaling model size and dataset size together.
The wall is about diminishing returns in scaling LLM pre-training. From GPT-1 to GPT-4, we've benefited from scaling model size and dataset size together.
Full-on reinforcement learning, based on the interaction of the agent with the environment!
LLM is the agent, and user is the environment. With OpenAI having hundred of millions of monthly active users, I think they can do RL with real-world interactions to keep improving their model.
Full-on reinforcement learning, based on the interaction of the agent with the environment!
LLM is the agent, and user is the environment. With OpenAI having hundred of millions of monthly active users, I think they can do RL with real-world interactions to keep improving their model.