✅ 3B active params
✅ 100% multimodal reasoning
✅ Visual reasoning, STEM, video understanding & “Thinking with Images”
✅ Tool use, precise grounding, dynamic zoom & search
👉 github.com/PaddlePaddle...
✅ 3B active params
✅ 100% multimodal reasoning
✅ Visual reasoning, STEM, video understanding & “Thinking with Images”
✅ Tool use, precise grounding, dynamic zoom & search
👉 github.com/PaddlePaddle...
This is an exciting step towards a more intuitive, responsive, and creative future of AI content creation.
joonghyuk.com/motionstream...
This is an exciting step towards a more intuitive, responsive, and creative future of AI content creation.
joonghyuk.com/motionstream...
Two different approaches to Document AI. Check them out!
github.com/deepseek-ai/...
Two different approaches to Document AI. Check them out!
github.com/deepseek-ai/...
✅ Beats Kimi-K2 & DeepSeek-V3
✅ Top in math (AIME’25)
✅ Efficient MoE design
✅ Strong multimodal & tool-use (~70% BFCL V3)
github.com/inclusionAI/Ling-V2
✅ Beats Kimi-K2 & DeepSeek-V3
✅ Top in math (AIME’25)
✅ Efficient MoE design
✅ Strong multimodal & tool-use (~70% BFCL V3)
github.com/inclusionAI/Ling-V2
👉 Key Insight: Demonstrates that high‑level reasoning on challenging tasks can be attained without large‑scale foundational models.
github.com/SamsungSAILM...
👉 Key Insight: Demonstrates that high‑level reasoning on challenging tasks can be attained without large‑scale foundational models.
github.com/SamsungSAILM...
fanegg.github.io/Human3R/
fanegg.github.io/Human3R/
github.com/deepseek-ai/...
github.com/deepseek-ai/...
The model has 80 billion parameters and is currently the most powerful and largest open‑source image‑generation model available.
github.com/Tencent-Huny...
The model has 80 billion parameters and is currently the most powerful and largest open‑source image‑generation model available.
github.com/Tencent-Huny...
arxiv.org/abs/2509.05276
arxiv.org/abs/2509.05276
humanaigc.github.io/wan-animate/
humanaigc.github.io/wan-animate/
github.com/DecartAI/luc...
github.com/DecartAI/luc...
3D Reconstruction, a simple, end-to-end trained transformer model that directly regresses the factored metric 3D geometry of a scene given various types of inputs (images, calibration, poses, or depth).
github.com/facebookrese...
3D Reconstruction, a simple, end-to-end trained transformer model that directly regresses the factored metric 3D geometry of a scene given various types of inputs (images, calibration, poses, or depth).
github.com/facebookrese...
The release includes model weights, technical report, model card, and starter code.
github.com/facebookrese...
The release includes model weights, technical report, model card, and starter code.
github.com/facebookrese...
github.com/QwenLM/Qwen3...
github.com/QwenLM/Qwen3...
Find a brief breakdown of current GPU types, sorted by performance in our blog article: www.aime.info/blog/en/deep...
Find a brief breakdown of current GPU types, sorted by performance in our blog article: www.aime.info/blog/en/deep...
👉 The AIME G500E is designed as maintainable efficient multi-GPU workstation with enough cooling and PSU capacity to host up to four high-end GPUs.
📺 Have a look: www.aime.info/en/shop/prod...
👉 The AIME G500E is designed as maintainable efficient multi-GPU workstation with enough cooling and PSU capacity to host up to four high-end GPUs.
📺 Have a look: www.aime.info/en/shop/prod...
Link to Paper: raw.githubusercontent.com/swiss-ai/ape...
Link to GitHub: github.com/swiss-ai/
Link to weights: huggingface.co/collections/...
Link to Paper: raw.githubusercontent.com/swiss-ai/ape...
Link to GitHub: github.com/swiss-ai/
Link to weights: huggingface.co/collections/...
It is a transformer-based architecture with 560 billion parameters and a 1 million token context window. Its multi-modal capabilities includes text, code, and image understanding and supports over 100 languages.
deepseek.ai/blog/deepsee...
It is a transformer-based architecture with 560 billion parameters and a 1 million token context window. Its multi-modal capabilities includes text, code, and image understanding and supports over 100 languages.
deepseek.ai/blog/deepsee...
The models are vLLM- and SGlang-ready!
github.com/zai-org/GLM-V/
The models are vLLM- and SGlang-ready!
github.com/zai-org/GLM-V/
The model is now natively supported in ComfyUI.
It’s said to outperform FLUX and comparable models.
github.com/QwenLM/Qwen-...
The model is now natively supported in ComfyUI.
It’s said to outperform FLUX and comparable models.
github.com/QwenLM/Qwen-...
www.aime.info/de/shop/prod...
www.aime.info/de/shop/prod...
👉 GLM-4.5: 355B total / 32B active parameters
👉 GLM-4.5-Air: 106B total / 12B active parameters
github.com/zai-org/GLM-...
👉 GLM-4.5: 355B total / 32B active parameters
👉 GLM-4.5-Air: 106B total / 12B active parameters
github.com/zai-org/GLM-...
github.com/MoonshotAI/K...
github.com/MoonshotAI/K...
github.com/wasserth/Tot...
github.com/wasserth/Tot...
bfl.ai/announcement...
bfl.ai/announcement...