Jan-v2-VL executes 49 steps without failure, while the base model stops at 5 and other similar-scale VLMs stop between 1 and 2.
Models: huggingface.co/collections...
Credit to the Qwen team for Qwen3-VL-8B-Thinking!
Jan-v2-VL executes 49 steps without failure, while the base model stops at 5 and other similar-scale VLMs stop between 1 and 2.
Models: huggingface.co/collections...
Credit to the Qwen team for Qwen3-VL-8B-Thinking!
Find the GGUF model on Hugging Face, click "Use this model" and select Jan, or copy the model link and paste it into Jan Hub.
Thanks Qwen 🧡
Find the GGUF model on Hugging Face, click "Use this model" and select Jan, or copy the model link and paste it into Jan Hub.
Thanks Qwen 🧡
Go to Settings, Model Providers, add Ollama, and set the Base URL to http://localhost:11434/v1.
Your 🦙 Ollama models will then be ready to use in 👋 Jan.
Go to Settings, Model Providers, add Ollama, and set the Base URL to http://localhost:11434/v1.
Your 🦙 Ollama models will then be ready to use in 👋 Jan.
Find it in Jan Hub, download, and you're good to go.
Find it in Jan Hub, download, and you're good to go.
This update brings:
- Projects to group related chats
- Model context stats
- Auto-loaded cloud models
- Support for Exa as an MCP Server
Update your Jan or download the latest version.
This update brings:
- Projects to group related chats
- Model context stats
- Auto-loaded cloud models
- Support for Exa as an MCP Server
Update your Jan or download the latest version.
Jan Hub searches @huggingface and runs them with @ggml's llama.cpp 💙
Jan Hub searches @huggingface and runs them with @ggml's llama.cpp 💙
- Import your vision models
- Experimental setting auto-adjusts llama.cpp for your system
- Fixed: image attachments, copy glitches, API key visibility, and more
Update your Jan or download the latest.
- Import your vision models
- Experimental setting auto-adjusts llama.cpp for your system
- Fixed: image attachments, copy glitches, API key visibility, and more
Update your Jan or download the latest.
Create your own assistant in 👋 Jan, make it a lawyer, finance, or customer support.
Choose an open-source model and keep every word private.
Create your own assistant in 👋 Jan, make it a lawyer, finance, or customer support.
Choose an open-source model and keep every word private.
Performance:
- Gains on reasoning & creativity benchmarks
- Small drop on SimpleQA (91.1 -> 90.7)
Model:
- huggingface.co/janhq/Jan-v...
- GGUF: huggingface.co/janhq/Jan-v...
Credit to the Qwen team for Qwen3-4B-Thinking and GGML for llama.cpp.
Performance:
- Gains on reasoning & creativity benchmarks
- Small drop on SimpleQA (91.1 -> 90.7)
Model:
- huggingface.co/janhq/Jan-v...
- GGUF: huggingface.co/janhq/Jan-v...
Credit to the Qwen team for Qwen3-4B-Thinking and GGML for llama.cpp.
We updated llama.cpp with the latest improvements and fixed a CUDA backend bug upstream. Jan stays pinned to v6324 due to flash-attention changes.
If you have auto-update on, you'll get this automatically - if not, we recommend turning it on.
We updated llama.cpp with the latest improvements and fixed a CUDA backend bug upstream. Jan stays pinned to v6324 due to flash-attention changes.
If you have auto-update on, you'll get this automatically - if not, we recommend turning it on.
We're testing how well web search & reasoning can transfer into a smaller 1.7B model.
- Jan-v1-edge: huggingface.co/janhq/Jan-v...
- Jan-v1-edge GGUF: huggingface.co/janhq/Jan-v...
We're testing how well web search & reasoning can transfer into a smaller 1.7B model.
- Jan-v1-edge: huggingface.co/janhq/Jan-v...
- Jan-v1-edge GGUF: huggingface.co/janhq/Jan-v...