Fine-tuning and LLMOps
My views
https://www.linkedin.com/in/erwinhuizenga/
🔹 Improved benchmark performance
🔹Enhanced front-end capabilities
🔹Supports JSON output & function calling
We've made sure its available from day one on Vertex AI for deployment. with Nvidia's H200
🔹 Improved benchmark performance
🔹Enhanced front-end capabilities
🔹Supports JSON output & function calling
We've made sure its available from day one on Vertex AI for deployment. with Nvidia's H200
Being part of this program will give startups early access to Google's latest AI models as well as resources, technical expertise and equity funding to accelerate their progress.
Apply here:
labs.google/aifuturesfund
Being part of this program will give startups early access to Google's latest AI models as well as resources, technical expertise and equity funding to accelerate their progress.
Apply here:
labs.google/aifuturesfund
Today, we're announcing:
1️⃣An A2A Python SDK AND an updated A2A protocol spec (v0.2).
2️⃣A new Java SDK for Agent Development Kit (ADK) AND the v1.0.0 stable release of our Python ADK.
Read more in our launch blog 👇
developers.googleblog.com/en/agents-ad...
Today, we're announcing:
1️⃣An A2A Python SDK AND an updated A2A protocol spec (v0.2).
2️⃣A new Java SDK for Agent Development Kit (ADK) AND the v1.0.0 stable release of our Python ADK.
Read more in our launch blog 👇
developers.googleblog.com/en/agents-ad...
I'm excited about our announcements. Watch the livestream to stay ahead of the curve.
Link below 👇
I'm excited about our announcements. Watch the livestream to stay ahead of the curve.
Link below 👇
What I find impressive is that a single system can contribute to numerous open problems in maths and computer science. Which proves that Agents can discover novel and useful algorithms
What I find impressive is that a single system can contribute to numerous open problems in maths and computer science. Which proves that Agents can discover novel and useful algorithms
What it offers:
🛠️ Build with the Agent Development Kit (ADK)
🏆 $50K cash prize pool
🧑🏫 Insights from Google experts
Sign up link below 👇
What it offers:
🛠️ Build with the Agent Development Kit (ADK)
🏆 $50K cash prize pool
🧑🏫 Insights from Google experts
Sign up link below 👇
Offering better visual quality and more accurate text rendering.
Notebook below to get you started 👇
Offering better visual quality and more accurate text rendering.
Notebook below to get you started 👇
It' got better at coding, with significant gains in front-end web development, editing, and transformation. We also fixed a bunch of function calling issue.
You can access it as gemini-2.5-pro-preview-05-06 on VertexAI and AIStudio
More details in 🧵
It' got better at coding, with significant gains in front-end web development, editing, and transformation. We also fixed a bunch of function calling issue.
You can access it as gemini-2.5-pro-preview-05-06 on VertexAI and AIStudio
More details in 🧵
> Agentic RAG
> Agent Evaluation
> Multi-Agent Evaluation
> Production Architectures
If you're comfortable with agent basics and looking to understand the next layer this is a great read 👇
> Agentic RAG
> Agent Evaluation
> Multi-Agent Evaluation
> Production Architectures
If you're comfortable with agent basics and looking to understand the next layer this is a great read 👇
@weightsbiases.bsky.social on the Agent2Agent (A2A) protocol.
The blog talks talks about how it works, its core components and how to use it with CrewAI and @langchain.bsky.social.
Blog below 👇
@weightsbiases.bsky.social on the Agent2Agent (A2A) protocol.
The blog talks talks about how it works, its core components and how to use it with CrewAI and @langchain.bsky.social.
Blog below 👇
- Up to 3.6x speedup vs HF+FA2
- ~50% memory reduction
- Now includes Sequence Parallelism, Cut-Cross Entropy & Liger kernels
Plus you can fine-tune Gemma-3 27B with just 21.6GB VRAM!
- Up to 3.6x speedup vs HF+FA2
- ~50% memory reduction
- Now includes Sequence Parallelism, Cut-Cross Entropy & Liger kernels
Plus you can fine-tune Gemma-3 27B with just 21.6GB VRAM!
↔️Offers a wide range of sizes, from 0.6B to 235B.
🧠 Includes MoE variants.
🎛️ Hybrid Thinking Control with `enable_thinking=True`.
🌍 Multilingual Support: with support for 119 languages and dialects.
↔️Offers a wide range of sizes, from 0.6B to 235B.
🧠 Includes MoE variants.
🎛️ Hybrid Thinking Control with `enable_thinking=True`.
🌍 Multilingual Support: with support for 119 languages and dialects.
⚡️High quality for best cost
🧩 Ability to dynamically think on complex tasks
🎛️ Configurable thinking budget (decide how many tokens to allocate to "thinking".
⚡️High quality for best cost
🧩 Ability to dynamically think on complex tasks
🎛️ Configurable thinking budget (decide how many tokens to allocate to "thinking".
Details & steps on the blog ⤵️
Details & steps on the blog ⤵️
On to the next.
On to the next.
google.github.io/adk-docs/get...
google.github.io/adk-docs/get...
Hereby an updated guide that helps you get started with Gemma 3 and
@ollama
on Colab for a fast start. Based on one of our most popular 2024 posts. ⤵️
Hereby an updated guide that helps you get started with Gemma 3 and
@ollama
on Colab for a fast start. Based on one of our most popular 2024 posts. ⤵️
Code below⤵️
Code below⤵️
You can find it using the model ID gemini-2.5-pro-exp-03-25 through both the UI and SDK.
See below how to get started ⤵️
You can find it using the model ID gemini-2.5-pro-exp-03-25 through both the UI and SDK.
See below how to get started ⤵️
arxiv.org/abs/2503.19786
arxiv.org/abs/2503.19786
Notebooks to get you started below ⤵️
Notebooks to get you started below ⤵️
🧮Enhanced math, science, and reasoning
🖥️Significantly improved coding capabilities
📏1 million context window
It's available on Google AI Studio now, with Vertex AI coming soon. Code below to get you started ⤵️
🧮Enhanced math, science, and reasoning
🖥️Significantly improved coding capabilities
📏1 million context window
It's available on Google AI Studio now, with Vertex AI coming soon. Code below to get you started ⤵️
Below you'll find docs that include fine-tuning, sharding, LoRA, PEFT, multimodality and tokenization in LLMs.
Below you'll find docs that include fine-tuning, sharding, LoRA, PEFT, multimodality and tokenization in LLMs.
The guide below is a great starting point to learn more about evaluation. ⤵️
The guide below is a great starting point to learn more about evaluation. ⤵️