Version 2025-04 was officially released on April 10, 2025.
Below are the most important fixes and features introduced in this version.
- New The Lab
- The Lab is designed for defining, managing, and orchestrating autonomous Agents. It provides a standardized model for representing Agents, their capabilities, and their interactions within complex workflows. The core components of The Lab include:
- The Lab aims to meet the growing demand for intelligent, self-sufficient Agents capable of collaborating and solving complex problems with minimal human intervention. It offers a flexible and adaptable model, allowing for the creation and management of a diverse range of Agents, from co-pilots working alongside humans to fully automated Agents executing complex tasks. Implemented as a new platform of Globant Enterprise AI, The Lab supports the development of intelligent Agents that can work autonomously or in collaboration with humans and other Agents.
- New features in Flows
- Agent Integration Component: You can now directly integrate Agents created with the The Lab into a Flow. These Agents can be exposed through platforms like WhatsApp, Teams, or Slack.
- File Upload Support from WhatsApp: Flows now support receiving file attachments such as documents, images, audio, and video directly from WhatsApp interactions.
- Audio and Video Attachment in Web Chat: The Web Chat component now allows users to attach audio and video files, enhancing the interaction experience.
- New Features for the Data Analyst Agent
- Reduced Configuration Requirements: The setup needed to enable the Assistant to respond to a wide range of questions has been minimized.
- Enhanced Analysis Module: An additional analysis module has been incorporated to complement the responses with relevant business conclusions and interpretations of the obtained data.
- New metrics to track processed tokens.
- LLMs:
- New Gemini 2.5 Pro (via providers Vertex AI and Gemini): Gemini 2.5 is Google’s latest reasoning model, engineered to tackle increasingly complex challenges. This model is designed for tasks that demand advanced analytical thinking and robust problem-solving capabilities. More details at LLM API.
- Migration to Gemini 2.0 series (Vertex AI): Based on recommendations from Vertex AI, we have migrated from the legacy Gemini 1.0 and 1.5 models to the more advanced Gemini 2.0 series, offering improved performance, scalability, and integration capabilities. For comprehensive information, please refer to the Deprecated Models section.
- New Azure OpenAI models (o1, o1-mini and o3-mini): We have expanded our model availability by introducing these models via Azure, providing the same high-quality capabilities as those offered through the OpenAI provider.
- DeepSeek-R1 via AWS Bedrock: Recently added through a Serverless API, DeepSeek-R1 offers reliable inference with a substantial 128K token context window and up to 32K maximum output tokens.
- OpenAI's new models with built-in web search tool: These specialized models integrate web search capabilities directly into the Chat Completions API, enabling them to both interpret and execute search queries in real time.
- openai/gpt-4o-search-preview
- openai/gpt-4o-mini-search-preview
- New models - Beta only:
- gemini/gemma-3-27b-it: Gemma is a family of lightweight, state-of-the-art open models from Google, built from the same research and technology used to create the Gemini models. Gemma 3 has a large, 128K context window, and multilingual support in over 140 languages.
- SambaNova:
- DeepSeek-R1: This Provider offers the fastest performance for running DeepSeek, processing up to 198 tokens per second per user, with a 16K token context window. The model is hosted in US data centers; privately and securely.
- DeepSeek-V3-0324: This model significantly outperforms its predecessor with enhanced reasoning benchmarks, improved code executability and refined web aesthetics, and superior Chinese writing aligned with the R1 style. It also offers better multi-turn interactive rewriting, translation quality, detailed report analysis, and more accurate function calling.
- QwQ-32B: SambaNova provides access to QwQ-32B-Preview, the best open source test-time compute model released by Alibaba.
- Llama 4 collection by Meta:
- Llama 4 Scout: A 17B-parameter multimodal MoE model with 16 experts that excels in text and image understanding. The model is currently in beta and is available via providers Nvidia, Groq and SambaNova. Via Groq, it supports a 128k tokens context window with fast inference at 460 tokens/sec, while SambaNova Cloud runs at 697 tokens/second/user.
- Llama 4 Maverick: Available via Nvidia, this 17-billion-parameter model featuring 128 experts supports a 32k-token context window.