DeepMind has launched Gemini 3.1 Flash-Lite, its fastest and most cost-effective AI model to date, offering developers a tool for high-volume AI applications. The model's enhanced speed and adjustable 'thinking levels' aim to accelerate the development of scalable AI-powered products and lower entry barriers for complex problem-solving.
The release of Gemini 3.1 Flash-Lite provides developers with a powerful, low-cost tool for building scalable AI applications. Its enhanced speed and efficiency can accelerate the development of real-time services, such as content moderation and interactive user interfaces, while its adjustable reasoning depth offers new flexibility for complex problem-solving. This could lower the barrier to entry for creating sophisticated AI-powered products.
Gemini 3.1 Flash-Lite is DeepMind's fastest and most cost-effective AI model to date.
It offers a 2.5x faster time-to-first-token and a 45% higher output speed compared to the previous version.
The model includes adjustable 'thinking levels' to optimize reasoning for different tasks.
The model includes adjustable 'thinking levels' to optimize reasoning for different tasks.
It is available in preview via the Gemini API in Google AI Studio and on Vertex AI for enterprise use.
Sign in to save notes on signals.
Sign In