Tech giant Google has unveiled its newest multimodal Large Language Model (LLM) called Gemini Flash. The announcement came during the recently concluded Google I/O, the annual developer conference organized by Google.
“Today, we’re introducing Gemini 1.5 Flash: a model that’s lighter-weight than 1.5 Pro, and designed to be fast and efficient to serve at scale”, stated Demis Hassabis CEO and Co-Founder of Google DeepMind. He goes on to explain that Flash is “optimized for high-volume, high-frequency tasks at scale”. Although this new model is a comparatively lighter weight model, it was still trained using the Gemini 1.5 pro model.
See Related: Google Launches Its Largest And Most Capable AI Model Yet – Google Gemini
Gemini Flash has been noted for its performance in summarization, chat applications, image and video captioning, data extraction from long documents and tables. The context window for the new model has also increased up to 1 million. This means the model can process one hour of video, 11 hours of audio, codebases with more than 30,000 lines of code, or over 700,000 words.
Gemini Flash is accessible for public preview in more than 200 regions across the globe. Currently, the model is available in 2 price plans. The “Free of charge” plan has a limit of 15 requests per minute (RPM) and 1,500 requests per day (RPD). The “pay-as-you-go” plan will cost users $0.35 to $0.70 per 1 million input token and $1.05 to $2.10 per 1 million output token. The paid version allows 360 RPM and 10,000 RPD.