Google unveils Gemini Embedding 2 with Multimodal Input Support and MRL technology

Date:

Google has introduced Gemini Embedding 2, a new multimodal embedding model built on the Gemini architecture. The model is designed to process multiple types of data, including text, images, video, audio and documents, placing them in a single integrated space.

Unlike previous embedding models that focused primarily on text, Gemini Embedding 2 extends capabilities to cross-modal understanding, enabling developers to build AI systems that analyze and connect different forms of media.

The model supports over 100 languages ​​and can power artificial intelligence applications such as Regression Augmented Generation (RAG), semantic search, sentiment analysis, and large-scale data mining.

Gemini Embedding 2 leverages the multimodal capabilities of the Gemini architecture to generate embeddings on various data types. The model can handle interwoven multimodal data, allowing developers to combine multiple forms of data within a single request.

For example, apps can analyze text descriptions alongside images or videos, enabling AI systems to understand relationships between different media formats.

This approach helps developers work with complex datasets containing multiple types of content.

Gemini Embedding 2 multimodal input

Main features

Multimodal login support

Gemini Embedding 2 supports a wide range of input formats:

  • Text: Up to 8,192 input characters
  • Images: Up to six images per request (PNG and JPEG formats)
  • Video: Up to 120 seconds per input (MP4 and MOV)
  • Audio: Live audio processing without transcription
  • Documents: Supports PDF files up to six pages

Cross-modal multimodal inputs

The model can process multiple media types in a single request, enabling contextual understanding of all input such as image and text together.

This capability is particularly useful for applications that require cross-media analysis and contextual search.

Matryoshka Representation Learning (MRL)

Gemini Embedding 2 includes Matryoshka Representation Learning (MRL), which allows embedding vectors to scale to different dimensions.

The default inset size is 3072 dimensions, but developers can reduce the size depending on their storage and performance requirements.

Recommended insert dimensions include:

This flexibility allows developers to optimize infrastructure performance and costs.

AI capabilities and use cases

Google says Gemini Embedding 2 enables multimodal embedding in text, image, video and speech tasks, while introducing native audio processing.

The model is designed for several AI applications, including:

  • Return Augmented Generation (RAG)
  • Semantic search
  • Sentiment analysis
  • Data collection
  • Large-scale knowledge management systems

Embedding twins 2 standard

Availability

Gemini Embedding 2 is currently available in Public Preview via:

Developers can integrate the model with popular AI frameworks and vector database tools such as:

  • LangChain
  • LlamaIndex
  • Straw
  • weave
  • Qdrant
  • ChromaDB

The model can also be combined with vector search systems to enable advanced multimodal data processing and retrieval.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Share post:

Subscribe

spot_imgspot_img

Popular

More like this
Related

Join Us WhatsApp