Gemini Embedding
Gemini Embedding’s first text model (gemini-embedding-001) is now generally available via the Gemini API and Gemini Enterprise Agent Platform, having held a top spot on the Massive Text Embedding Benchmark Multilingual leaderboard since its experimental launch in March, thanks to superior performance across retrieval, classification, and other embedding tasks compared to both legacy Google and external proprietary models. Exceptionally versatile, it supports over 100 languages with a 2,048‑token input limit and employs the Matryoshka Representation Learning (MRL) technique to let developers choose output dimensions of 3072, 153,6, or 768 for optimal quality, performance, and storage efficiency.
Learn more
Pinecone
The AI Knowledge Platform.
The Pinecone Database, Inference, and Assistant make building high-performance vector search apps easy. Developer-friendly, fully managed, and easily scalable without infrastructure hassles.
Once you have vector embeddings, manage and search through them in Pinecone to power semantic search, recommenders, and other applications that rely on relevant information retrieval.
Ultra-low query latency, even with billions of items. Give users a great experience. Live index updates when you add, edit, or delete data. Your data is ready right away. Combine vector search with metadata filters for more relevant and faster results.
Launch, use, and scale your vector search service with our easy API, without worrying about infrastructure or algorithms. We'll keep it running smoothly and securely.
Learn more
Universal Sentence Encoder
The Universal Sentence Encoder (USE) encodes text into high-dimensional vectors that can be utilized for tasks such as text classification, semantic similarity, and clustering. It offers two model variants: one based on the Transformer architecture and another on Deep Averaging Network (DAN), allowing a balance between accuracy and computational efficiency. The Transformer-based model captures context-sensitive embeddings by processing the entire input sequence simultaneously, while the DAN-based model computes embeddings by averaging word embeddings, followed by a feedforward neural network. These embeddings facilitate efficient semantic similarity calculations and enhance performance on downstream tasks with minimal supervised training data. The USE is accessible via TensorFlow Hub, enabling seamless integration into various applications.
Learn more
Cohere Embed
Cohere's Embed is a leading multimodal embedding platform designed to transform text, images, or a combination of both into high-quality vector representations. These embeddings are optimized for semantic search, retrieval-augmented generation, classification, clustering, and agentic AI applications. The latest model, embed-v4.0, supports mixed-modality inputs, allowing users to combine text and images into a single embedding. It offers Matryoshka embeddings with configurable dimensions of 256, 512, 1024, or 1536, enabling flexibility in balancing performance and resource usage. With a context length of up to 128,000 tokens, embed-v4.0 is well-suited for processing large documents and complex data structures. It also supports compressed embedding types, including float, int8, uint8, binary, and ubinary, facilitating efficient storage and faster retrieval in vector databases. Multilingual support spans over 100 languages, making it a versatile tool for global applications.
Learn more