Amazon OpenSearch Service as a Vector Database

Build vector-driven search and enterprise AI applications with a scalable, secure, and high-performance vector database.

Overview

The vector engine for OpenSearch Service offers a scalable, secure, and high-performance vector database for modern generative AI applications. Effortlessly store and search billions of high-dimensional vectors in milliseconds using advanced k-Nearest Neighbors (k-NN) and Approximate Nearest Neighbors (ANN) algorithms with Hierarchical Navigable Small World (HNSW) and Inverted File (IVF) implementations. Seamlessly combine vector embeddings with text-based keywords for vector-driven search capabilities (semantic, multi-modal, conversational, and others), recommendation systems, chatbots, and other modern generative AI applications. Build Retrieval-Augmented Generation (RAG) applications that securely connect foundation models (FMs) to your business data for accurate, context-aware responses—eliminating the need for fine-tuning or retraining. Optimize costs with intelligent data lifecycle management while maintaining fast query performance across all storage tiers. The vector database is available in a fully managed or severless configuration.

Vector Engine for Amazon OpenSearch Serverless (0:30)

Benefits

Seamlessly combine vector embeddings with text-based keyword queries in a single search request and use advanced nearest neighbor algorithms like ANN (across HNSW and IVF) and exact k-NN vector search with auto-scaling to deliver low-latency similarity searches across billions of vectors. This reduces system complexity, eliminates the need for multiple systems, and accelerates time-to-market for AI-powered applications like vector-driven search (hybrid, semantic, multi-modal, conversational, and others), recommendation systems, and AI chatbots, and other modern generative AI applications.

Scale to billions of high-dimensional vectors while optimizing storage costs with disk-based vector storage and intelligent data lifecycle management. OpenSearch Service simplifies vector database operations, offering an easy-to-use interface for both fully managed and serverless configurations. You can choose between precise control with managed clusters or automatic resource optimization with serverless to efficiently scale your vector workloads without incurring unnecessary costs. Both options ensure fast query responses across all storage levels while leveraging intelligent data lifecycle management to optimize costs as your workloads grow. OpenSearch Service's intuitive console and APIs make it straightforward to deploy, manage, and scale your vector database, reducing operational complexity.

Add, update, or delete vector embeddings in real-time without re-indexing or impacting query performance. This capability ensures that AI models and search applications remain responsive to dynamic data changes, making it ideal for use cases like e-commerce personalization or anomaly detection where data evolves frequently.

Amazon OpenSearch Service integrates with AWS services and third-party AI platforms to support modern generative AI applications. zero-ETL integration with Amazon DynamoDB and Amazon DocumentDB enables you to enhance your generative AI applications with vector search across operational data without building complex pipelines. Native two-way integration with Amazon Bedrock streamlines generative AI workflows, allowing you to connect foundation models to your knowledge base for efficient embedding generation and retrieval-augmented generation (RAG) applications. OpenSearch Service is the AWS recommended vector database for Amazon Bedrock. Developers can harness the power of Amazon SageMaker for model training and deployment or connect effortlessly to Amazon Titan or third-party models like OpenAI, Cohere, DeepSeek, and others through pre-built connectors. This enables secure, efficient, and scalable development while maximizing the value of your existing data and infrastructure investments.

A fully managed service that manages OpenSearch delivering enterprise reliability while leveraging open-source innovation. The global open-source community actively contributes to and enhances OpenSearch (now part of the Linux Foundation), driving continuous advancement while the managed service eliminates infrastructure management overhead. This approach provides high availability (99.99% SLA), automated scaling, patching, and updates, plus the flexibility and vendor neutrality of Apache 2.0 licensed technology. The open-source community also helps guide the project's direction, ensuring continuous innovation that benefits all users.

Use cases

Enhance search experiences by combining traditional keyword search with vector similarity for improved relevance. Support natural language understanding, multi-modal queries (text, images, audio), and hybrid search capabilities to deliver contextually relevant results across diverse content types.
Power personalized recommendations at scale, using vector similarity to match user preferences across billions of items, delivering relevant suggestions in near real time.
Build trustworthy AI chatbots, assistants, and applications by connecting foundation models to your business data, enabling accurate, context-aware responses and task execution. Eliminate hallucinations and improve accuracy through vector-based information retrieval while maintaining fast response times for both simple queries and complex interactions.
Identify patterns and anomalies at scale by comparing vector similarities across large datasets, enabling real-time detection of potential fraud, counterfeits, or suspicious activities.

Customers and partners

riskCanvas customer review

riskCanvas is a subsidiary of Genpact. It is a SaaS product offering for a financial crime compliance solution that uses cutting-edge big data, automation, and machine learning technologies to deliver compliance, efficiency, and automation to its clients.

"riskCanvas directly integrates with the vector engine for Amazon OpenSearch Serverless, allowing us to expose our existing client operational data through AWS’s generative AI capabilities. This is a game changer, as we can now leverage summarization to accelerate the analysis of investigations, author seed narratives of financial crimes reporting, and make recommendations on escalations—all while using real data kept contained within the riskCanvas secure enclave. With the vector engine, we are reducing handle time across financial crimes use cases, improving consistency of narratives with fewer errors, driving higher efficacy via straight-through processing, and shifting human involvement to deeper analysis."

Ryan Skousen, Chief Technology Officer (riskCanvas) and Vice President of Technology, Genpact Financial Crimes 

riskCanvas logo

Academia customer review

Academia is a platform for sharing academic research. The mission of Adademia.edu is to accelerate the world's research.

“Amazon OpenSearch Service advances Academia’s mission of accelerating the world’s research by enabling us to efficiently index and search through millions of vectors and find the most relevant academic papers to recommend to our users. Switching to Amazon OpenSearch Service has driven a 20% increase in engagement from our users with our content recommendations compared to our previous recommendation solutions. "

Bob Tucker, Director of Engineering, Academia

Academia logo

Intuit customer review

Intuit Inc. is a global financial technology platform that helps consumers and small businesses prosper by delivering financial management, compliance, and marketing products and services.

"Our platform team worked closely with AWS to build high-level capabilities to efficiently store, manage, and query vector embeddings produced by state-of-the-art ML models, unlocking new possibilities for natural language processing applications and services. This solution is now the default store for all vector needs across Intuit - thanks to Amazon OpenSearch Service. We are excited to expand our OpenSearch-based vector database adoption to tackle new and upcoming use cases in the coming months."

Achal Kumar, Director of Data Capabilities, Intuit

Intuit logo