Fully managed and integrated with Google Cloud, Azure, and AWS.
Build the fastest, most reliable GenAI apps with our advanced vector database.
Self-managed software with enterprise-grade compliance and reliability.
Synchronize data in near-real time to make data fast—without writing code.
In-memory database for caching & streaming.
An overview of Redis for AI documentation
Redis stores and indexes vector embeddings that semantically represent unstructured data including text passages, images, videos, or audio. Store vectors and the associated metadata within hashes or JSON documents for indexing and querying.
Vector | RAG | RedisVL |
---|---|---|
![]() |
![]() |
![]() |
This page organized into a few sections depending on what you’re trying to do:
FLAT
and HNSW
vector index types.Learn to perform vector search and use gateways and semantic caching in your AI/ML projects.
Search | LLM memory | Semantic caching | Semantic routing | AI Gateways |
---|---|---|---|---|
![]() |
![]() |
![]() |
![]() |
![]() |
Quickstarts or recipes are useful when you are trying to build specific functionality. For example, you might want to do RAG with LangChain or set up LLM memory for your AI agent. Get started with the following Redis Python notebooks.
Vector search retrieves results based on the similarity of high-dimensional numerical embeddings, while hybrid search combines this with traditional keyword or metadata-based filtering for more comprehensive results.
Retrieval Augmented Generation (aka RAG) is a technique to enhance the ability of an LLM to respond to user queries. The retrieval part of RAG is supported by a vector database, which can return semantically relevant results to a user’s query, serving as contextual information to augment the generative capabilities of an LLM.
AI agents can act autonomously to plan and execute tasks for the user.
LLMs are stateless. To maintain context within a conversation chat sessions must be stored and resent to the LLM. Redis manages the storage and retrieval of chat sessions to maintain context and conversational relevance.
An estimated 31% of LLM queries are potentially redundant. Redis enables semantic caching to help cut down on LLM costs quickly.
Routing is a simple and effective way of preventing misuses with your AI application or for creating branching logic between data sources etc.
Build a facial recognition system using the Facenet embedding model and RedisVL.
Need a deeper-dive through different use cases and topics?
See how we stack up against the competition.
See how leaders in the industry are building their RAG apps.