Fully managed and integrated with Google Cloud, Azure, and AWS.
Build the fastest, most reliable GenAI apps with our advanced vector database.
Self-managed software with enterprise-grade compliance and reliability.
Synchronize data in near-real time to make data fast—without writing code.
In-memory database for caching & streaming.
The Redis Query Engine (RQE) is a powerful tool for executing complex search and query operations on structured, semi-structured, and unstructured data. Indexes are the backbone of this functionality, enabling fast and efficient data retrieval. Proper management of these indexes is essential for optimal performance, scalability, and resource utilization.
This guide outlines best practices for managing RQE indexes throughout their lifecycle. It provides recommendations on:
Indexes directly impact query speed and resource consumption. Poorly managed indexes can lead to increased memory usage, slower query times, and challenges in maintaining data consistency. By following the strategies outlined in this guide, you can:
Planning your indexes strategically requires understanding your application’s query patterns and tailoring indexes to match.
Begin by identifying the types of searches your application performs—such as full-text search, range queries, or geospatial lookups—and the fields involved.
Categorize fields based on their purpose: searchable fields (e.g., TEXT
for full-text searches), filterable fields (e.g., TAG
for exact match searches), and sortable fields (e.g., NUMERIC
for range queries or sorting).
Match field types to their intended use and avoid indexing fields that are rarely queried to conserve resources. Here's the list of index types:
TEXT
: use TEXT
for free-text searches and set weights if some fields are more important.TAG
: use TAG
for categorical data (e.g., product categories) that benefit from exact matching and filtering.NUMERIC
: use NUMERIC
for numeric ranges (e.g., prices, timestamps).GEO
: use GEO
for geospatial coordinates (e.g., latitude/longitude).GEOSHAPE
: use GEOSHAPE
to represent locations as points, but also to define shapes and query the interactions between points and shapes (e.g., to find all points that are contained within an enclosing shape).VECTOR
: use VECTOR
for high-dimensional similarity searches.See these pages for discussions and examples on how best to use these index types.
Next, simulate queries on a sample dataset to identify potential bottlenecks.
Use tools like FT.PROFILE
to analyze query execution and refine your schema if needed.
For example, assign weights to TEXT
fields for prioritizing results or use the PREFIX
option of FT.CREATE
to limit indexing to specific key patterns. Note that you can use multiple PREFIX
clauses when you create an index (see below)
After creating the index, validate its performance with real queries and monitor usage with the available tools:
FT.EXPLAIN
and FT.EXPLAINCLI
allow you to see how Redis Query Engine parses a given search query. FT.EXPLAIN
returns a structured breakdown of the query execution plan, while FT.EXPLAINCLI
presents a more readable, tree-like format for easier interpretation. These commands are useful for diagnosing query structure and ensuring it aligns with the intended logic.FT.INFO
provides detailed statistics about an index, including the number of indexed documents, memory usage, and configuration settings. It helps in monitoring index growth, assessing memory consumption, and verifying index structure to detect potential inefficiencies.FT.PROFILE
runs a query while capturing execution details, which helps to reveal query performance bottlenecks. It provides insights into processing time, key accesses, and filter application, making it a crucial tool for fine-tuning complex queries and optimizing search efficiency.Avoid over-indexing. Indexing every field increases memory usage and can slow down updates. Only index the fields that are essential for your planned queries.
FT.CREATE
command to define an index schema.TEXT
fields to prioritize certain fields in full-text search results.PREFIX
option to restrict indexing to keys with specific patterns.
Using multiple PREFIX clauses when creating an index allows you to index multiple key patterns under a single index. This is useful in several scenarios:
If your Redis database stores different types of entities under distinct key prefixes (e.g., user:123
, order:456
), a single index can cover both by specifying multiple prefixes. For example:
FT.CREATE my_index ON HASH PREFIX 2 "user:" "order:" SCHEMA name TEXT age NUMERIC status TAG
This approach enables searching across multiple entity types without needing separate indexes.
Instead of querying multiple indexes separately, you can search across related data structures using a single query. This is particularly helpful when data structures share common fields, such as searching both customer and vendor records under a unified contacts index.
Maintaining multiple indexes for similar data types can be inefficient in terms of memory and query performance. By consolidating data under one index with multiple prefixes, you reduce overhead while still allowing for distinct key organization.
If your data model evolves and new key patterns are introduced, using multiple PREFIX
clauses from the start ensures future compatibility without requiring a full reindexing.
ON HASH
or ON JSON
options to match the data structure.Index aliases act as abstracted names for the underlying indexes, enabling applications to reference the alias instead of the actual index name. This approach simplifies schema updates and index management.
There are several use cases for index aliasing, including:
products_v1
initially and later to products_v2
when the schema evolves.Best practices for aliasing:
users_current
or orders_live
).tenant1_products
and tenant2_products
to different indexes for isolated query performance.Tools for managing aliases:
FT.ALIASADD
my_alias my_index
FT.ALIASUPDATE
my_alias new_index
FT.ALIASDEL
my_alias
Monitoring and troubleshooting aliases:
FT.INFO
command to check which aliases are associated with an index.Use the FT.INFO
command to monitor the num_docs
and indexing
fields, to check that all expected documents are indexed.
FT.INFO my_new_index
Validate data with sample queries to ensure proper indexing:
FT.SEARCH my_new_index "*"
Use FT.PROFILE
to analyze query plans and validate performance:
FT.PROFILE my_new_index SEARCH QUERY "your_query"
Implement scripts to periodically verify document counts and query results. For example, in Python:
import re
def check_index_readiness(index_name, expected_docs):
r = redis.StrictRedis(host='localhost', port=6379, decode_responses=True)
info = r.execute_command('FT.INFO', index_name)
num_docs = int(info[info.index('num_docs') + 1])
return num_docs >= expected_d
if check_index_readiness('my_new_index', 100000):
print("Index is fully populated!")
else:
print("Index is still populating...")
FT.PROFILE
command to analyze query performance and identify bottlenecks.INFO
memory
and FT.INFO
commands to detect growth patterns and optimize resource allocation.Use FT.ALTER
when you need to add new fields to an existing index without rebuilding it, minimizing downtime and resource usage. However, FT.ALTER
cannot remove or modify existing fields, limiting its flexibility.
Use index aliasing when making schema changes that require reindexing, such as modifying field types or removing fields. In this case, create a new index with the updated schema, populate it, and then use FT.ALIASUPDATE
to seamlessly switch queries to the new index without disrupting application functionality.
FT.DROPINDEX
command to remove unused indexes and free up memory. Be cautious with the DD
(Delete Documents) flag to avoid unintended data deletion.