My main use case for Redis is caching frequently accessed data to improve performance and reduce database load. For example, I cache API responses and user-related data so that repeated requests can be served quickly without hitting the database every time. I use TTL to automatically expire stale data and ensure caching freshness. In some cases, I also use Redis for session management and handling short-lived data efficiently. I have used Redis for session management in a back-end system, where the main idea was to store user session data in Redis instead of keeping it in memory on a single server, which helps me scale across multiple instances. When a user logs in, we generate a session ID or token and store session-related data like user ID and metadata in Redis, and this session is associated with a TTL. It automatically expires after a certain period of time or after a certain time of inactivity. On each request, the session ID is validated by fetching data from Redis, which is very fast due to its in-memory nature, ensuring low latency and allowing us to handle the highest traffic efficiently. This approach helps us achieve horizontal scalability and avoids issues concerning session stickiness. Additionally, we ensure security by expiring inactive sessions or occasionally refreshing TTL for active users. Apart from caching and session management, I worked on interesting challenges using Redis, particularly around caching consistency and handling stale data. Initially, we faced issues where cached data would become outdated after database updates, and to solve this, we implemented a cache-aside strategy where we explicitly invalidated or updated the cache whenever the underlying data changed. Another scenario was handling cache misses during high traffic to avoid multiple requests hitting the database simultaneously, where we introduced techniques such as setting approaches, TTLs, and in some cases, using locking to ensure only one request rebuilds the cache. We also tuned invocation policies and memory usage to ensure Redis remains performant under load. These experiences helped me understand how to use Redis not just as a cache, but as a critical component in system performance and scalability. For maintaining the high traffic system, we also explored using Redis for rate limiting and short-lived counters, which further reduced our load on our core system.
Database Admin and Architect at D-EDGE Hospitality Solutions
Real User
Top 20
Jul 10, 2025
Redis is used for a part of a booking engine for travel, specifically for the front part to get some sessions and information about the sessions. If a customer or user is using the sites in different parts, we use Redis to get this information in cache.
Senior FullStack Developer/Engineer/Architect at Capitbrok
Real User
Apr 4, 2025
My primary use case for Redis has been leveraging it as a high-performance, in-memory data store to support real-time features in web applications. For instance, I use Redis to manage live chat systems and pub/sub patterns for real-time notifications, ensuring low latency and fast data access. Additionally, I’ve implemented Redis for caching database query results and session data, significantly improving performance and scalability under heavy load. In some microservice environments, I’ve also used Redis to support distributed locking and rate limiting, enhancing overall system reliability and throughput.
We use Redis ( /products/redis-reviews ) for several purposes, including ranking, counting, saving, sharing, caching, and setting time-to-live notifications. These functionalities are employed across various AI projects and in data processing tools, where Redis ( /products/redis-reviews ) helps with the ongoing data pipeline process.
I use Redis as a tool in building projects, specifically for in-memory caching. My backend API uses Redis to cache information retrieved from the database.
My primary use case for Redis is to enhance the performance of our web applications by using it as a caching layer. By caching frequently accessed data, we reduce the load on our primary databases, resulting in faster data retrieval and a more responsive user experience. We also use Redis for session storage, managing user sessions in a stateless manner, which ensures quick access to session data, supporting high-traffic scenarios without compromising performance. Additionally, Redis handles real-time analytics and leaderboard features, providing fast and efficient data processing capabilities. For real-time notifications and updates, we utilize Redis' Pub/Sub messaging feature. This facilitates real-time communication and synchronization between our services. Our Redis setup includes replication for high availability, persistence for data durability, and clustering for scalability. This integration of Redis in our environment has significantly boosted the performance, scalability, and reliability of our applications, making it an essential component of our infrastructure.
Software Engineer at a tech vendor with 5,001-10,000 employees
Real User
Top 5
Sep 5, 2024
I've used Redis mainly to improve application performance through caching. I've also used it as a small project broker and an external lock mechanism. I used it in previous companies.
I use Redis for data deduplication or identifying duplicate records I receive. I have a very basic usage with Redis, where I insert a key value record.
I'm the only user utilizing Redis in my company, as my platform is relatively new and has around fifty users. However, in my previous role at Profit Clicks, we had approximately ten thousand users. We utilized a combination of databases, including Redis, MongoDB, and MariaDB. We needed to handle push notifications for our users, sending them three times a day, and managing a large volume of notifications efficiently was crucial. Redis played a role in managing this workload.
We use it primarily for real-time applications. In our web application, we added a feature where hundreds of people could play a quiz in real time. Instead of using traditional databases like SQL, we implemented Redis to make everything happen in real time – all those quick calculations, data hashing for easy retrieval, and so on. It was a live quiz feature, so Redis helped a lot. I also use Redis for caching and similar general use cases.
Software Engineer at a financial services firm with 10,001+ employees
Real User
Apr 6, 2023
In our company, Redis is used for read operations. When we receive data from a user via the Read API, we cache the response for two or three minutes to minimize API calls to the database and preserve resources within the DB. This practice also helps prevent bot-triggered dummy requests to the DB.
Software Development Engineer at a tech vendor with 1-10 employees
Real User
Feb 3, 2023
It was a classic caching use case. In my previous company, we were running MongoDB on the cloud, and we were running Redis for caching. We were just starting, so the data size was pretty small there.
Redis offers high-speed, in-memory storage, renowned for real-time performance. It supports quick data retrieval and is used commonly in applications like analytics and gaming.Renowned for real-time performance, Redis delivers high-speed in-memory storage, making it a favorite for applications needing quick data retrieval. Its diverse data structures and caching capabilities support a broad array of use cases, including analytics and gaming. Redis ensures robust scalability with master-slave...
My main use case for Redis is caching frequently accessed data to improve performance and reduce database load. For example, I cache API responses and user-related data so that repeated requests can be served quickly without hitting the database every time. I use TTL to automatically expire stale data and ensure caching freshness. In some cases, I also use Redis for session management and handling short-lived data efficiently. I have used Redis for session management in a back-end system, where the main idea was to store user session data in Redis instead of keeping it in memory on a single server, which helps me scale across multiple instances. When a user logs in, we generate a session ID or token and store session-related data like user ID and metadata in Redis, and this session is associated with a TTL. It automatically expires after a certain period of time or after a certain time of inactivity. On each request, the session ID is validated by fetching data from Redis, which is very fast due to its in-memory nature, ensuring low latency and allowing us to handle the highest traffic efficiently. This approach helps us achieve horizontal scalability and avoids issues concerning session stickiness. Additionally, we ensure security by expiring inactive sessions or occasionally refreshing TTL for active users. Apart from caching and session management, I worked on interesting challenges using Redis, particularly around caching consistency and handling stale data. Initially, we faced issues where cached data would become outdated after database updates, and to solve this, we implemented a cache-aside strategy where we explicitly invalidated or updated the cache whenever the underlying data changed. Another scenario was handling cache misses during high traffic to avoid multiple requests hitting the database simultaneously, where we introduced techniques such as setting approaches, TTLs, and in some cases, using locking to ensure only one request rebuilds the cache. We also tuned invocation policies and memory usage to ensure Redis remains performant under load. These experiences helped me understand how to use Redis not just as a cache, but as a critical component in system performance and scalability. For maintaining the high traffic system, we also explored using Redis for rate limiting and short-lived counters, which further reduced our load on our core system.
Redis is used for a part of a booking engine for travel, specifically for the front part to get some sessions and information about the sessions. If a customer or user is using the sites in different parts, we use Redis to get this information in cache.
My primary use case for Redis has been leveraging it as a high-performance, in-memory data store to support real-time features in web applications. For instance, I use Redis to manage live chat systems and pub/sub patterns for real-time notifications, ensuring low latency and fast data access. Additionally, I’ve implemented Redis for caching database query results and session data, significantly improving performance and scalability under heavy load. In some microservice environments, I’ve also used Redis to support distributed locking and rate limiting, enhancing overall system reliability and throughput.
We use Redis ( /products/redis-reviews ) for several purposes, including ranking, counting, saving, sharing, caching, and setting time-to-live notifications. These functionalities are employed across various AI projects and in data processing tools, where Redis ( /products/redis-reviews ) helps with the ongoing data pipeline process.
I use Redis as a tool in building projects, specifically for in-memory caching. My backend API uses Redis to cache information retrieved from the database.
My primary use case for Redis is to enhance the performance of our web applications by using it as a caching layer. By caching frequently accessed data, we reduce the load on our primary databases, resulting in faster data retrieval and a more responsive user experience. We also use Redis for session storage, managing user sessions in a stateless manner, which ensures quick access to session data, supporting high-traffic scenarios without compromising performance. Additionally, Redis handles real-time analytics and leaderboard features, providing fast and efficient data processing capabilities. For real-time notifications and updates, we utilize Redis' Pub/Sub messaging feature. This facilitates real-time communication and synchronization between our services. Our Redis setup includes replication for high availability, persistence for data durability, and clustering for scalability. This integration of Redis in our environment has significantly boosted the performance, scalability, and reliability of our applications, making it an essential component of our infrastructure.
I used Redis for a one-time implementation while working on a college project. I implemented it locally on a machine and used it for a vector search.
In the past, I used Redis as a cache, and currently, I'm using it as a messaging system.
I've used Redis mainly to improve application performance through caching. I've also used it as a small project broker and an external lock mechanism. I used it in previous companies.
I use Redis for data deduplication or identifying duplicate records I receive. I have a very basic usage with Redis, where I insert a key value record.
I'm the only user utilizing Redis in my company, as my platform is relatively new and has around fifty users. However, in my previous role at Profit Clicks, we had approximately ten thousand users. We utilized a combination of databases, including Redis, MongoDB, and MariaDB. We needed to handle push notifications for our users, sending them three times a day, and managing a large volume of notifications efficiently was crucial. Redis played a role in managing this workload.
Our organization's main use case for Redis is to store and load static content for our multi-client products.
We use it primarily for real-time applications. In our web application, we added a feature where hundreds of people could play a quiz in real time. Instead of using traditional databases like SQL, we implemented Redis to make everything happen in real time – all those quick calculations, data hashing for easy retrieval, and so on. It was a live quiz feature, so Redis helped a lot. I also use Redis for caching and similar general use cases.
Sometimes, Redis is used as a cache and sometimes as the main database.
I use the solution for saving user sessions and user data for different applications.
In our company, Redis is used for read operations. When we receive data from a user via the Read API, we cache the response for two or three minutes to minimize API calls to the database and preserve resources within the DB. This practice also helps prevent bot-triggered dummy requests to the DB.
It was a classic caching use case. In my previous company, we were running MongoDB on the cloud, and we were running Redis for caching. We were just starting, so the data size was pretty small there.