- RedisAI

High-performance runtime for AI serving.

Redis AI

Co-developed with Redis Labs, RedisAI breeds the fastest DL/ML runtimes with the power of Redis.

How it works
- A Redis module for DL/ML.

RedisAI is a module bringing tensors to the Redis database and the ability to run computations on them. This brings the operational flexibility and the performance of Redis into the landscape of AI serving. By keeping the data local to the computation, roundtrip is zero, inference happens right where the data resides. RedisAI supports HA, Cluster and Redis Enterprise.

Redis AI How It Works
Redis AI How It Works
- Multi-backend, all sizes.

RedisAI supports the most popular inference engines: PyTorch, TensorFlow, ONNX and ONNX-ML. Whatever the framework is, RedisAI exposes a uniform API to developers. Changing the backend has an extremely limited impact on a codebase. RedisAI fits all scales, from edge to web-scale, from small ARM devices to multi-node clusters on the cloud.

- Composable computations.

RedisAI is not your usual serving solution. Build DAGs of RedisAI commands for maximum performance. Or programmatically compose computations using TorchScript, managing ensambles, model aliases, A/B testing, monitoring metrics, uncertainty estimation. Optimize calls with automated batching, leverage clustering by running commands on nodes or replicas. Compose away.

Redis AI How It Works

Co-developed with Redis Labs, RedisAI breeds the fastest DL/ML runtimes with the power of Redis.

How it works
Redis AI How It Works
- A Redis module for DL/ML.

RedisAI is a module bringing tensors to the Redis database and the ability to run computations on them. This brings the operational flexibility and the performance of Redis into the landscape of AI serving. By keeping the data local to the computation, roundtrip is zero, inference happens right where the data resides. RedisAI supports HA, Cluster and Redis Enterprise.

Redis AI How It Works
- Multi-backend, all sizes.

RedisAI supports the most popular inference engines: PyTorch, TensorFlow, ONNX and ONNX-ML. Whatever the framework is, RedisAI exposes a uniform API to developers. Changing the backend has an extremely limited impact on a codebase. RedisAI fits all scales, from edge to web-scale, from small ARM devices to multi-node clusters on the cloud.

Redis AI How It Works
- Composable computations.

RedisAI is not your usual serving solution. Build DAGs of RedisAI commands for maximum performance. Or programmatically compose computations using TorchScript, managing ensambles, model aliases, A/B testing, monitoring metrics, uncertainty estimation. Optimize calls with automated batching, leverage clustering by running commands on nodes or replicas. Compose away.

RedisAI provides Tensors as a new data structure and allows Redis to run models at the fastest rate on compute hardware.

Benefits.
Computing

Deploy models from TensorFlow, PyTorch, ONNX, ONNX-ML. Run TorchScript natively for model composition.

Data locality

Dramatically accelerate & simplify AI serving by bringing model execution to where the data lives.

Clustering

Leverage Redis master-replica and cluster capabilities to design your compute architecture.

Lifecycle integration

Use RedisAI as a deployment target for your lifecycle management framework of choice.