While building a feature store to handle the massive growth of our machine-learning (“ML”) platform, we learned that using a mix of different databases can yield significant gains in efficiency and operational simplicity.
Tag Archives: redis
How We Applied Client-Side Caching to Improve Feature Store Performance by 70%
At DoorDash, we make millions of predictions every second to power machine learning applications to enhance our search, recommendation, logistics, and fraud areas, and scaling these complex systems along with our feature store is continually a challenge.
Building a Gigascale ML Feature Store with Redis, Binary Serialization, String Hashing, and Compression
When a company with millions of consumers such as DoorDash builds machine learning (ML) models, the amount of feature data can grow to billions of records with millions actively retrieved during model inference under low latency constraints.
Speeding Up Redis with Compression
One of challenges we face almost everyday is to keep our API latency low.