March 26, 2024
Container orchestration platforms, like Kubernetes, were, from the beginning, an ideal solution for microservice architectures running a lot of stateless services. This was also the case for Datadog, which is run on dozens of self-managed Kubernetes clusters in a multi-cloud environment, adding up to hundreds of thousands of pods. But what about stateful applications? What are the best practices to run and scale those without losing data?
The team at Datadog owning our Kafka clusters has been running big business critical storage workloads on our Kubernetes clusters for a long time. Over the years, they have gained experience on how to run this type of workload at scale, and have created tooling around it.
The team at Datadog owning our Postgres databases, on the other hand, is currently working on the transition to move their workloads from managed cloud instances to Kubernetes.
In this live session from the Datadog London Summit, Ara Pulido, Staff Developer Advocate, will chat with Martin Dickson, Senior Software Engineer in the Datadog Kafka team, and Edward Dale, Engineering Manager in the Postgres team at Datadog, about their experience, their tooling, and their stories (good and bad) on running stateful workloads in Kubernetes.
Datadog on Building Reliable Distributed Applications Using Temporal →
Datadog on OpenTelemetry →
Datadog on Secure Remote Updates →
Datadog on Data Science →
Datadog on Kubernetes Autoscaling →
Datadog on Kubernetes Node Management →
Datadog on Caching →
Datadog on Data Engineering Pipelines: Apache Spark at Scale →
Datadog on Site Reliability Engineering →
Datadog on Building an Event Storage System →
Datadog on gRPC →
Datadog on Gamedays →
Datadog on Chaos Engineering →
Datadog on Serverless →
Datadog on Kubernetes Monitoring →
Datadog on Software Delivery →
Datadog on Incident Management →
Datadog on RocksDB →
Datadog on Kafka →
Datadog on Kubernetes →