Keynote: How Spotify Accidentally Deleted All its Kube Clusters with No User Impact - David Xia
HTML-код
- Опубликовано: 27 июл 2024
- Join us for Kubernetes Forums Seoul, Sydney, Bengaluru and Delhi - learn more at kubecon.io
Don't miss KubeCon + CloudNativeCon 2020 events in Amsterdam March 30 - April 2, Shanghai July 28-30 and Boston November 17-20! Learn more at kubecon.io. The conference features presentations from developers and end users of Kubernetes, Prometheus, Envoy, and all of the other CNCF-hosted projects
Keynote: How Spotify Accidentally Deleted All its Kube Clusters with No User Impact - David Xia, Infrastructure Engineer, Spotify
During Spotify's Kubernetes migration, David's team deleted most of their production Kubernetes clusters. Accidentally. Twice. With little to no user impact. David shares how they recovered and learned to operate many clusters automatically and safely.
In 2017, Spotify planned the migration of hundreds of teams, thousands of services, and tens of thousands of hosts to Google Kubernetes Engine (GKE). In the last half of 2018, Spotify migrated 50 teams and hundreds of services, including critical ones, onto multiple production clusters.
David describes what led to the cluster deletions and how they barely affected users. Since the postmortem, Spotify has minimized downtime and human error by declaratively defining clusters in code with Terraform, backing up and restoring clusters with Ark, and increasing scalability and availability by running many more clusters.
sched.co/MQbb Наука
I believe Spotify is company that appreciate the engineering team so much. Culture of Learning is new thing.
This is like one of those water cooler conversations you get to have from that seasoned architect at your work who has had enough interesting mistakes ! Also I feel isolating your bundles with your infra can actually help avoid these errs
This is super entertaining. Sorry you guys had to deal with a mangled tfstate file in production. It's a terrible rite of passage.
loved it, honest , great learning
2 Teams - Kubernetes [Cluster Operators and Cluster Users]
Wonderful test case even for a beginner like me. Good talk!
If the internal slack channel was "eerily quiet" it's probably because teams don't have enough alerting set up to notify them that their applications/services aren't running in production.
Go through your envs and protect those clusters from deletion now that you can!
let's do it before bad things happen 😂
Ouch this was really painful. I guess you were running -auto-approve or no manual review of the tf plan before application.
ya, Guess so...quite an ignorant and not recommended approach for Prod.
yes many times deleted k8s cluster accidentally...
:clappepe:
Patrick Marshall :pepego: