About Me
About 630 wordsAbout 2 min
2025-03-28
Hi, I'm Cedric β but most people know me as cedi.
I'm a Senior Site Reliability Engineering and Tech Lead at Microsoft Azure, working on distributed systems, chaos engineering, and platform resilience at scale. If it's complex, distributed, and needs to stay up β I'm into it.
I specialize in building and maintaining large-scale distributed systems, driving reliability, and leading technical initiatives to improve platform resilience.
π§ What I Do
- Building reliable, large-scale systems with a focus on resilience, SLOs, and automation
- Leading teams and setting technical direction in high-stakes, high-scale environments
- Designing chaos experiments, improving release workflows, and modernizing infrastructure
- Evangelizing good SRE practices through talks, docs, and community work
π οΈ Community & Chaos
I regularly help with infrastructure, planning, and logistics for events like the Chaos Communication Congress. During the pandemic, I helped build:
- jitsi.rocks - A collection of Jitsi Video Servers to stay connected
- Open Infrastructure - A collective of people building desperately needed infrastructure for education institutions to keep classes going
- The Kubernetes stack powering the rC3 - NOW HERE virtual workd
π§ͺ What I Tinker With
- Home lab with Raspberry Pi K3s cluster, CEPH storage, and a Stratum 1 NTP/PTP time server
- Cluster API managed cloud Kubernetes cluster running a full Grafana LGTM Stack
- Kernel recompilation just for fun (and for weird hardware drivers)
- Low-level distributed systems algorithms to explore gossip and consensus protocols
πΈ When Not Writing YAML
Iβm also a hobbyist analog photographer with a small collection of 35mm and medium format cameras (Leica M6, Hasselblad 500 c/m, Canon A1). I develop film at home and have a tiny darkroom with a 35mm enlarger.
π‘ Things I Believe in
- Be excellent to each other π€
- Focus on fundamentals > chasing hype
- Alert on symptoms, not vitals
- Incidents are opportunities to learn
- There is no single "root cause"
- How Complex Systems Fail is required reading
- Your beloved system architecture exists mostly in your head and
behavesfails differently than you'd expect. (See the Above the line/below the line framework)
β Letβs Connect
- Technical Leadership & Mentorship
- Key Stakeholder Management
- Staff Training
- Distributed Systems & Low-Level Systems Programming
- Chaos Engineering & Resilience Testing
- CI/CD & Deployment Orchestration
- Infrastructure as Code
- Reliability Engineering
- Software Engineering
- Cloud Computing & Cloud-Native Platforms
- Automation & Infrastructure as Code
- Distributed Systems Architecture
- Site Reliability Engineering (SRE)
- Chaos Engineering & Resilience Testing
- Kubernetes & Container Orchestration
- Observability & Monitoring (Metrics, Logging, Tracing)
- Low-Level Systems Programming
- Networking & Security
- DevOps & CI/CD Pipelines
- Technical Leadership & Mentorship
- Open Source Contributions & Cloud-Native Advocacy
- Public Speaking & Knowledge Sharing
- Home Lab Experimentation (Kubernetes, IoT, LoRaWAN, ESP32)
- Container Orchestration
- Kubernetes
- Docker / Docker-Compose
- Kubernetes
- GitOps with ArgoCD
- Cilium CNI
- K3s
- ClusterAPI
- Observability
- Grafana LGTM Stack
- Loki (Self-hosted in K8s)
- Grafana Dashboard (Self-hosted in K8s)
- Tempo (Self-hosted in K8s)
- Mimir (Self-hosted in K8s)
- Service Level Indicators & Service Level Objectives
- Grafana LGTM Stack
- CI/CD
- GitLab CI/CD
- GitHub Actions
- Project Management Tools
- JIRA
- Confluence
- Azure DevOps
- Azure Wiki
- DocFX
- Hugo
- VuePress
- Chaos Engineering
- Chaos Mesh (for K8s)
- Azure Chaos Studio (for Azure infrastructure)
- Infrastructure as Code
- Terraform
- Azure Resource Manager (ARM)
- Bicep
- Ansible
- Programming Languages
- Go
- Rust
- C++
- Python
- TypeScript
- BASH
- Networking
- Hardware
- Arista EOS
- HPE / Aruba
- Protocols
- IPsec IKEv2
- BGP
- Wireguard
- Technologies
- Tailscale
- mininet
- Hardware