<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Reliability Enablers]]></title><description><![CDATA[Soon, focus will shift to the emerging intersection of resilient and intelligent systems. Because “five nines” won’t survive AI ignorance. Newsletter and podcast inside.]]></description><link>https://read.srepath.com</link><image><url>https://substackcdn.com/image/fetch/$s_!hjhf!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ee1dc2-77bf-4ffa-b056-f66dac8ad0d0_128x128.png</url><title>Reliability Enablers</title><link>https://read.srepath.com</link></image><generator>Substack</generator><lastBuildDate>Thu, 30 Apr 2026 07:01:55 GMT</lastBuildDate><atom:link href="https://read.srepath.com/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Ash P]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[srepath@substack.com]]></webMaster><itunes:owner><itunes:email><![CDATA[srepath@substack.com]]></itunes:email><itunes:name><![CDATA[Ash Patel]]></itunes:name></itunes:owner><itunes:author><![CDATA[Ash Patel]]></itunes:author><googleplay:owner><![CDATA[srepath@substack.com]]></googleplay:owner><googleplay:email><![CDATA[srepath@substack.com]]></googleplay:email><googleplay:author><![CDATA[Ash Patel]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[More telemetry makes reliability worse (until you fix the loop)]]></title><description><![CDATA[Every reliability engineer eventually learns the same painful truth: you can have a thousand dashboards showing you xyz and still miss the real signal.]]></description><link>https://read.srepath.com/p/more-telemetry-makes-reliability</link><guid isPermaLink="false">https://read.srepath.com/p/more-telemetry-makes-reliability</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 23 Dec 2025 15:05:15 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/4dc47b11-b51a-4875-b488-48a59c97282f_1280x869.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Every reliability engineer eventually learns the same painful truth: you can have a thousand dashboards showing you <em>xyz</em> and still miss the real signal.</p><p>This might feel like an insurmountable hurdle at first glance.</p><p>One of those &#8220;it is what it is&#8221; situations. After all:</p><p>The more data we collect &#8594; the more noise we face &#8594; the less trust we have in our alerts &#8594; the slower we respond &#8594; the more incidents worsen &#8594; the more data we collect to compensate.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!ZhJL!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!ZhJL!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png 424w, https://substackcdn.com/image/fetch/$s_!ZhJL!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png 848w, https://substackcdn.com/image/fetch/$s_!ZhJL!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png 1272w, https://substackcdn.com/image/fetch/$s_!ZhJL!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!ZhJL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png" width="907" height="547" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/b1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:547,&quot;width&quot;:907,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:72469,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://read.srepath.com/i/180281648?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!ZhJL!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png 424w, https://substackcdn.com/image/fetch/$s_!ZhJL!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png 848w, https://substackcdn.com/image/fetch/$s_!ZhJL!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png 1272w, https://substackcdn.com/image/fetch/$s_!ZhJL!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb1ec914f-0d24-46fa-b047-6d8bbbcf1813_907x547.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>In MIT&#8217;s world of System Dynamics, this noise amplification problem is what we&#8217;d call a <em>reinforcing loop</em>.</p><p>It&#8217;s a spiralling up of information overload as the loop continues to reinforce, or metaphorically snowball, onto itself. But here&#8217;s the thing&#8230;</p><p>Observability (o11y) isn&#8217;t just telemetry.</p><p>It&#8217;s also who interprets, triages, and learns from the telemetry: a <em>balancing loop</em>.</p><p>In a healthy system, every new signal that enters should trigger an equal and opposite stabilizing action, essentially a check-and-balance. That&#8217;s the balancing loop at work.</p><p>For example, when noise increases, teams should automatically slow alert creation or tighten signal thresholds to wait until trust recovers.</p><p>When signal quality improves, they can loosen up again.</p><p>Without that feedback control, the system loses balance, and the painful reinforcing loop that I mentioned earlier takes over.</p><p>If your team doesn&#8217;t trust the data, or worse, doesn&#8217;t have time to translate it, your observability system isn&#8217;t truly &#8220;seeing everything&#8221;.</p><p>That&#8217;s why engineers with Staff+ potential treat incident retros and observability reviews like a process tuning. They ask:</p><ul><li><p>Who sees which o11y signals, and when?</p></li><li><p>What incentives drive our attention to o11y signals?</p></li><li><p>Where does learning from outputs feed back into o11y design?</p></li></ul><p>Small interventions like taking the time to prune unhelpful alerts can have an outsized impact in the long run because they restore the balancing loop between data and actionability.</p><p>This should be your takeaway: reliability improves when observability helps people modify their impact from using the system, not just seeing the outputs of their services.</p>]]></content:encoded></item><item><title><![CDATA[Real DevOps happens between commits, not pipelines]]></title><description><![CDATA[Your CI/CD runs faster than your feedback.]]></description><link>https://read.srepath.com/p/real-devops-happens-between-commits</link><guid isPermaLink="false">https://read.srepath.com/p/real-devops-happens-between-commits</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 16 Dec 2025 15:01:11 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/7492caa8-4242-48e0-916b-50fd4ea7ac3c_1280x848.webp" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Your CI/CD runs faster than your feedback. That&#8217;s not DevOps &#8212; that&#8217;s DevOops.</p><p>Most engineers still picture DevOps as a conveyor belt more than a (software development life)cycle. The ground-level reality of DevOps in most organizations is that code goes in, product comes out, and every new automation step makes the process run smoother and faster.</p><p>That&#8217;s the story most of us were sold and still work toward to this day.</p><p>&#8220;DevOps is efficiency.&#8221;</p><p>It&#8217;s understandable. That&#8217;s how DevOps got operationalised in most places &#8212; as tooling and velocity. But <strong>DevOps wasn&#8217;t conceived to </strong><em><strong>only</strong></em><strong> make delivery faster.</strong></p><p>Patrick Debois, the &#8220;godfather of DevOps&#8221;, never framed DevOps with a singular focus on delivery speed. He framed it through the lens of <em>feedback.</em></p><p>The DevOps Handbook, which he co-authored with Gene Kim et al., talks about DevOps as flow, feedback, and continuous learning.</p><p>That&#8217;s system-dynamics thinking, even if nobody used that term at the time&#8230; and well, even now.</p><p>Debois cared about how software engineering could take advantage of this.</p><p>His thinking focused on:</p><ol><li><p>How fast signals travelled</p></li><li><p>How quickly the drift was corrected, and</p></li><li><p>How learning could compound improvements across teams</p></li></ol><p>Coming back to today, we should not merely be optimizing for delivery speed. We should also optimize for feedback speed.</p><p>If you zoom into any software delivery flow today, for the most part, the commits themselves aren&#8217;t the problem. In fact, they are fast. Pipelines are flowing fast. Rollouts are fast.</p><p><strong>What&#8217;s slow is everything </strong><em><strong>between</strong></em><strong> commits:</strong></p><ul><li><p>How long it takes to interpret an alert</p></li><li><p>How long it takes for an insight to become a design change</p></li><li><p>How fast one team&#8217;s learning propagates to another</p></li></ul><p>This &#8220;between commits&#8221; zone is the <strong>rate-limiting step for truly smooth DevOps.</strong></p><p>Here&#8217;s a composite story of the kinds I have heard over the last 3 years alone:</p><blockquote><p>During a large-scale outage, the alert fired instantly, but it took 45 minutes before anyone realised the signal was noise-shaped by a mis-configured dashboard. The root cause was understood the same day, but the design change sat unscheduled for eight weeks. Worse, another team triggered the <em>same</em> failure pattern three months later because the learning never left the postmortem.</p></blockquote><p>Feedback is the layer that determines whether DevOps creates stability or chaos.</p><p>In system-dynamics language, DevOps is a dance between two loops. The:</p><ul><li><p>balancing loop of continuous improvement (detect &#8594; understand &#8594; act) and</p></li><li><p>reinforcing loop of (deploy &#8594; drift &#8594; more deploys).</p></li></ul><p>This zone is where the balancing loop either outruns (stability) or falls behind (chaos) the reinforcing loop. <em>Which is more common in your team or organization?</em></p><p>When your feedback travels slower than your commits, the reinforcing loop wins.</p><p>That&#8217;s when system reliability begins to decay quietly in the background.</p><p>But when your feedback travels <em>faster</em> than your commits, the balancing loop wins.</p><p>That&#8217;s when reliability earns its rightful place as a default system property. Really, it should not be seen as a sociocultural artifact like team effort or heroic effort.</p><p>Optimizing pipelines makes delivery <em>fast.</em> Optimizing feedback makes delivery <em>safe</em>. Doing both gives you real DevOps.</p><p>High-performing teams don&#8217;t ask, &#8220;How do we deploy more often?&#8221;</p><p>They ask, &#8220;How do we <em>learn</em> faster than we deploy?&#8221;</p>]]></content:encoded></item><item><title><![CDATA[Humans, the pesky side of system design]]></title><description><![CDATA[Most reliability engineers hear &#8216;system design&#8217; and picture diagrams wrangling load balancers, queues, and failover zones.]]></description><link>https://read.srepath.com/p/humans-the-pesky-side-of-system-design</link><guid isPermaLink="false">https://read.srepath.com/p/humans-the-pesky-side-of-system-design</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 09 Dec 2025 15:30:33 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/8a42c75e-1817-4622-9132-3d0ced00bd71_640x421.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Most reliability engineers hear &#8216;system design&#8217; and picture diagrams wrangling load balancers, queues, and failover zones.</p><p>Fair enough. That&#8217;s the world we were educated in.</p><p>But every design workshop I&#8217;ve sat in this year has ended with very human questions:</p><p><em>&#8220;Who&#8217;ll own this?&#8221;</em></p><p><em>&#8220;How do we know when to alert?&#8221;</em></p><p><em>&#8220;How do we ramp these 37 changes into our flow?&#8221;</em></p><p>Guess what? That&#8217;s system design, too. It&#8217;s just the part we never diagram.</p><p>You already know the cognitive load of infra churn&#8230;</p><p>AWS &#8594; Kubernetes &#8594; AI pipelines &#8594; whatever&#8217;s next.</p><p>Across every shift, the engineers who stay ahead are the ones who <strong>shape the system around the system,</strong> so its behaviors support sustainable reliability.</p><p>That&#8217;s not fluff &#8212; that&#8217;s the beginning of Staff+ thinking.</p><p>No one gets promoted for invisible work.</p><p>You advance by showing <em>evidence</em> of potential, and the human side of the system is where it shows early.</p><p>You already know the rituals that humans in your org get involved in. Most orgs run them &#8212; some lightly, some intensely, some religiously.</p><p>Some examples of rituals include postmortems/retrospectives, SLO writing sessions, and incident bridges with <em>all</em> involved parties.</p><p>But there&#8217;s a quiet crisis building up within these rituals.</p><p><strong>Reliability flatlines when rituals replace mechanisms.</strong></p><p>Rituals mean well with the intent to turn responses to events into regular action, so it makes sense that many teams try to improve reliability through rituals.</p><p>Over time, you end up with:</p><ul><li><p>more dashboards</p></li><li><p>more alerts</p></li><li><p>more retros</p></li><li><p>more action items</p></li><li><p>more processes</p></li><li><p>more OKRs</p></li><li><p>more monitoring &#8220;initiatives&#8221;</p></li></ul><p>These things <em>look</em> like progress. They feel responsible.</p><p>But they don&#8217;t change system behaviour.</p><p>That&#8217;s the ritual mindset: <strong>if we do more of the same things, reliability will improve.</strong> Except when it doesn&#8217;t. Now, let&#8217;s look at what happens to a lot of those rituals in the real world.</p><p>And in practice, they often falter:</p><ul><li><p>Retro held, but nothing changes despite learnings</p></li><li><p>Action items logged, never acted on</p></li><li><p>SLOs are written but don&#8217;t fully influence prioritization</p></li><li><p>Everyone &#8220;cares&#8221;, but nobody changes behaviour</p></li><li><p>&#8220;We added alerts!&#8221; (&#8230;and made things worse)</p></li></ul><p>&#129300; <strong>Why don&#8217;t rituals consistently improve reliability?</strong></p><p>Because they don&#8217;t change the parts of the system that <em>produce</em> reliability. Reliability is produced by what MIT&#8217;s System Dynamics group defines as <strong>feedback loops.</strong></p><p>These loops can include delays in judgment and action, handoff frequency, review cadence, and how fast learning propagates. Rituals don&#8217;t touch any of that. They create <em>activity</em>, not <em>loop correction</em>.</p><p>You can run a retro every sprint, but if:</p><ul><li><p>the process stays the same,</p></li><li><p>the information flows stay the same,</p></li><li><p>the handoffs stay the same,</p></li><li><p>and the delays stay the same&#8230;</p></li></ul><p><strong>the loop never changes.</strong></p><p>The ritual happens around the system, but nothing shifts inside the system.</p><p>Next time, I&#8217;ll map out the loops underneath all this &#8212; the mechanisms rituals never touch.</p>]]></content:encoded></item><item><title><![CDATA[You (and AI) can't automate reliability away]]></title><description><![CDATA[What if the hardest part of reliability has nothing to do with tooling or automation?]]></description><link>https://read.srepath.com/p/you-and-ai-cant-automate-reliability</link><guid isPermaLink="false">https://read.srepath.com/p/you-and-ai-cant-automate-reliability</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 02 Dec 2025 13:03:47 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/179318513/1b715e971fdb5ab78a4c7fd545894f54.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p><em>What if the hardest part of reliability has nothing to do with tooling or automation? Jennifer Petoff explains why real reliability comes from the human workflows wrapped around the engineering work.</em></p><p><strong>Everyone seems to think AI will automate reliability away.</strong> </p><p>I keep hearing the same story: </p><p><em>&#8220;Our tooling will catch it.&#8221;</em> </p><p><em>&#8220;Copilots will reduce operational load.&#8221;</em> </p><p><em>&#8220;Automation will mitigate incidents before they happen.&#8221;</em></p><p>But here&#8217;s a hard truth to swallow: AI only automates the mechanical parts of reliability &#8212; the machine in the machine.</p><p><strong>The hard parts haven&#8217;t changed at all.</strong></p><p>You still need teams with clarity on system boundaries.<br>You still need consistent approaches to resolution.<br>You still need postmortems that drive learning rather than blame.</p><p><strong>AI doesn&#8217;t fix any of that.</strong> If anything, it exposes every organizational gap we&#8217;ve been ignoring. And that&#8217;s exactly why I wanted today&#8217;s guest on.</p><p>Jennifer Petoff is Director of Program &#8202;Management for Google Cloud Platform and Technical Infrastructure education.  Every day, she works with SREs at Google, as well as with SREs at other companies through her public speaking and Google Cloud Customer engagements.</p><p>Even if you have never touched GCP, you have still been influenced by her work at some point in your SRE career. She is co-editor of Google&#8217;s original Site Reliability Engineering book from 2016. </p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!EQLo!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!EQLo!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp 424w, https://substackcdn.com/image/fetch/$s_!EQLo!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp 848w, https://substackcdn.com/image/fetch/$s_!EQLo!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp 1272w, https://substackcdn.com/image/fetch/$s_!EQLo!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!EQLo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp" width="232" height="296.96" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:512,&quot;width&quot;:400,&quot;resizeWidth&quot;:232,&quot;bytes&quot;:null,&quot;alt&quot;:&quot;Site Reliability Engineering&quot;,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="Site Reliability Engineering" title="Site Reliability Engineering" srcset="https://substackcdn.com/image/fetch/$s_!EQLo!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp 424w, https://substackcdn.com/image/fetch/$s_!EQLo!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp 848w, https://substackcdn.com/image/fetch/$s_!EQLo!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp 1272w, https://substackcdn.com/image/fetch/$s_!EQLo!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F93f293e8-14b4-4dc2-8a3e-e88582286cb9_400x512.webp 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Yeah, that one!</p><p>It was my immense pleasure to have her join me to discuss the internal dynamics behind successful reliability initiatives. Here are 5 highlights from our talk:</p><h2>3 issues stifling individual SREs&#8217; work</h2><p>To start, I wanted to know from Jennifer the kinds of <strong>challenges she has seen individual SREs face</strong> when attempting to introduce or reinforce reliability improvements within their teams or the broader organization.</p><p>She categorized these challenges into 3 main categories</p><ol><li><p>Cultural issues (with a look into Westrum&#8217;s typology of organizational culture)</p></li><li><p>Insufficient buy-in from stakeholders</p></li><li><p>Inability to communicate the value of reliability work</p></li></ol><div class="pullquote"><p>Organizations with generative cultures have 30% better organizational performance.</p></div><p>A key highlight from this topic came from her look at DORA research, an annual survey of thousands of tech professionals and the research upon which the book <em>Accelerate</em> is based.</p><p>It showed that organizations with generative cultures have 30% better organizational performance. In other words, you can have the best technology, tools, and processes to get good results, but culture further raises the bar. </p><p>A generative culture also makes it easier to implement the more technical aspects of DevOps or SRE that are associated with improved organizational performance.</p><h2>Hands-on is the best kind of training</h2><p>We then explored structured approaches that ensure consistency, build capability, and deliberately shape reliability culture. As they say &#8211; <em><strong>Culture eats strategy for breakfast!</strong></em></p><p>One key example Jennifer gave was the hands-on approach they take at Google. She believes that <strong>adults learn by doing. In other words, SREs</strong> <strong>gain confidence by</strong> <strong>doing hands-on work</strong>. </p><p>Where possible, training programs should move away from passive listening to lectures toward hands-on exercises that mimic real SRE work, especially troubleshooting.</p><p>One specific exercise that Google has built internally is <strong>Simulating Production Breakages.  </strong></p><p>Engineers undergoing that training have a chance to troubleshoot a real system built for this purpose in a safe environment. </p><p>The results have been profound, with a tremendous amount of confidence that Jennifer&#8217;s team saw in survey results. </p><blockquote><p>This confidence is focused on job-related behaviors, which when repeated over time reinforce that culture of reliability.</p></blockquote><h2>Reliability is mandatory for <em>everybody</em></h2><p>Another thing Jennifer told me Google did differently was making reliability a mandatory part of <em>every </em>engineer&#8217;s curriculum, not only SREs.</p><blockquote><p>When we first spun up the SRE Education team, our focus was squarely on our SREs. However, that&#8217;s like preaching to the choir. SREs are usually bought into reliability. </p><p>A few years in, our leadership was interested in propagating the reliability-focused culture of SRE to all of Google&#8217;s development teams, a challenge an order of magnitude greater than training SREs. </p></blockquote><p>How did they achieve this mandate?</p><ul><li><p>They developed a short and engaging (and mandatory) production safety training</p></li><li><p>That training has now been taken by tens of thousands of Googlers</p></li><li><p>Jennifer attributes this initiative&#8217;s success to how they&#8220;SRE&#8217;ed the program&#8221;. </p><p><em>&#8220;We ran a canary followed by a progressive roll-out. We instituted monitoring and set up feedback loops so that we could learn and drive continuous improvement.&#8221;</em></p></li></ul><p>The result of this massive effort? </p><p>A very respectable 80%+ net promoter score with open text feedback: &#8220;best required training ever.&#8221;</p><pre><code>What made this program successful is that Jennifer and her team SRE&#8217;d its design and iterative improvement. 

You can learn more about &#8220;<strong>How to SRE anything</strong>&#8221; (from work to life) using her rubric: 
https://www.reliablepgm.com/how-to-sre-anything/</code></pre><h2>Reliability gets rewarded just like feature work</h2><p>Jennifer then talked about how Google mitigates a risk that I think every reliability engineer wishes could be solved at their organization. </p><p>That is, having great reliability work rewarded at the same level as great feature work.</p><p>For development and operations teams alike at Google, this means making sure &#8220;grungy work&#8221; like tech debt reduction, automation, and other activities that improve reliability are rewarded equally to shiny new product features. </p><p>Organizational reward programs that recognize outstanding work typically have committees. These committees not only look for excellent feature development work, but also reward and celebrate foundational activities that improve reliability. </p><p>This is explicitly built into the rubric for judging award submissions.</p><h2>Keep a scorecard of reliability performance</h2><p>Jennifer gave another example of how Google judges reliability performance, but more specifically for SRE teams this time. </p><div class="pullquote"><p>Google&#8217;s <strong>Production Excellence (ProdEx) program</strong> was created in 2015 to <strong>assess and improve production excellence (aka reliability improvements)</strong> across SRE teams.</p></div><p>ProdEx acts like a <strong>central scorecard</strong> <strong>to aggregate metrics from various production health domains</strong> to provide a comprehensive overview of an SRE team&#8217;s health and the reliability of the services they manage. </p><p>Here are some specifics from the program:</p><ul><li><p>Domains include SLOs, on-call workload, alerting quality, and postmortem discipline</p></li><li><p>Reviews are conducted live every few quarters by senior SREs (directors or principal engineers) who are not part of the team&#8217;s direct leadership</p></li><li><p>There is a focus on coaching and accountability without shame (to elicit psychological safety)</p></li></ul><p>ProdEx serves various levels of the SRE organization through:</p><ol><li><p>providing strategic situational awareness regarding organizational and system health to leadership <em>and</em></p></li><li><p>keeping forward momentum around reliability and surfacing team-level issues early to support engineers in addressing them</p></li></ol><h2>Wrapping up</h2><p>Having an inside view of reliability mechanisms within a few large organizations, I know that few are actively doing all &#8212; or sometimes any &#8212; of the reliability enhancers that Google uses and Jennifer has graciously shared with us. </p><p>It&#8217;s time to get the ball rolling. What will you do today to make it happen?</p>]]></content:encoded></item><item><title><![CDATA[Reliability engineers are the best SDLC problem solvers because...]]></title><description><![CDATA[they look at the system with fresh eyes and can spot patterns people too close to the problem have learned to ignore. I'll cover how you can do this later in the post.]]></description><link>https://read.srepath.com/p/reliability-engineers-are-the-best</link><guid isPermaLink="false">https://read.srepath.com/p/reliability-engineers-are-the-best</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 18 Nov 2025 12:20:41 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/13356861-c0b6-4a30-9add-91c738df2d0f_640x427.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>One thing I&#8217;ve learned from watching reliability masters doing their magic is that <strong>SREs end up solving the SDLC problems no one else sees &#8212; at least not clearly.</strong></p><p>It&#8217;s not because SREs are better at debugging.<br>It&#8217;s not because SREs are the &#8220;process optimizers&#8221; (that one still surprises people when I say it).</p><p>It&#8217;s because SREs look at the system from a different altitude. </p><p>They can see the parts of the SDLC that quietly shape everything but rarely show up as activities/outcomes in Jira, Git, or dashboards.</p><p>Most engineering teams have been indoctrinated into an industrialized version of DevOps. The daily mantra: <strong>ship new features, open more branches, close more tickets.</strong></p><p>But a surprising amount of reliability pain comes from the parts of the SDLC that don&#8217;t fit neatly into that cycle. The things no one &#8220;owns,&#8221; because they&#8217;re not obviously tied to delivery velocity.</p><p>A few examples you&#8217;ll recognise instantly:</p><ul><li><p><strong>The release cadence frozen in 2022</strong> - everyone remembers <a href="https://www.cnbc.com/2022/11/09/tech-layoffs-2022.html">what kind of year </a><em><a href="https://www.cnbc.com/2022/11/09/tech-layoffs-2022.html">that</a></em><a href="https://www.cnbc.com/2022/11/09/tech-layoffs-2022.html"> was</a></p></li><li><p><strong>The ticket queue that magically moves</strong> because one engineer quietly triages it every morning</p></li><li><p><strong>The API contract that changed three times</strong>, but half the dependent services never updated their call parameters</p></li></ul><p>These things fade into the background because everyone&#8217;s racing to grow services, not slowing down to prune the system with bonsai-level care.</p><p>That&#8217;s where the <em>outsider advantage</em> kicks in. </p><p>Because SREs work across multiple teams and tech stacks, they spot patterns that day-to-day insiders simply stop seeing. </p><p>Not because those engineers lack skill, but because <strong>familiarity blinds you to the system&#8217;s oddities.</strong></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://read.srepath.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Reliability Enablers! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p>This clicked for me while recently watching a live lecture by William Dalrymple, the Scottish historian known for his work on India&#8217;s colonial period. </p><p>His biggest discoveries came from re-reading old sources differently. </p><p>He reviewed materials that other historians had access to but rarely prioritized, including Persian court chronicles, Maratha records, Portuguese archives, and private letters sent by English sailors.</p><p>He said his <strong>craft isn&#8217;t about hunting for hidden facts. It&#8217;s about noticing the things filtered out over time</strong> by local and foreign historians.</p><p>He compared it to walking through London as a tourist. Suddenly, every red postbox, royal crest, and double-decker bus jumps out at you.</p><p>Residents never see them that way. Their brains auto-classify those details as &#8220;background texture.&#8221;</p><p>Masterful SREs do something similar in the SDLC.</p><p>They notice the &#8220;background texture&#8221; that quietly shapes reliability but has become invisible to the people living inside the system every day.</p><p>Their edge isn&#8217;t extreme technical depth (though a baseline is obviously required).<br><strong>Their edge is that they</strong> <strong>refuse to go numb to the system&#8217;s defaults.</strong></p><p>My own soft landing into reliability looked exactly like this:</p><pre><code>In the early 2010s, I co-founded a startup in a high-stakes finance vertical. Reliability wasn&#8217;t optional. It was the credibility signal that clients judged us on. If the system shook, the business shook. That forces one to notice what others gloss over.

Years later, the healthcare organization I worked in went through a messy public&#8211;private partnership digital-transformation program. You can imagine how well <em>that</em> went. Azure was still new to most of the engineers, incidents piled up, and we kept hitting reliability problems nobody could quite trace.

Somehow I ended up being the one who had to spot what everyone else missed and push it through the chain every week. Not the flashy issues &#8212; the structural ones: workflow drift, mismatched expectations, fuzzy ownership, and architecture shaped by people with PhDs in bureaucracy and technical skills ending with Internet Explorer.</code></pre><p>Here&#8217;s the part engineers sometimes underestimate:<br><strong>Fresh eyes are a skill. You can train this skill.</strong></p><p>A few practices could make a massive difference for your own reliability work:</p><p><strong>1. Become a tourist every week</strong><br>Write yourself a weekly tourist pass that lets you pick a workflow or pipeline at random, and then walk through it like you&#8217;ve never seen it.<br><em>What steps only make sense because you already know the history or how your org works? Because the next person working on this workflow or pipeline might not.</em></p><p><strong>2. Audit the default settings</strong><br>Most regressions result from defaults no one has revisited on a regular cadence, so review the retry logic and thresholds, but also the ownership and handoffs.<br><em>If you say to yourself, &#8220;Why is this the way it is?&#8221;, that&#8217;s your sign to dig deeper.</em></p><p><strong>3. Seek out organizational complacency </strong><br>Identify 3 things your org treats as normal &#8212; the &#8220;that&#8217;s how we&#8217;ve always done it&#8221; &#8212; but really shouldn&#8217;t be. For example, tech debt is easy to blame, but coordination debt is harder to pin down. <em>How does it cause downstream issues?</em></p><div><hr></div><p>Reliability improves the moment you start paying attention to the parts of the SDLC that everyone else filters out. That&#8217;s the real SRE advantage, and it&#8217;s available to anyone who wants to develop it.</p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://read.srepath.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Reliability Enablers! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[What a Slovak test engineer taught me about SRE career growth in a post-ZIRP world]]></title><description><![CDATA[This post was inspired by my chance meeting with a Slovak test engineer who had corporates fighting to hire him and sponsor his high-salary migration in a tight job market.]]></description><link>https://read.srepath.com/p/what-a-slovak-test-engineer-taught</link><guid isPermaLink="false">https://read.srepath.com/p/what-a-slovak-test-engineer-taught</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 28 Oct 2025 12:39:26 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!hjhf!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F99ee1dc2-77bf-4ffa-b056-f66dac8ad0d0_128x128.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><em>I can&#8217;t make up a story like this &#8212; and neither can GPT-5 (though it told me to use the em dash in this sentence).</em></p><p>I met a Slovak test engineer recently at a social event. When I go to these meetups, my goal is to unwind &#8212; not to talk about tech, or business, or the business of tech.</p><p>But within twenty minutes, we&#8217;d touched on all of it. He was simply happy to meet socially open people at our weekly pub gathering in my hometown.</p><p>And yet it all came out. </p><p>He&#8217;s a test engineer in the banking sector. I didn&#8217;t want to know more, but everyone else at the table did. So I got pulled into the vortex of systemizing his experience through the familiar frameworks and tropes of our industry.</p><p>Interestingly, he had no formal qualification in computing. No CS degree, minimal certifications, no fancy title. Yet every time a core-banking integration failed, everyone waited for <em>him</em> to show up.</p><p>He&#8217;d been seconded from his Slovak post at a multinational into a complex project serving one of Australia&#8217;s largest companies &#8212; think US$150 billion market cap. What was more interesting than his r&#233;sum&#233; was his <em>approach</em> to the problem. </p><blockquote><p>He did not understand why everyone was jumping up and down, trying to keep him in country for the long-run. I did&#8230;</p></blockquote><p>Projects like this, in the <a href="https://newsletter.pragmaticengineer.com/p/zirp-software-engineers">post-ZIRP world</a>, after years of zero-interest rates and cheap capital fueling endless &#8220;digital transformations&#8221;, now have to justify every line of spend with actual reliability.</p><p>The same tightening is visible in Australia&#8217;s internal tech sector. During the ZIRP decade, domestic headcount grew on the back of cheap money and &#8220;digital-transformation&#8221; budgets that rewarded expansion over efficiency.</p><p>Now, with the cost of capital normalized, those roles are being quietly unwound through redundancies and restructures. What remains are the engineers who can <em>stabilize</em> systems &#8212; like our Slovak friend.</p><p>He&#8217;s the kind of engineer who has quietly mastered<strong> context, not code</strong>. Whether at a domain (banking) or a system (multi-service dependencies) level, he seeks beyond the baseline and digs deeper.</p><p>Most of us in the industry, especially those shaped by the traditional IT career pipeline, are products of a linear system<strong>. </strong>We&#8217;re trained early in life to think in progressions<strong>:</strong> clear the exam, earn the certification, reach the next title.</p><p>The whole infrastructure rewards throughput and correctness &#8212; not exploration.</p><p>The Slovak engineer had come up through something very different.</p><p>His early training was hands-on, constraint-driven, the kind you find in vocational settings with:</p><ul><li><p>limited tooling</p></li><li><p>tight budgets</p></li><li><p>systems that had to keep running because no vendor contract could bail you out</p></li></ul><p>He learned by <em>fixing what broke</em>, not by <em>studying what should work and hoping that it would</em>.</p><p>That&#8217;s the real divide I saw that night &#8212; not East versus West, or degree versus no degree &#8212; but <strong>theoretical progression versus practical insight.</strong></p><p>He&#8217;d been forced to see systems for what they really are, and how they behave in reality, not how they&#8217;re diagrammed in a Visio document. </p><blockquote><p>Where many engineers measure growth in certifications, he measured it in <em>failure modes understood</em>.</p></blockquote><p>That&#8217;s why he looked puzzled when people praised him or tried to negotiate hard to keep him. In his world, competence wasn&#8217;t a r&#233;sum&#233; advantage. It was the minimum requirement for keeping the system alive.</p><p>At one point, he squirmed on his hardwood stool as I mentioned the likely negotiations between his seconding firm and the Australian companies trying to secure him permanently.</p><p>I expected pride. Instead, he shrugged. He cared more about what he could do with the system and the problems it threw at him. That was the real excitement for him.</p><p>So how does he <em>system up</em>?</p><p>In every outage call, he wasn&#8217;t asking <em>&#8220;What failed?&#8221;</em></p><p>He was asking, <em>&#8220;Why did the system allow this to fail?&#8221;</em></p><p>That question alone places him a decade ahead of most engineers in leadership capability.</p><p>It&#8217;s easy to romanticize stories like this, but a real shift is happening underneath:</p><div class="pullquote"><p>The industry is rediscovering that <strong>sensemaking of systems beats formal credentials</strong>.</p></div><p>He didn&#8217;t strategize ways to climb up fast; he <em>dug deep </em>and was rewarded with a faster climb up than any certification could deliver. </p><p>He tested through shaky pipelines until they whispered their secrets.</p><p>He understood the underlying system&#8217;s <em>personality </em>&#8212; its hidden assumptions and dependencies built up through years of tech debt. He&#8217;s the kind of engineer high-paying Australian banks are quietly fighting to keep long-term.</p><p>That encounter reminded me of something we often miss in our own careers, especially in the Indian tech ecosystem, which I&#8217;ve become more cognizant of in recent years. </p><p><strong>Many engineers are trained to think in</strong> <strong>levels </strong>&#8212; L4 to L5, Senior to Staff, Team Lead to Manager. Every promotion signals a &#8220;level up in capability.&#8221; But does it count in an era when systems literacy is critical in managing incidents?</p><p>Reliability work doesn&#8217;t reward hierarchy. It rewards <strong>systems literacy</strong>: the ability to see across components, people, and processes, and to hold that complexity in your head while you&#8217;re trying to fix things.</p><blockquote><p>&#8220;Leveling up&#8221; is about scope and title.<br>&#8220;Systeming up&#8221; is about building insight and ownership.</p></blockquote><p>It&#8217;s the difference between managing a dashboard and understanding the feedback loop behind it.</p><p>When I run workshops with senior engineers, I don&#8217;t care how many frameworks they&#8217;ve memorized. I care whether they can answer:</p><blockquote><p>&#8220;When the alert fires, what happens next in the real world &#8212; and why?&#8221;</p></blockquote><p>That question separates framework followers from system thinkers.</p><p>The best engineers &#8212; whether from India, the United States, or central Europe &#8212; are no longer measured by tenure, but by how reliably they can <em>stabilize the systems that keep the enterprise &#8212; and everyone&#8217;s salary &#8212; running.</em></p><p>So if you&#8217;re planning your growth in this career of reliability engineering, don&#8217;t just chase a higher level.</p><ol><li><p>Map the system you operate in <em>and</em> </p></li><li><p>Find the failure modes no one else has taken the time to understand.</p></li></ol><p>That&#8217;s how you truly <em>level up</em> in<em> y</em>our SRE career.</p>]]></content:encoded></item><item><title><![CDATA[AWS us-east-1 sneezed, the world felt its breeze]]></title><description><![CDATA[Race conditions 101, a global ripple, and looking at our hidden dependencies. We'll go through questions, like "Why did one region dropping cause global level issues?"]]></description><link>https://read.srepath.com/p/aws-us-east-1-sneezed-the-world-felt</link><guid isPermaLink="false">https://read.srepath.com/p/aws-us-east-1-sneezed-the-world-felt</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Fri, 24 Oct 2025 12:30:44 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/c1c4ff42-4221-460d-9787-11888c79cd4e_640x426.jpeg" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>As you are already aware, on October 20, 2025, AWS experienced its <strong>longest outage in a decade</strong>. You&#8217;d be forgiven for assuming it was a data center meltdown, or one of the usual suspects: a cyberattack, catastrophic hardware failure, or misconfigured network deployment. </p><p><em>It was none of those.</em></p><p>It was something more ordinary, but at the same time more hidden. A <strong>race condition in DNS management </strong>brought the whole house down. It&#8217;s the kind of gremlin that lives quietly inside distributed systems for years, waiting for the perfect moment to strike.</p><p>When us-east-1 (Northern Virginia) sneezed, the internet &#8212; from Seattle to Melbourne &#8212; caught a cold. So what actually happened, and what can teams learn <em>without</em> grumbling through another tick-box incident retro?</p><div><hr></div><h3><strong>So what actually happened on October 20th?</strong></h3><pre><code>~14:10 UTC</code></pre><p>A change to Amazon&#8217;s internal DNS configuration triggered a <strong>race condition between replication processes</strong> responsible for updating authoritative name servers. For a few minutes, different replicas disagreed on which version of a DNS zone file was &#8220;true.&#8221; That mismatch rippled outward. It was cached by resolvers, retried by clients, and multiplied by automation loops expecting stable DNS.</p><pre><code>~14:20 UTC</code></pre><p>Services that depended on AWS&#8217;s control plane began timing out. Route 53, EC2 instance metadata, and IAM authentication all started wobbling hard and in the process, took out a wide slice of internet plumbing. Everything from CI/CD pipelines to food-delivery apps and payment gateways slowed or failed outright.</p><pre><code>by 15:00 UTC</code></pre><p>AWS engineers isolated the offending processes within the us-east-1 region and began rolling back by 15:00 UTC. But if you&#8217;ve ever made DNS changes, you&#8217;d know that DNS caches live in the wild, and so, recovery was protracted. For some users, the internet, their conduit to the outside world, vanished for 20-odd hours.</p><p>Now, let&#8217;s address some questions that arise from this event:</p><div><hr></div><h3><strong>First, what is a race condition in DNS management, and why does it matter?</strong></h3><p>Think of a race condition as two parts of a system trying to update the same record at the same time, and the system saying, &#8220;Both of you go ahead and update this record simultaneously. I&#8217;ll sort it out later.&#8221;</p><p>In DNS terms, that can mean two replication processes trying to modify or propagate zone data simultaneously, where <strong>timing rather than logic decides who wins. </strong>For a few milliseconds, caches see stale or inconsistent entries, and suddenly, half the internet can&#8217;t find the right IP for a perfectly healthy service.</p><p>Normally, DNS heals quickly from this. But when the affected zone is AWS&#8217;s<em> own internal</em> DNS infrastructure &#8212; the guidepost that every AWS service uses to find the other &#8212; you&#8217;re well past a rapid self-healing situation.</p><p>This is the <strong>digital equivalent of deleting your city&#8217;s street signs</strong> while heavy out-of-town traffic is streaming in.  That&#8217;s why this outage hurt: it wasn&#8217;t a &#8220;catastrophic&#8221; failure &#8212; it was a consistency failure. </p><div><hr></div><h3><strong>Okay, but why did a Virginia outage break things in faraway regions like Australia?</strong></h3><p>Because <strong>not all dependencies are local</strong>, even if your data is. Many AWS control-plane services &#8212; DNS, API endpoints, authentication flows &#8212; are partly centralized in us-east-1, where AWS runs out of 100 or so computer warehouses. </p><p>So it&#8217;s no surprise to learn that when that region sneezes, the rest of the world gets a cold. Your app mirrored in Sydney may never touch Virginia&#8217;s compute nodes, but if it needs to resolve an AWS-managed domain, for example, it&#8217;s now more susceptible to a single point of failure (SPoF). </p><p>And that&#8217;s why popular banking tools like PayID in Australia could not function for some time.</p><div><hr></div><h3><strong>What kind of SPoF is this?</strong></h3><p>A logical one. The kind you can&#8217;t rack-and-stack. Engineering organizations spend fortunes removing physical SPOFs from their architectures. Things like multiple AZs, multi-region DR&#8212;but ignore the logical couplings baked into platform design. </p><p>DNS, IAM, and ancillary services are the invisible backbone of cloud-based software. They don&#8217;t look like dependencies until they break, and then you realize half your automation is pointing at one set of nameservers in a <em>down status</em> region.</p><div><hr></div><h3><strong>Could companies have avoided this?</strong></h3><p>Depends on what &#8220;avoid&#8221; means. Some engineers argued you could just use multiple providers so everyone doesn&#8217;t fail together. Others pointed out that testing, syncing data, and keeping a cross-provider setup functional costs more time and money than the occasional global outage. They&#8217;re both right. </p><p>The <strong>trade-off is between the</strong> <strong>cost of preparation</strong> <strong>and the</strong> <strong>cost of surprise</strong>. For many orgs, the latter still <em>feels</em> cheaper &#8212; until they&#8217;re on the wrong side of it.</p><div><hr></div><h3><strong>Why does everyone still rely on us-east-1?</strong></h3><p>Because that&#8217;s where AWS started, and inertia drives dependency. New services often launch there first. Docs default to it. SDKs assume it. It&#8217;s Conway&#8217;s Law of cloud regions: your architecture mirrors the provider&#8217;s organizational chart. As long as AWS itself treats us-east-1 as the &#8220;mothership,&#8221; customers will have to as well.</p><div><hr></div><h3><strong>What blind spots about our system of work does this reveal?</strong></h3><p>Resilience isn&#8217;t just about adding on extra regions for redundancy. It&#8217;s also about the conversations teams never have:</p><p><em>Who owns DNS? Who tests failover? Which global endpoints does our CI/CD depend on?</em></p><p>You can&#8217;t buy that awareness. You need to practice it &#8212; through game-days, dependency mapping, and post-incident reviews that ask &#8220;what assumptions failed?&#8221; instead of &#8220;who missed an alert?&#8221;. I know it sounds obvious, but how robust is your practice in these areas?</p><div><hr></div><h3><strong>Should teams actually do something about this?</strong></h3><p>Yes &#8212; but not as much as you&#8217;d think. You don&#8217;t need to make every service bulletproof. You just need to know which ones matter most and how they&#8217;ll behave when your cloud infrastructure starts shaking again. </p><div><hr></div><h2><strong>3 takeaways for practical engineers</strong></h2><ol><li><p><strong>Map hidden dependencies.</strong> Every cloud architecture has ghosts in the control plane &#8212; learn to know where yours live.</p></li><li><p><strong>Design for graceful degradation.</strong> Read-only mode beats &#8220;down for maintenance&#8221; pages, or worse, uncommunicated outages.</p></li><li><p><strong>Treat provider outages as mirrors.</strong> They reflect your own coupling, not just theirs.</p></li></ol><div><hr></div><h2><strong>Final thoughts</strong></h2><p>This wasn&#8217;t intended as a teardown of failure. It was designed to act as a reminder that complexity doesn&#8217;t collapse cleanly. For all our talk about multi-AZ, multi-region, and multi-cloud, most systems are still one race condition away from a very bad day. </p><p>Sometimes the most reliable fix isn&#8217;t more redundancy &#8212; it&#8217;s more insight.</p>]]></content:encoded></item><item><title><![CDATA[#67 Why the SRE Book Fails Most Orgs — Lessons from a Google Veteran]]></title><description><![CDATA[Listen now | Dave O'Connor has been an SRE leadership practitioner and coach for many years. Prior to this, he was an SRE and reliability leader at the director-level at Google for close to 16 years.]]></description><link>https://read.srepath.com/p/ex-googler-on-driving-reliability</link><guid isPermaLink="false">https://read.srepath.com/p/ex-googler-on-driving-reliability</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 15 Jul 2025 13:05:13 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/165789886/9c56332d942071a8f50da4da198173da.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>A new or growing SRE team. A copy of the book. A company that says it cares about reliability. <em>What happens next? </em>Usually&#8230; not much.</p><p>In this episode, I sit down with <strong>Dave O&#8217;Connor</strong>, <strong>a 16-year Google SRE veteran</strong>, to talk about what happens when organizations cargo-cult reliability practices without understanding the context they were born in.</p><p>You might know him for his self-deprecating wit and legendary USENIX blurb about being &#8220;complicit in the development of the SRE function.&#8221;</p><p>This one&#8217;s a treat &#8212; less &#8220;here&#8217;s a shiny new tool&#8221; and more &#8220;here&#8217;s what reliability <em>actually</em> looks like when you&#8217;ve seen it all.&#8221;</p><p>&#10024; <em>No vendor plugs from Dave at all, just a good old-fashioned chat about what works and what doesn&#8217;t.</em></p><p>Here&#8217;s what we dive into:</p><ul><li><p><strong>The adoption trap</strong>: Why SRE efforts often fail before they begin&#8212;especially when new hires care more about reliability than the org ever intended.</p></li><li><p><strong>The SRE book dilemma</strong>: Dave&#8217;s take on why following the SRE book chapter-by-chapter is a trap for most companies (and what to do instead).</p></li><li><p><strong>The cost of &#8220;caring too much&#8221;</strong>: How engineers burn out trying to force reliability into places it was never funded to live.</p></li><li><p><strong>You build it, you run it (but should you?)</strong>: Not everyone&#8217;s cut out for incident command&#8212;and why pretending otherwise sets teams up to fail.</p></li><li><p><strong>Buying vs. building</strong>: The real reason even conservative enterprises are turning into software shops &#8212; and the reliability nightmare that follows.</p></li></ul><p>We also discuss the evolving role of reliability in organizations today, from being mistaken for &#8220;just ops&#8221; to becoming a strategic investment (when done right).</p><p>Dave's seen the waves come and go in SRE &#8212; and he's still optimistic. That alone is worth a listen.</p>]]></content:encoded></item><item><title><![CDATA[#66 - Unpacking 2025 SRE Report’s Damning Findings ]]></title><description><![CDATA[This episode was prompted by the 2025 Catchpoint SRE Report, which dropped some damning but all-too-familiar findings. Sebastian joined me for this episode so you know it'll have great insights.]]></description><link>https://read.srepath.com/p/unpacking-2025-sre-reports-damning</link><guid isPermaLink="false">https://read.srepath.com/p/unpacking-2025-sre-reports-damning</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 01 Jul 2025 01:10:26 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/165791277/d94303c52fc563110b9a2ed26846b24f.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>I know it&#8217;s already six months into 2025, but we recorded this almost three months ago. I&#8217;ve been busy with my foray into the world of tech consulting and training &#8212;and, well, editing these podcast episodes takes time and care.</p><p>This episode was prompted by the <strong>2025 Catchpoint SRE Report</strong>, which dropped some damning but all-too-familiar findings:</p><ul><li><p><strong>53% of orgs still define reliability as uptime only</strong>, ignoring degraded experience and hidden toil</p></li><li><p><strong>Manual effort is creeping back in</strong>, reversing five years of automation gains</p></li><li><p><strong>41% of engineers feel pressure to ship fast</strong>, even when it undermines long-term stability</p></li></ul><p>To unpack what this actually means inside organizations, I sat down with <strong>Sebastian Vietz</strong>, Director of Reliability Engineering at Compass Digital and co-host of the Reliability Enablers podcast. </p><p>Sebastian doesn&#8217;t just talk about technical fixes &#8212; he focuses on the organizational frictions that stall change, burn out engineers, and leave &#8220;reliability&#8221; as a slide deck instead of a lived practice.</p><p>We dig into:</p><ul><li><p>How SREs get stuck as messengers of inconvenient truths</p></li><li><p>What it really takes to move from advocacy to adoption &#8212; without turning your whole org into a cost center</p></li><li><p>Why tech is more like <em>milk</em> than <em>wine</em> (Sebastian explains)</p></li><li><p>And how SREs can strengthen&#8212;not compete with&#8212;security, risk, and compliance teams</p></li></ul><p>This one&#8217;s for anyone tired of reliability theatrics. No kumbaya around K8s here. Just an exploration of the messy, human work behind making systems and teams more resilient.</p>]]></content:encoded></item><item><title><![CDATA[#65 - In Critical Systems, 99.9% Isn’t Reliable — It’s a Liability]]></title><description><![CDATA[In most SaaS, 99.9% uptime gets you promoted. In critical infrastructure like the energy sector, it gets flagged as a failure mode.]]></description><link>https://read.srepath.com/p/65-in-critical-systems-999-isnt-reliable</link><guid isPermaLink="false">https://read.srepath.com/p/65-in-critical-systems-999-isnt-reliable</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 17 Jun 2025 13:05:09 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/165835533/0767760aa43e11aa1fa534b9bf7af421.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Most teams talk about reliability with a margin for error. &#8220;What&#8217;s our SLO? What&#8217;s our budget for failure?&#8221; </p><p>But in the energy sector? <strong>There is no acceptable downtime. Not even a little.</strong></p><p>In this episode, I talk with <strong>Wade Harris, Director of FAST Engineering in Australia</strong>, who&#8217;s spent 15+ years designing and rolling out monitoring and control systems for critical energy infrastructure like power stations, solar farms, SCADA networks, you name it.</p><p>What makes this episode different is that Wade isn&#8217;t a reliability engineer by title, but it&#8217;s baked into everything his team touches. And that matters more than ever as software creeps deeper into operational technology (OT), and the cloud tries to stake its claim in critical systems.</p><p>We cover:</p><ul><li><p>Why <strong>100% uptime is the minimum bar</strong>, not a stretch goal</p></li><li><p>How the rise of renewables has <strong>increased system complexity</strong> &#8212; and what that means for monitoring</p></li><li><p>Why <strong>bespoke integration and SCADA spaghetti</strong> are still normal (and here to stay)</p></li><li><p>The <strong>reality of cloud risk</strong> in critical infrastructure (&#8220;the cloud is just someone else&#8217;s computer&#8221;)</p></li><li><p>What software engineers need to understand if they want their products used in serious environments</p></li></ul><p>This isn&#8217;t about observability dashboards or DevOps rituals. This is reliability <strong>when the lights go out and people risk getting hurt</strong> if you get it wrong.</p><p>And it&#8217;s a reminder: not every system lives in a feature-driven world. Some systems just have to work. Always. No matter what.</p>]]></content:encoded></item><item><title><![CDATA[A video on why reliability doesn't scale like it should]]></title><description><![CDATA["Why reliability work gets sidelined &#8212; even when everyone says it matters" (A takeaway from a former Google SRE Director featured in this video)]]></description><link>https://read.srepath.com/p/a-video-on-why-reliability-doesnt</link><guid isPermaLink="false">https://read.srepath.com/p/a-video-on-why-reliability-doesnt</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 01 Apr 2025 15:21:29 GMT</pubDate><enclosure url="https://substackcdn.com/image/youtube/w_728,c_limit/nmW-IrzAKas" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Hey SRE friend,</p><p>There&#8217;s a good chance I don&#8217;t know much about you &#8212; maybe not even your first name &#8212; thanks to how newsletters work these days.</p><p>But I do know you&#8217;re interested in reliability engineering. And chances are, you&#8217;re doing that work somewhere that isn&#8217;t exactly MAANG or a VC-fueled startup.</p><p>That&#8217;s why I made this research-backed video. </p><p>It looks at why reliability efforts stall in most orgs &#8212; with support from interviews, journal references, and real-world examples.</p><div id="youtube2-nmW-IrzAKas" class="youtube-wrap" data-attrs="{&quot;videoId&quot;:&quot;nmW-IrzAKas&quot;,&quot;startTime&quot;:null,&quot;endTime&quot;:null}" data-component-name="Youtube2ToDOM"><div class="youtube-inner"><iframe src="https://www.youtube-nocookie.com/embed/nmW-IrzAKas?rel=0&amp;autoplay=0&amp;showinfo=0&amp;enablejsapi=0" frameborder="0" loading="lazy" gesture="media" allow="autoplay; fullscreen" allowautoplay="true" allowfullscreen="true" width="728" height="409"></iframe></div></div><p>I&#8217;d love to know if it resonates with your experience.</p><p>&#8212; Ash</p><p><em>P.S. This is still in preview mode &#8212; unlisted for now and shared only with subscribers and collaborators. Feel free to pass it on to teammates if it sparks something useful.</em></p><div><hr></div><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://read.srepath.com/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Reliability Enablers. Subscribe for free to receive new posts.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[#64 - Using AI to Reduce Observability Costs]]></title><description><![CDATA[Exploring how to manage observability tool sprawl, reduce costs, and leverage AI to make smarter, data-driven decisions.]]></description><link>https://read.srepath.com/p/64-using-ai-to-reduce-observability</link><guid isPermaLink="false">https://read.srepath.com/p/64-using-ai-to-reduce-observability</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 28 Jan 2025 14:03:35 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/155579851/21d55b415ca03894981571709c73bff5.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p><em>Exploring how to manage observability tool sprawl, reduce costs, and leverage AI to make smarter, data-driven decisions.</em></p><div><hr></div><p>It's been a hot minute since the last episode of the Reliability Enablers podcast.</p><p>Sebastian and I have been working on a few things in our realms. On a personal <em>and </em>work front, I&#8217;ve been to over 25 cities in the last 3 months and need a breather.</p><p>Meanwhile, listen to this interesting vendor, Ruchir Jha from Cardinal, working on the cutting edge of o11y to help reduce costs from spiraling out of control. </p><p>(To the skeptics, he did not pay me for this episode)</p><p>Here&#8217;s an AI-generated summary of what you can expect in our conversation:</p><p>In this conversation, we explore cutting-edge approaches to FinOps i.e. cost optimization for observability. </p><p>You'll hear about three pressing topics:</p><ol><li><p><strong>Managing Tool Sprawl</strong>: Insights into the common challenge of juggling 5-15 tools and how to identify which ones deliver real value.</p></li><li><p><strong>Reducing Observability Costs</strong>: Techniques to track and trim waste, including how to uncover cost hotspots like overused or redundant metrics.</p></li><li><p><strong>AI for Observability Decisions</strong>: Practical ways AI can simplify complex data, empowering non-technical stakeholders to make informed decisions.</p></li></ol><p>We also touch on the balance between open-source solutions like OpenTelemetry and commercial observability tools. </p><p>Learn how these strategies, informed by Ruchir's experience at Netflix, can help streamline observability operations and cut costs without sacrificing reliability.</p>]]></content:encoded></item><item><title><![CDATA[#63 - Does "Big Observability" Neglect Mobile?]]></title><description><![CDATA[Andrew Tunall is a product engineering leader focused on pushing the boundaries of reliability with a current focus on mobile observability.]]></description><link>https://read.srepath.com/p/63-mobile-apps-and-how-observability</link><guid isPermaLink="false">https://read.srepath.com/p/63-mobile-apps-and-how-observability</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 12 Nov 2024 13:03:12 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/151260045/334ab290bb5ba3269054b9dba051f92d.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Andrew Tunall is a product engineering leader focused on pushing the boundaries of reliability with a current focus on mobile observability. Using his experience from AWS and New Relic, he&#8217;s vocal about the need for a more user-focused observability, especially in mobile, where traditional practices fall short. </p><p></p><ol><li><p><strong>Career Journey and Current Role</strong>: Andrew Tunall, now at Embrace, a mobile observability startup in Portland, Oregon, started his journey at AWS before moving to New Relic. He shifted to a smaller, Series B company to learn beyond what corporate America offered.</p></li><li><p><strong>Specialization in Mobile Observability</strong>: At Embrace, Andrew and his colleagues build tools for consumer mobile apps, helping engineers, SREs, and DevOps teams integrate observability directly into their workflows.</p></li><li><p><strong>Gap in Mobile Observability</strong>: Observability for mobile apps is still developing, with early tools like Crashlytics only covering basic crash reporting. Andrew highlights that more nuanced data on app performance, crucial to user experience, is often missed.</p></li><li><p><strong>Motivation for User-Centric Tools</strong>: Leaving &#8220;big observability&#8221; to focus on mobile, Andrew prioritizes tools that directly enhance user experience rather than backend metrics, aiming to be closer to end-users.</p></li><li><p><strong>Mobile's Role as a Brand Touchpoint</strong>: He emphasizes that for many brands, the primary consumer interaction happens on mobile. Observability needs to account for this by focusing on user experience in the app, not just backend performance.</p></li><li><p><strong>Challenges in Measuring Mobile Reliability</strong>: Traditional observability emphasizes backend uptime, but Andrew sees a gap in capturing issues that affect user experience on mobile, underscoring the need for end-to-end observability.</p></li><li><p><strong>Observability Over-Focused on Backend Systems</strong>: Andrew points out that &#8220;big observability&#8221; has largely catered to backend engineers due to the immense complexity of backend systems with microservices and Kubernetes. Despite mobile being a primary interface for apps like Facebook and Instagram, observability tools for mobile lag behind backend-focused solutions.</p></li><li><p><strong>Lack of Mobile Engineering Leadership in Observability</strong>: Reflecting on a former Meta product manager&#8217;s observations, Andrew highlights the lack of VPs from mobile backgrounds, which has left a gap in observability practices for mobile-specific challenges. This gap stems partly from frontend engineers often seeing themselves as creators rather than operators, unlike backend teams.</p></li><li><p><strong>OpenTelemetry&#8217;s Limitations in Mobile</strong>: While OpenTelemetry provides basic instrumentation, it falls short in mobile due to limited SDK support for languages like Kotlin and frameworks like Unity, React Native, and Flutter. Andrew emphasizes the challenges of adapting OpenTelemetry to mobile, where app-specific factors like memory consumption don&#8217;t align with traditional time-based observability.</p></li><li><p><strong>SREs as Connective Tissue</strong>: Andrew views Site Reliability Engineers (SREs) as essential in bridging backend observability practices with frontend user experience needs. Whether through service level objectives (SLOs) or similar metrics, SREs help ensure that backend metrics translate into positive end-user experiences&#8212;a critical factor in retaining app users.</p></li><li><p><strong>Amazon&#8217;s Operational Readiness Review</strong>: Drawing from his experience at AWS, Andrew values Amazon&#8217;s practice of operational readiness reviews before launching new services. These reviews encourage teams to anticipate possible failures or user experience issues, weighing risks carefully to maintain reliability while allowing innovation.</p></li><li><p><strong>Shifting Focus to &#8220;Answerability&#8221; in Observability</strong>: For Andrew, the goal of observability should evolve toward &#8220;answerability,&#8221; where systems provide engineers with actionable answers rather than mere data. He envisions a future where automation or AI could handle repetitive tasks, allowing engineers to focus on enhancing user experiences instead of troubleshooting.</p></li></ol>]]></content:encoded></item><item><title><![CDATA[#62 - Early Youtube SRE shares Modern Reliability Strategy]]></title><description><![CDATA[Hear about Andrew Fong's thoughts on modern SRE as an early employee at Youtube, VP of Infra @ Dropbox, co-founder of a cloud infra startup and Senior Director of Engineering at Databricks.]]></description><link>https://read.srepath.com/p/62-early-youtube-sre-shares-modern</link><guid isPermaLink="false">https://read.srepath.com/p/62-early-youtube-sre-shares-modern</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 05 Nov 2024 13:40:50 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/151182017/df3b19cd2e88d6c3cefe0b329d5b3e86.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Andrew Fong&#8217;s take on engineering cuts through the usual role labels, urging teams to start with the problem they&#8217;re solving instead of locking into rigid job titles. He sees reliability, inclusivity, and efficiency as the real drivers of good engineering. </p><p>In his view, SRE is all about keeping systems reliable and healthy, while platform engineering is geared toward speed, developer enablement, and keeping costs in check. It&#8217;s a values-first, practical approach to tackling tough challenges that engineers face every day.</p><p>Here&#8217;s a slightly deeper dive into the concepts we discussed:</p><ol><li><p><strong>Career and Evolution in Tech</strong>: Andrew shares his journey through various roles, from early SRE at Youtube to VP of Infrastructure at Dropbox to Director of Engineering at Databricks, with extensive experience in infrastructure through three distinct eras of the internet. He emphasized the transition from early infrastructure roles into specialized SRE functions, noting the rise of SRE as a formalized role and the evolution of responsibilities within it.</p></li><li><p><strong>Building Prodvana and the Future of SRE</strong>: As CEO of startup, Prodvana, they're focused on an "intelligent delivery system" designed to simplify production management for engineers, addressing cognitive overload. They highlight SRE as a field facing new demands due to AI, discussing insights shared with Niall Murphy and Corey Bertram around AI's potential in the space, distinguishing it from "web three" hype, and affirming that while AI will transform SRE, it will not eliminate it.</p></li><li><p><strong>Challenges of Migration and Integration</strong>: Reflecting on experiences at YouTube post-acquisition by Google, the speaker discusses the challenges of migrating YouTube&#8217;s infrastructure onto Google&#8217;s proprietary, non-thread-safe systems. This required extensive adaptation and &#8220;glue code,&#8221; offering insights into the intricacies and sometimes rigid culture of Google&#8217;s engineering approach at that time.</p></li><li><p><strong>SRE&#8217;s Shift Toward Reliability as a Core Feature</strong>: The speaker describes how SRE has shifted from system-level automation to application reliability, with growing recognition that reliability is a user-facing feature. They emphasize that leadership buy-in and cultural support are essential for organizations to evolve beyond reactive incident response to proactive, reliability-focused SRE practices.</p></li><li><p><strong>Organizational Culture and Leadership Influence</strong>: Leadership&#8217;s role in SRE success is highlighted as crucial, with examples from Dropbox and Google emphasizing that strong, supportive leadership can shape positive, reliability-centered cultures. The speaker advises engineers to gauge leadership attitudes towards SRE during job interviews to find environments where reliability is valued over mere incident response.</p></li><li><p><strong>Outcome-Focused Work Over Titles</strong>: Emphasis on assembling the right team based on skills, not titles, to solve technical problems effectively. Titles often distract from focusing on outcomes, and fostering a problem-solving culture over role-based thinking accelerates teamwork and results.</p></li><li><p><strong>Engineers as Problem Solvers</strong>: Engineers, especially natural ones, generally resist job boundaries and focus on solving problems rather than sticking rigidly to job descriptions. This echoes how iconic engineers like Steve Jobs valued versatility over predefined roles.</p></li><li><p><strong>Culture as Core Values</strong>: Organizational culture should be driven by core values like reliability, efficiency, and inclusivity rather than rigid processes or roles. For instance, Dropbox's infrastructure culture emphasized being a &#8220;force multiplier&#8221; to sustain product velocity, an approach that ensured values were integrated into every decision.</p></li><li><p><strong>Balancing SRE and Platform Priorities</strong>: The fundamental difference between SRE (Site Reliability Engineering) and platform engineering is their focus: SRE prioritizes reliability, while platform engineering is geared toward increasing velocity or reducing costs. Leaders must be cautious when assigning both roles simultaneously, as each requires a distinct focus and expertise.</p></li><li><p><strong>Strategic Trade-Offs in Smaller Orgs</strong>: In smaller companies with limited resources, leaders often face challenges balancing cost, reliability, and other objectives within single roles. It's advised to sequence these priorities rather than burden one individual with conflicting objectives. Prioritizing platform stability, for example, can help improve reliability in the long term.</p></li><li><p><strong>DevOps as a Philosophy</strong>: DevOps is viewed here as an operational philosophy rather than a separate role. The approach enhances both reliability and platform functions by fostering a collaborative, efficient work culture.</p></li><li><p><strong>Focus Investments for Long-Term Gains</strong>: Strategic technology investments, even if they might temporarily hinder short-term metrics (like reliability), can drive long-term efficiency and reliability improvements. For instance, Dropbox invested in a shared metadata system to enable active-active disaster recovery, viewing this as essential for future reliability.</p></li></ol>]]></content:encoded></item><item><title><![CDATA[#61 Scott Moore on SRE, Performance Engineering, and More]]></title><description><![CDATA[Scott's got a few interesting things to say about these topics and software operations in general!]]></description><link>https://read.srepath.com/p/61-scott-moore-on-sre-performance</link><guid isPermaLink="false">https://read.srepath.com/p/61-scott-moore-on-sre-performance</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 22 Oct 2024 11:24:50 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/150400932/37055ad2597cbf99d5955d8cca950aa0.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p></p>]]></content:encoded></item><item><title><![CDATA[#60 How to NOT fail in Platform Engineering]]></title><description><![CDATA[Ankit Wal from ThoughtWorks Asia Pacific gave me the inside word on this hot topic]]></description><link>https://read.srepath.com/p/60-how-to-not-fail-in-platform-engineering</link><guid isPermaLink="false">https://read.srepath.com/p/60-how-to-not-fail-in-platform-engineering</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 01 Oct 2024 13:01:34 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/149602222/f1990dbfa464f6d47fb1626c9cdd62db.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Here&#8217;s what we covered:</p><h3>Defining Platform Engineering</h3><ul><li><p><strong>Platform engineering</strong>: Building compelling internal products to help teams reuse capabilities with less coordination.</p></li><li><p><strong>Cloud computing connection</strong>: Enterprises can now compose platforms from cloud services, creating mature, internal products for all engineering personas.</p></li></ul><h3>Ankit&#8217;s career journey</h3><ul><li><p>Didn't choose platform engineering; it found him.</p></li><li><p>Early start in programming (since age 11).</p></li><li><p>Transitioned from a product engineer mindset to building internal tools and platforms.</p></li><li><p>Key experience across startups, the public sector, unicorn companies, and private cloud projects.</p></li></ul><h3>Singapore Public Sector Experience</h3><ul><li><p><strong>Public sector</strong>: Highly advanced digital services (e.g., identity services for tax, housing).</p></li><li><p><strong>Exciting environment</strong>: Software development in Singapore&#8217;s public sector is fast-paced and digitally progressive.</p></li></ul><h3>Platform Engineering Turf Wars</h3><ul><li><p><strong>Turf wars</strong>: Debate among DevOps, SRE, and platform engineering.</p><ul><li><p><strong>DevOps</strong>: Collaboration between dev and ops to think systemically.</p></li><li><p><strong>SRE</strong>: Operations done the software engineering way.</p></li><li><p><strong>Platform engineering</strong>: Delivering operational services as internal, self-service products.</p></li></ul></li></ul><h3>Dysfunctional Team Interactions</h3><ul><li><p><strong>Issue</strong>: Requiring tickets to get work done creates bottlenecks.</p><ul><li><p><strong>Ideal state</strong>: Teams should be able to work autonomously without raising tickets.</p></li><li><p><strong>Spectrum of dysfunction</strong>: From one ticket for one service to multiple tickets across teams leading to delays and misconfigurations.</p></li></ul></li></ul><h3>Quadrant Model (Autonomy vs. Cognitive Load)</h3><ul><li><p><strong>Challenge</strong>: Balancing user autonomy with managing cognitive load.</p></li><li><p><strong>Goal</strong>: Enable product teams with autonomy while managing cognitive load.</p></li><li><p><strong>Solution</strong>: Platforms should abstract unnecessary complexity while still giving teams the autonomy to operate independently.</p><h3>How it pans out</h3><ul><li><p><strong>Low autonomy, low cognitive load</strong>: Dependent on platform teams but a simple process.</p></li><li><p><strong>Low autonomy, high cognitive load</strong>: Requires interacting with multiple teams and understanding technical details (worst case).</p></li><li><p><strong>High autonomy, high cognitive load</strong>: Teams have full access (e.g., AWS accounts) but face infrastructure burden and fragmentation.</p></li><li><p><strong>High autonomy, low cognitive load</strong>: Ideal situation&#8212;teams get what they need quickly without detailed knowledge.</p></li></ul></li></ul><h3>Shift from Product Thinking to Cognitive Load</h3><ul><li><p><strong>Cognitive load focus</strong>: More important than just product thinking&#8212;consider the human experience when using the system.</p></li><li><p><strong>Team Topologies</strong>: Mentioned as a key reference on this concept of cognitive load management.</p></li></ul><h3>Platform as a Product Mindset</h3><ul><li><p><strong>Collaboration</strong>: Building the platform in close collaboration with initial users (pilot teams) is crucial for success.</p></li><li><p><strong>Product Management</strong>: Essential to have a product manager or team dedicated to communication, user journeys, and internal marketing.</p></li></ul><h3>Self-Service as a Platform Requirement</h3><ul><li><p><strong>Definition</strong>: Users should easily discover, understand, and use platform capabilities without human intervention.</p></li><li><p><strong>User Testing</strong>: Watch how users interact with the platform to understand stumbling points and improve the self-service experience.</p></li></ul><h3>Platform Team Cognitive Load</h3><ul><li><p><strong>Burnout Prevention</strong>: Platform engineers need low cognitive load as well. Moving from a reactive (ticket-based) model to a proactive, self-service approach can reduce the strain.</p></li><li><p><strong>Proactive Approach</strong>: Self-service models allow platform teams to prioritize development and avoid being overwhelmed by constant requests.</p></li></ul>]]></content:encoded></item><item><title><![CDATA[#59 Who handles monitoring in your team and how?]]></title><description><![CDATA[Monitoring responsibilities vary between organizations, and how your team handles them might differ significantly from others, especially companies like Google.]]></description><link>https://read.srepath.com/p/59-who-handles-monitoring-in-your</link><guid isPermaLink="false">https://read.srepath.com/p/59-who-handles-monitoring-in-your</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 24 Sep 2024 11:35:49 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/149286810/e9971afe5b1fd5e3e4822b8c8923c90a.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<h2>Why many copy Google&#8217;s monitoring team setup</h2><ol><li><p><strong>Google&#8217;s Influence. </strong>Google played a key role in defining the concept of software reliability.</p></li><li><p><strong>Success in Reliability.</strong> Few can dispute Google&#8217;s ability to ensure high levels of reliability <em>and</em> its ability to share useful ways to improve it in other settings</p><p></p><p><em><strong>BUT there&#8217;s a problem:</strong></em></p></li><li><p><strong>It&#8217;s not always replicable.</strong> While Google's practices are admired, they may not be a perfect fit for every team.</p></li></ol><h2>What is Google&#8217;s monitoring approach within teams?</h2><p>Here&#8217;s the thing that Google does:</p><ul><li><p>Google assigns one or two people per team to manage monitoring.</p></li><li><p>Even with centralized infrastructure, a dedicated person handles monitoring.</p></li><li><p>Many organizations use a separate observability team, unlike Google's integrated approach</p></li></ul><p>If your org is large enough <em>and</em> prioritizes reliability highly enough, you might find it feasible to follow Google&#8217;s model to the tee. </p><p>Otherwise, a centralized team with occasional &#8220;embedded x engineer&#8221; secondments might be more effective.</p><h2>Can your team mimic Google&#8217;s model?</h2><p>Here are a few things you should factor in:</p><h3>Size matters</h3><p>Google's model works because of its scale and technical complexity. Many organizations don&#8217;t have the size, resources, or technology to replicate this.</p><h3>What are the options for your team?</h3><h4>Dedicated monitoring team (very popular but $$$)</h4><p>If you have the resources, you might create a dedicated observability team. This might call for a ~$500k+ personnel budget so it&#8217;s not something that a startup or SME can easily justify. </p><h4>Dedicate SREs to monitoring work (effective but difficult to manage)</h4><p>You might do this on rotation or make an SRE permanently &#8220;responsible for all monitoring matters&#8221;. Putting SREs on permanent tasks might lead to burnout as it might not suit their goals, and rotation work requires effective planning.</p><h4>Internal monitoring experts (useful but hard capability)</h4><p>One or more engineers within teams could take on monitoring/observability responsibilities as needed and support the team&#8217;s needs. This should be how we get monitoring work done, but it&#8217;s hard to get volunteers across a majority of teams. </p><h2>Transitioning monitoring from project work to maintenance</h2><h3>2 distinct phases</h3><h4>Initial Setup (the &#8220;project&#8221;) </h4><p>SREs may help set up the monitoring/observability infrastructure. </p><p>Since they have breadth of knowledge across systems, they can help connect disparate services and instrument applications effectively.</p><h4>Post-project phase (&#8220;keep the lights on&#8221;)</h4><p>Once the system is up, the focus shifts from project mode to ongoing operational tasks. But who will do that?</p><h3>Who will maintain the monitoring system?</h3><h4>Answer: usually not the same team</h4><p>After the project phase, a new set of people&#8212;often different from the original team&#8212;typically handles maintenance.</p><h4>Options to consider (once again)</h4><ol><li><p><strong>Spin up a monitoring/observability team.</strong> Create a dedicated team for observability infrastructure.</p></li><li><p><strong>Take a decentralized approach. </strong>Engineers across various teams take on observability roles as part of their regular duties.</p></li><li><p><strong>Internal monitoring/observability experts.</strong> They can take responsibility for monitoring and ensure best practices are followed.</p></li></ol><p>The key thing to remember here is&#8230;</p><h3>Adapt to Your Organizational Context</h3><h4>One size doesn&#8217;t fit all</h4><p>Google's model may not work for everyone. Tailor your approach based on your organization&#8217;s specific needs.</p><h4>The core principle to keep in mind</h4><p>As long as people understand why monitoring/observability matters and pay attention to it, you're on the right track.</p><h4>Work according to engineer awareness</h4><p><em><strong>If engineers within product and other non-operations teams are aware of monitoring:</strong> </em>You can attempt to **decentralize the effort** and involve more team members.</p><p><em><strong>If awareness or interest is low:</strong></em><strong> c</strong>onsider **dedicated observability roles** or an SRE team to ensure monitoring gets the attention it needs.</p><h2>In conclusion</h2><p>There&#8217;s no universal solution. </p><p>Whether you centralize or decentralize monitoring depends on your team&#8217;s structure, size, and expertise. </p><p>The important part is ensuring that observability practices are understood and implemented in a way that works best for your organization.</p><div><hr></div><p>PS. Rather than spend an hour on writing, I decided to write in the style I normally use in a work setting i.e. &#8220;executive short-hand&#8221;. Tell me what you think.</p>]]></content:encoded></item><item><title><![CDATA[#58 Fixing Monitoring's Bad Signal-to-Noise Ratio]]></title><description><![CDATA[Sebastian and I looked further into common pitfalls in monitoring. A major issue is the poor signal-to-noise ratio of data. This often results from having too much irrelevant... (read below)]]></description><link>https://read.srepath.com/p/58-fixing-monitorings-bad-signal</link><guid isPermaLink="false">https://read.srepath.com/p/58-fixing-monitorings-bad-signal</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 17 Sep 2024 12:13:22 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/148997261/9271a7ba5a94bf6df96c07d83a9078a1.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>Monitoring in the software engineering world continues to grapple with poor signal-to-noise ratios. It&#8217;s a challenge that&#8217;s been around since the beginning of software development and will persist for years to come. </p><p>The core issue is the overwhelming noise from non-essential data, which floods systems with useless alerts. </p><p>This interrupts workflows, affects personal time, and even disrupts sleep.</p><p>Sebastian dove into this problem, highlighting that the issue isn't just about having meaningless pages but also the struggle to find valuable information amidst the noise. </p><p>When legitimate alerts get lost in a sea of irrelevant data, pinpointing the root cause becomes exceptionally hard.</p><p>Sebastian proposes a fundamental fix for this data overload: <strong>be deliberate with the data you emit.</strong> </p><p>When instrumenting your systems, be intentional about what data you collect and transport. </p><p>Overloading with irrelevant information makes it tough to isolate critical alerts and find the one piece of data that indicates a problem.</p><p>To combat this, focus on:</p><ol><li><p><strong>Being Deliberate with Data</strong>. Make sure that every piece of telemetry data serves a clear purpose and aligns with your observability goals.</p></li><li><p><strong>Filtering Data Effectively</strong>. Improve how you filter incoming data to eliminate less relevant information and retain what's crucial.</p></li><li><p><strong>Refining Alerts</strong>. Optimize alert rules such as creating tiered alerts to distinguish between critical issues and minor warnings.</p></li></ol><p>Dan Ravenstone, who leads platform at Top Hat, discussed &#8220;triaging alerts&#8221; recently.</p><p> He shared that managing millions of alerts, often filled with noise, is a significant issue. </p><p>His advice: <strong>scrutinize alerts for value, ensuring they meet the criteria of a good alert, and discard those that don&#8217;t impact the user journey</strong>.</p><p>According to Dan, the anatomy of a good alert includes:</p><ul><li><p>A run book</p></li><li><p>A defined priority level</p></li><li><p>A corresponding dashboard</p></li><li><p>Consistent labels and tags</p></li><li><p>Clear escalation paths and ownership</p></li></ul><p>To elevate your approach, consider using aggregation and correlation techniques to link otherwise disconnected data, making it easier to uncover patterns and root causes.</p><p>The learning point is simple: <strong>aim for quality over quantity.</strong> </p><p>By refining your data practices and focusing on what's truly valuable, you can enhance the signal-to-noise ratio, ultimately allowing more time for deep work rather than constantly managing incidents.</p>]]></content:encoded></item><item><title><![CDATA[#57 How Technical Leads Support Software Reliability]]></title><description><![CDATA[You might be familiar with the term, &#8220;technical lead&#8221;. You might even be working with one or a few right now. But how well do you know their ability to support your reliability work?]]></description><link>https://read.srepath.com/p/57-how-technical-leads-support-software</link><guid isPermaLink="false">https://read.srepath.com/p/57-how-technical-leads-support-software</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Tue, 10 Sep 2024 12:10:22 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/148676546/8510cd70601fee529f8e83dbc8905a1a.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>The question then condenses down to: <em>Can technical leads support reliability work?</em> </p><p>Yes, they can! </p><p>Anemari has been a technical lead for years &#8212; even spending a few years doing that at the coveted consultancy, Thoughtworks &#8212; and now coaches others. </p><p>She and I discussed the link between this role and software reliability.</p><h2>Technical lead &#8800; Engineering manager</h2><p>According to Anemari, a tech lead is a person with technical knowledge who is responsible for leading a technical team to align them with a common technical strategy. </p><p>For the most part, engineering managers are focused on the people aspect of the work. They coach engineers and other technical folk to improve their performance. </p><p><strong>In terms of activities, engineering managers</strong> coach, mentor, and support the development of their team members or direct reports. They&#8217;ll also go and bat for the team within the organization. </p><p><strong>Technical leads are more focused on</strong> guiding the technical work that these people do. Their focus includes work like architecture, design patterns, and implementing projects. They offer technical insights and mentorship to the team.</p><p>Anemari found in her work experience that the use of technical leads depends from company to company:</p><blockquote><p>I've worked with teams that don't have tech leads and then the engineering manager takes a more hands-on approach. Then you have teams where you have a tech lead only focused on tech, and then you have like a team lead or engineering manager, doing the people side.</p></blockquote><p>Interestingly, she found that it was very difficult to singularly focus on the technical side of the work:</p><blockquote><p>[Even] if you only want to focus on tech as a tech lead, you still end up having to deal with the people side because most tech problems are people problems in the end and so you kind of require both.  </p></blockquote><h2>How can technical leads drive reliability principles?</h2><p>Anemari advised me that tech leads are often required to think about reliability principles as part of their technical strategy.</p><p>Sometimes, teams don&#8217;t have an SRE team supporting them and are operating in a <strong>&#8220;you build it, you run it&#8221; mode. In this situation, reliability becomes a 90+ percent responsibility of the team.</strong></p><p>In other words, the product team has had to develop the reliability-focused knowledge required to build and run their product reliably.</p><p>The technical lead can help the less experienced team members pick up reliability concepts and not just have that mindset of creating more features and making the burndown chart look pretty.</p><p>However, their preference usually goes on working with a Site Reliability Engineering (SRE) team if it is possible.</p><h2>Working with reliability teams like SREs </h2><p>Anemari believes it&#8217;s important to be aware of reliability principles. She put it as knowing:</p><div class="pullquote"><p>What does it mean for my product to be reliable?</p></div><p>So, the first critical step is to align your definition of reliability from pondering this question with the known constraints of the product. These constraints can be:</p><ul><li><p>performance limits</p></li><li><p>operational environments</p></li><li><p>user expectations</p></li></ul><p>Anemari went on in our conversation to challenge the notion of simply stating reliability requirements to product teams. </p><p>It may seem obvious to reliability folk, but it&#8217;s one thing in a very large pile of considerations for a product team:</p><blockquote><p>So I think it's important when we're talking about [reliability]&#8230; when we're talking about uptime, like, what does that mean? And why do we need 99 percent or whatever?</p></blockquote><p>The onus falls on reliability engineers to support this thinking.</p><p>Technical leads can be a good point of contact for that initial conversation of alignment. The reliability team can advise about the standards we mentioned before. </p><p>So <strong>the</strong> <strong>first step is co-creating a common version of reliability</strong>. </p><p>The second step is continuous communication between the product and reliability teams. Anemari believes there must be an easy way for either party to communicate with the other.</p><p>&#8220;Creating a ticket&#8221; to reach the reliability team or vice versa does not work well.</p><p>Anemari has seen having a Slack channel between the SRE team and each product team as something that has worked well. </p><blockquote><p>You can go in there, pop up a question and take it from there and solve problems.</p></blockquote><p>Anemari is also a fan of the embedded SRE model.</p><blockquote><p>If you have to build something new from scratch and you have to build a whole infrastructure, another strategy that I've seen is bringing a person with SRE knowledge [into] the team for a while, helping us define [technical work] and being hands on.</p></blockquote><p>A long-term embedded SRE, however, might be an antipattern.</p><blockquote><p>I&#8217;ve had some teams that had an SRE team member&#8230; all the time, but I&#8217;ve also seen them struggle with the fact that at some point  there is not so much to do.</p></blockquote><p>An embedded SRE should be brought in to enable the teams to do more and take care of their products  By training the team on reliability topics &#8212; even at a high level &#8212; like understanding all the different dashboards that they have.</p><h2>What can SREs do better to support product teams with reliability efforts?</h2><p>Anemari told me that it would be great if SREs first put more effort into helping the product team understand why reliability is so important. It&#8217;s obvious to us in this space, but as I mentioned earlier, people have n+100 other things on their minds.</p><p>Product teams need to be made more aware of what the reliability engineers are taking care of and how that can affect their products. </p><p>There needs to be more conversation than what Anemari has experienced in a lot of encounters with SRE teams:</p><blockquote><p>The SRE teams are like, &#8220;Just give it to us. We'll take care of it. You know, like,   we, we know what to do. I don't have to explain all of these things to you.&#8221; </p></blockquote><p>She found it more effective to sit down with the reliability team and try to understand what might seem trivial, but are important questions like:</p><ul><li><p>Why is our service not processing enough requests and how can we change it?</p></li><li><p>Why can we just drop this?</p></li><li><p>Why do we need to have all of these services? </p></li></ul><p>Asking these kinds of questions helped the collaborators come up with a better solution than what the SRE team could come up with on their own.</p><p>Anemari has seen in the past when an SRE team came in and introduced themselves to the product team. They then covered how their services ran in the background. This led the product team to ask questions to brainstorm on what could be improved.</p><p>The intentionality is important: the SRE team did not come with horns blaring that they were going to change the systems. They came in to discuss and brainstorm ideas. </p><h2>But will the software engineers cooperate?</h2><p>A lot of  SREs I've spoken with are frustrated and think developer teams do not want to understand reliability.</p><p>But I'm sure at least <em>some</em> teams want to understand how reliability works. </p><p>Sometimes you have to just swallow that bitter pill and say, &#8220;Hey, look, I'm going to spell every single thing I'm doing out to you, just so that maybe in the future you can do it yourself.&#8221; </p><p>I think a lot of developers remember their early days and want to be able to run their product entirely on their own. </p><p>Anemari confirmed with her experience that developers want to have full power over what's happening with their product. Dependency is not necessarily something developers like very much.</p><p>There's of course the consideration of the learning journey.  </p><blockquote><p>There's that sweet spot that is hard to find early on, &#8220;How much do I need to know? Like, do I need to become an SRE expert now to run my service? Or how much do I need to know to just run  my service?&#8221; </p></blockquote><p>Anemari recalled one particular high-performing team that she led. The reason why it was high-performing was partly because they were able to run services fully with very little support from an external SRE team.</p><blockquote><p>We [brought] someone we knew with knowledge&#8230; to fully understand how does our AWS work and how to restart our services and everything so that we were able to fully run it and be on call for it.</p></blockquote><p>The sweet spot notion comes in once again where you have to work out, <em>&#8220;How much do I need to develop and spend time on developing this SRE knowledge, and how much do put into developing the product?&#8221;</em></p><p>It's complicated, but communication is the key to it.</p><h2>How can product teams handle conflicting priorities?  </h2><p>Anemari told me that the answer depends on the context: </p><ul><li><p>where you are</p></li><li><p>what situation your product is in </p></li><li><p>how the developer and SRE teams are laid out</p></li></ul><p>It&#8217;s very different to start a project from scratch versus having a whole monolith running for years and trying to make things better. </p><p>Anemari added:</p><blockquote><p>It might sound a little bit crazy but I still think that it's very important for you to like write down all of these different things that you think you need to solve&#8230; </p><p>A big part of the role of a tech lead is to make sure that all of these different parts agree and align on a strategy. So that means the product, that means the developer, that means the SREs, that means the customer support, etc.</p></blockquote><p>A tech lead would work to bring all of this together and then propose to all of these different stakeholders. You might have to put it as simply as, &#8220;This is what we have to focus on right now, given that as a product team. We have to keep delivering.&#8221;</p><p>Then come the compromises. For example, &#8220;20 percent of our time would go into improving our delivery pipeline.&#8221; </p><p>It is crucial to work out compromises before getting agreement from the various stakeholders in the software delivery organization. </p><div><hr></div><p>This write-up was just a preview of what Anemari and I talked about in this episode of the Reliability Enablers. Be sure to listen to get the other half of our conversation.</p><div><hr></div>]]></content:encoded></item><item><title><![CDATA[#56 Resolving DORA Metrics Mistakes]]></title><description><![CDATA[I asked Nathen Harvey who is lead DORA advocate at Google about some of the mistakes people make. The major ones seem to come down to management misconceptions about what it can do.]]></description><link>https://read.srepath.com/p/56-resolving-dora-metrics-mistakes</link><guid isPermaLink="false">https://read.srepath.com/p/56-resolving-dora-metrics-mistakes</guid><dc:creator><![CDATA[Ash Patel]]></dc:creator><pubDate>Wed, 04 Sep 2024 12:02:44 GMT</pubDate><enclosure url="https://api.substack.com/feed/podcast/148441662/fc4197978ba67641e1eb7586a50ae120.mp3" length="0" type="audio/mpeg"/><content:encoded><![CDATA[<p>We're already well into 2024 and it&#8217;s sad that people still have enough fuel to complain about various aspects of their engineering life.  </p><p>DORA seems to be turning into one of those problem areas.</p><p>Not at every organization, but some places are turning it into a case of &#8220;hitting metrics&#8221; without caring for the underlying capabilities and conversations.</p><p>Nathen Harvey is no stranger to this problem.</p><p>He used to talk a lot about SRE at Google as a developer advocate. Then, he became the lead advocate for DORA when Google acquired it in 2018. </p><p>His focus has been on questions like:</p><div class="pullquote"><p>How do we help teams get better at delivering and operating software? </p></div><p>You and I can agree that this is an important question to ask. </p><p>I&#8217;d listen to what he has to say about DORA because he&#8217;s got a wealth of experience behind him, having also run community engineering at Chef Software.</p><p>Before we continue, let&#8217;s explore <strong>What is DORA? </strong>in Nathen&#8217;s (paraphrased) words:</p><p>DORA is a software research program that's been running since 2015.</p><p>This research program looks to figure out:</p><div class="pullquote"><p>How do teams get good at delivering, operating, building, and running software? </p></div><p>The researchers were able to draw out the concept of the metrics based on <strong>correlating teams that have good technology practices with highly robust software delivery outcomes</strong>.</p><p>They found that this positively impacted organizational outcomes like profitability, revenue, and customer satisfaction.</p><p>Essentially, all those things that matter to the business.</p><p>One of the challenges the researchers found over the last decade was working out: <em>how do you measure something like software delivery?</em> </p><p>It's not the same as a factory system where you can go and count the widgets that we're delivering necessarily.</p><p>The unfortunate problem is that the factory mindset I think still leaks in. I&#8217;ve personally noted some silly metrics over the years like lines of code.</p><p>Imagine being asked constantly: &#8220;How many lines of code did you write this week?&#8221;</p><p>You might not have to imagine. It might be a reality for you. </p><p>DORA&#8217;s researchers agreed that the factory mode of metrics cannot determine whether or not you are a productive engineer. </p><p>They settled on and validated 4 key measures for software delivery performance.</p><p>Nathen elaborated that <strong>2 of these measures look at throughput</strong>:</p><blockquote><p>[Those] two [that] look at throughput really ask two questions:</p><ol><li><p>How long does it take for a change of any kind, whether it's a code change, configuration change, whatever, a change to go from the developer's workstation. right through to production?</p></li></ol><p>And then the second question on throughput is:</p><ol start="2"><li><p>How frequently are you updating production?</p></li></ol></blockquote><p>In plain English, these 2 metrics are:</p><ol><li><p><strong>Deployment Frequency</strong>. How often code is deployed to production? This metric reflects the team's ability to deliver new features or updates quickly.</p></li><li><p><strong>Lead Time for Changes</strong>: Measures the time it takes from code being committed to being deployed to production. </p></li></ol><p>Nathen recounted his experience of working at organizations that differed in how often they update production from once every six months to multiple times a day.  </p><p>They're both very different types of organizations, so their perspective on throughput metrics will be wildly different. </p><p>This has some implications for the <em>speed</em><strong> </strong>of software delivery.</p><p>Of course, <strong>everyone wants to move faster, but there&#8217;s this other thing that comes in and that's stability</strong>.</p><p>And so, the other two stability-oriented metrics look at:</p><blockquote><p>What happens when you do update production and... something's gone horribly wrong. &#8220;Yeah, we need to roll that back quickly or push a hot fix.&#8221; </p></blockquote><p>In plain English, they are:</p><ol start="3"><li><p><strong>Change Failure Rate</strong>: Measures the percentage of deployments that cause a failure in production (e.g., outages, bugs). </p></li><li><p><strong>Failed Deployment Recovery Time</strong>: Measures how long it takes to recover from a failure in production. </p></li></ol><p>You might be thinking the same thing as me. These stability metrics might be a lot more interesting to reliability folks than the first 2 throughput metrics.</p><p>But keep in mind, it&#8217;s about balancing all 4 metrics. </p><p>Nathen believes it&#8217;s fair to say today that across <strong>many organizations, they look at these concepts of throughput and stability as tradeoffs of one another</strong>. </p><div class="pullquote"><p>We can either be fast or we can be stable. </p></div><p>But the interesting thing that the DORA researchers have learned from their decade of collecting data is that throughput and stability aren't trade-offs of one another.</p><p>They tend to move together. They&#8217;ve seen organizations of every shape and size, in every industry, doing well across all four of those metrics. </p><p>They are the best performers. </p><p>The interesting thing is that the size of your organization doesn't matter the industry that you're in.</p><p>Whether you&#8217;re working in a highly regulated or unregulated industry, it doesn't matter.</p><p>The key insight that Nathen thinks we should be searching for is: <strong>how do you get there?</strong>  </p><p>To him, it's about shipping smaller changes. </p><p>When you ship small changes, they're easier to move through your pipeline. </p><p>They're easier to reason about. </p><p>And when something goes wrong, they're easier to recover from and restore service.</p><p>But along with those small changes, we need to think about those feedback cycles.</p><p>Every line of code that we write is in reality a little bit of an experiment. </p><p>We think it's going to do what we expect and it's going to help our users in some way, but we need to get feedback on that as quickly as possible.</p><p>Underlying all of this, both small changes and getting fast feedback, is a real climate for learning. Nathen drew up a few thinking points from this:</p><blockquote><p>So what is the learning culture like within our organization? </p><p>Is there a climate for learning? </p><p>And are we using things like failures as opportunities to learn, so that we can ever be improving?</p></blockquote><p>&#8202;I don&#8217;t know if you&#8217;re thinking the same as me already, but we're already learning that DORA is a lot more than just metrics.</p><p>&#8202;To Nathen (and me), <strong>the metrics should be one of the least interesting parts of DORA because it digs into useful capabilities, like small changes and fast feedback</strong>. </p><p>That&#8217;s what truly helps determine how well you're going to do against those performance metrics.</p><p>Not saying &#8220;We are a low to medium performer. Now go and improve the metrics!&#8221;</p><p>I think the issue is that a lot of organizations emphasize the metrics because it's something that can sit on an executive dashboard  </p><p>But the true reason we have metrics is to help drive conversations.</p><p>Through those conversations, we drive improvement.</p><p>That&#8217;s important because currently an unfortunately noticeable amount of organizations are doing this according to Nathen:</p><blockquote><p>I've seen organizations [where it&#8217;s like]: &#8220;Oh, we're going to do DORA. Here's my dashboard. Okay, we're done. We've done DORA. I can look at these metrics on a dashboard.&#8221;  </p><p>That doesn't change anything. </p><p>We have to go the step further and put those metrics into action.</p></blockquote><p>We should be treating the metrics as a kind of compass on a map. </p><p>You can use those metrics to help orient yourself and understand, &#8220;Where are we heading?&#8221;.</p><p>But then you have to choose how are you going to make progress toward whatever your goal is.</p><p>The capabilities enabled by the DORA framework should help answer questions like:</p><ul><li><p>Where are our bottlenecks?</p></li><li><p>Where are our constraints?</p></li><li><p>Do we need to do some improvement work as a team?</p></li></ul><p>We also talked about the SPACE framework, which is a follow-on tool from DORA metrics. </p><p>It is a framework for understanding developer productivity.  </p><p>It encourages teams or organizations to<strong> look at five dimensions when trying to measure something from a productivity perspective</strong>.</p><p>It stands for:</p><ul><li><p><strong>S</strong> &#8212; satisfaction and well-being</p></li><li><p><strong>P</strong> &#8212; performance</p></li><li><p><strong>A</strong> &#8212; activity</p></li><li><p><strong>C</strong> &#8212; communication and collaboration</p></li><li><p><strong>E</strong> &#8212; efficiency and flow</p></li></ul><p>What the SPACE framework recommends is that you</p><p>First, pick metrics from two to three of those five categories. </p><p>(You don't need a metric from every one of those five but find something that works well for your team.)</p><p>Then write down those metrics and start measuring them. </p><p>Here&#8217;s the interesting thing: <strong>DORA is an implementation of SPACE.</strong> </p><p>You can correlate each metric with the SPACE acronym!</p><ul><li><p>Lead time for changes is a measure of <strong>E</strong>fficiency and flow</p></li><li><p>Deployment frequency is an <strong>A</strong>ctivity</p></li><li><p>Change fail rate is about <strong>P</strong>erformance.</p></li><li><p>Failed deployment recovery time is about <strong>E</strong>fficiency and flow</p></li></ul><p>Keep in mind that SPACE itself has no metrics. </p><p>It is a framework for identifying metrics.</p><p>Nathen reiterated that you can't use the space metrics because there is no such thing. </p><p>I mentioned earlier how DORA is a means of identifying the capabilities that can improve the metrics.</p><p>These can be technical practices like using continuous integration.</p><p>But they can also be capabilities like collaboration and communication. </p><p>As an example, you might look at what your change approval process looks like. </p><p>You might look at how collaboration and communication have failed when you&#8217;ve had to send changes off to an external approval board like a CAB (change approval board).</p><p>DORA&#8217;s research backs the above up:</p><blockquote><p>What our research has shown through collecting data over the years, is that while they do exist on the whole, <strong>an external change approval body will slow you down.</strong></p><p>That's no surprise. So <strong>your change lead time is going to increase, your deployment frequency will decrease</strong>.  </p><p>But, <strong>at best, they have zero impact on your change fail rate.</strong> In most cases, they have a negative impact on your change fail rate. So you're failing more often.</p></blockquote><p>It goes back to the idea of smaller changes, faster feedback, and being able to validate that. Building in audit controls and so forth.</p><p>This is something that reliability-focused engineers should be able to help with because one of the things Sebastian and I talk about a lot is embracing and managing risk effectively and not trying to mitigate it through stifling measures like CABs. </p><p>In short,  DORA and software reliability are not mutually exclusive concepts.</p><p>They're certainly in the same universe.</p><p>Nathen went as far as to say that <strong>some SRE practices necessarily get a little bit deeper than sort of the capability level that DORA has</strong> and provide even more sort of specific guidance on how to do things.</p><p>He clarified a doubt I had because a lot of people have argued with me (mainly at conferences) that DORA is this thing that developers do, earlier in the SDLC.</p><p>And then SRE is completely different because it focuses on the production side. </p><div class="pullquote"><p>The worst possible situation could be turning to developers and saying, &#8220;These 2 throughput metrics, they&#8217;re yours. Make sure they go up no matter what,&#8221; and then turn to our SREs and say &#8220;Those stability metrics, they're yours. Make sure they stay good&#8221; </p><p>All that does is put these false incentives in place and we're just fighting against each other.</p></div><p>We talked a little more about the future of DORA in our podcast episode (player/link right at the top of this post) if you want to hear about that.</p><p>Here are some useful links from Nathen for further research:</p><p><a href="https://dora.community/">DORA online community of practice</a></p><p><a href="https://dora.dev">DORA homepage</a></p><p><a href="https://queue.acm.org/detail.cfm?id=3454124">[Article] The SPACE of Developer Productivity</a></p><p><a href="https://linktr.ee/nathenharvey">Nathen Harvey's Linktree</a></p>]]></content:encoded></item></channel></rss>