TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

Block Storage Issues Across All Regions: Incident Report for DigitalOcean

70 点作者 juliand超过 5 年前

9 条评论

redredrobot超过 5 年前
That&#x27;s not seriously the real engineering portmortem is it? That looks more like a &#x27;Resolved the issue&#x27; update - it is way too shallow and vague.<p>If this was sent out at AWS as a COE (postmortem), it would be ripped apart - it is not going to satisfy anyone reading it that they should have confidence this class of failure isn&#x27;t going to happen again. It looks like they haven&#x27;t even identified the root cause(s) of the failure...
评论 #21169743 未加载
unilynx超过 5 年前
Unfortunately, this doesn&#x27;t explain why this change was applied to 5 datacenters at the same time, or why they didn&#x27;t do that but it still affected 5 of them.<p>I would have liked to hear more about how they are going to reduce the blast radius of such a change, because it sounds like something that could have been deployed to a single datacenter first
评论 #21168171 未加载
ebcode超过 5 年前
Adding my voice to the chorus here as a DO customer. This 188-word &quot;postmortem&quot; gives postmortems a bad name. I would like to know the details of the &quot;network configuration change&quot; and <i>why</i> it &quot;caused incompatibilities&quot;. And also <i>how</i> you will ensure that this particular failure will not re-occur.<p>Trust and transparency are the currencies of the internet, in the same way that cigarettes and contraband are the currencies in prison. This post is worth approx. a 1&#x2F;2 smoked cigarette.
caiobegotti超过 5 年前
It&#x27;s because of &quot;reports&quot; like these I didn&#x27;t feel like staying as their customer. Whomever is in charge of [limiting scope and wording of] these reports should listen to a few things in private at their HQs.
评论 #21168862 未加载
notacoward超过 5 年前
This happened, and was apparently resolved, on Monday. Am I the only one who wonders if it was released on a Saturday to minimize the amount of attention&#x2F;commentary it would get (e.g. here)?
lucb1e超过 5 年前
Aside from a timeline and some promises, this is the full post-mortem analysis of what happened:<p>&gt; The outage was triggered as a result of a networking configuration change on the Block Storage clusters to improve handling packet loss scenarios. The new setting caused incompatibilities<p>So that doesn&#x27;t tell us very much about the cause (&quot;a networking configuration change&quot;) nor the effect (&quot;incompatibilities&quot;).
mlthoughts2018超过 5 年前
&gt; “The outage was triggered as a result of a networking configuration change on the Block Storage clusters to improve handling packet loss scenarios. The new setting caused incompatibilities, which led to network interfaces becoming unavailable.”<p>I wonder if this just means someone changed an MTU configuration and it led to tons of fragmentation in different components of the network, and especially for any large file transfer making things timeout constantly to render an outage. Just a wild guess, but I’ve seen this happen with in-house datacenters before, so perhaps.
exabrial超过 5 年前
Nice details in the report. Come on DO, this was the most annoyingly PC, lawyer sanitized, non-scientific RCA anyone here has ever read. Here&#x27;s the BLUF line: Don&#x27;t cause issues. That causes problems.
adreamingsoul超过 5 年前
Sigh, I really like DO but this just shows how much they still need to learn about operations. AWS does “raise” the bar for that, but unfortunately you can only really see that from the inside.