TE
科技回声
首页24小时热榜最新最佳问答展示工作
GitHubTwitter
首页

科技回声

基于 Next.js 构建的科技新闻平台,提供全球科技新闻和讨论内容。

GitHubTwitter

首页

首页最新最佳问答展示工作

资源链接

HackerNews API原版 HackerNewsNext.js

© 2025 科技回声. 版权所有。

January 28th Incident Report

451 点作者 Oompa超过 9 年前

27 条评论

eric_h超过 9 年前
&gt; One of the biggest customer-facing effects of this delay was that status.github.com wasn&#x27;t set to status red until 00:32am UTC, eight minutes after the site became inaccessible. We consider this to be an unacceptably long delay, and will ensure faster communication to our users in the future.<p>Amazon could learn a thing or two from Github in terms of understanding customer expectations.
评论 #11031069 未加载
评论 #11032830 未加载
bosdev超过 9 年前
There&#x27;s no mention of why they don&#x27;t have redundant systems in more than one datacenter. As they say, it is unavoidable to have power or connectivity disruptions in a datacenter. This is why reliable configurations have redundancy in another datacenter elsewhere in the world.
评论 #11030063 未加载
评论 #11031046 未加载
评论 #11030806 未加载
评论 #11030049 未加载
danielvf超过 9 年前
For all that work to be done in just two hours is amazing, especially with degraded internal tools, and both hardware and ops teams working simultaneously.
评论 #11030187 未加载
评论 #11030624 未加载
评论 #11032426 未加载
DarkTree超过 9 年前
I don&#x27;t know enough about server infrastructure to comment on whether or not Github was adequately prepared or reacted appropriately to fix the problem.<p>But wow it is refreshing to hear a company take full responsibility and own up to a mistake&#x2F;failure and apologize for it.<p>Like people, all companies will make mistakes and have momentary problems. It&#x27;s normal. So own up to it and learn how to avoid the mistake in the future.
评论 #11031081 未加载
pedalpete超过 9 年前
Does Github run anything like Netflix Simbian Army against it&#x27;s services? As a company by engineers for engineers with the scale that github has reached, I&#x27;m a bit surprised they are lacking a bit more redundancy. Though they may not need the uptime of netflix, an outage of more than a few minutes on github could affect businesses that rely on the service.
评论 #11030967 未加载
评论 #11032052 未加载
评论 #11030335 未加载
onetwotree超过 9 年前
Every time I read about a massive systems failure, I think of Jurassic Park and am mildly grateful that the velociraptor padock wasn&#x27;t depending on the systems operation.
评论 #11030427 未加载
评论 #11030429 未加载
mjevans超过 9 年前
This just shows how difficult it is to avoid hidden dependencies without a complete, cleanly isolated, testing environment of sufficient scale to replicate production operations and do strange system fault scenarios somewhere that won&#x27;t kill production.
评论 #11030065 未加载
评论 #11030081 未加载
viraptor超过 9 年前
&gt; ... Updating our tooling to automatically open issues for the team when new firmware updates are available will force us to review the changelogs against our environment.<p>That&#x27;s an awesome idea. I wish all companies published the firmware releases in simple rss feeds, so everyone could easily integrate them with their trackers.<p>(If someone&#x27;s bored, that may be a nice service actually ;) )
评论 #11030763 未加载
评论 #11030325 未加载
matt_wulfeck超过 9 年前
&gt; Remote access console screenshots from the failed hardware showed boot failures because the physical drives were no longer recognized.<p>I&#x27;m getting flashbacks. All of the servers in the DC reboot and NONE of them come online. No network or anything. Even remotely rebooting them again we had nothing. Finally getting a screen (which is a pain in itself) we saw they were all stuck on a grub screen. Grub detected an error and decided not to boot automatically. Needless to say we patched grubbed and removed this &quot;feature&quot; promptly!
gaius超过 9 年前
You can very clearly see two kinds of people posting on this thread: those who have actually dealt with failures of complex distributed systems, and those who think it&#x27;s easy.
Animats超过 9 年前
<i>&quot;We identified the hardware issue resulting in servers being unable to view their own drives after power-cycling as a known firmware issue that we are updating across our fleet.&quot;</i><p>Tell us which vendor shipped that firmware, so everyone else can stop buying from them.
评论 #11032286 未加载
merqurio超过 9 年前
I feel it was good incident for the Open Source community, to see how dependent we are on GitHub today. I feel sad whenever I see another large project like Python moving to GitHub, a closed-sourced company. I know, GitLab is there as an alternative, but I would love to see all the big Open Source projects putting pressure over GitHub to make them open their source code, as right they are big player in open source, like it or not.
评论 #11030259 未加载
评论 #11030245 未加载
评论 #11030084 未加载
评论 #11030736 未加载
rqebmm超过 9 年前
It must be nice to know that the majority of your customers are familiar enough with the nature of your work that they&#x27;ll actually understand a relatively complex issue like this. Almost by definition, we&#x27;ve all been there.
dsmithatx超过 9 年前
If only Bitbucket could give such comprehensive reports. A few months back outages seemed almost daily. Things are more stable now. I hope for the long term.
评论 #11030128 未加载
guelo超过 9 年前
Weird that they didn&#x27;t say what caused the power outage and what the mitigations are for that.
评论 #11030940 未加载
评论 #11033633 未加载
评论 #11031168 未加载
tmsh超过 9 年前
&gt; Over the past week, we have devoted significant time and effort towards understanding the nature of the cascading failure which led to GitHub being unavailable for over two hours.<p>I don&#x27;t mean to be blasphemous, but from a high level, is the performance issues with Ruby (and Rails) that necessitate close binding with Redis (i.e., lots of caching) part of the issue?<p>It sounds like the fundamental issue is not Ruby, nor Redis, but the close coupling between them. That&#x27;s sort of interesting.
评论 #11031663 未加载
评论 #11031206 未加载
评论 #11030917 未加载
cognivore超过 9 年前
Um, work from your local cache for a few hours? It&#x27;s that the one of the main reasons for git?
评论 #11032900 未加载
timiblossom超过 9 年前
If you use Redis, you should try out Dynomite at <a href="http:&#x2F;&#x2F;github.com&#x2F;Netflix&#x2F;Dynomite" rel="nofollow">http:&#x2F;&#x2F;github.com&#x2F;Netflix&#x2F;Dynomite</a>. It can provide HA for Redis servers
rurounijones超过 9 年前
I would have expected there to be a notification system owned by the DC that literally send an email to clients saying &quot;Power blipped &#x2F; failed&quot;.<p>That would have given them immediate co text and not wasting time on DDOS protection
spydum超过 9 年前
So, while it sounds like they have reasonable HA, they fell down on DR. unrelated, I could not comprehend what this means?..: technicians to bring these servers back online by draining the flea power to bring<p>Flea power?
评论 #11030691 未加载
tonylxc超过 9 年前
TL;DR: &quot;We don’t believe it is possible to fully prevent the events that resulted in a large part of our infrastructure losing power, ...&quot;<p>This doesn&#x27;t sound very good.
评论 #11032174 未加载
评论 #11030592 未加载
评论 #11031986 未加载
评论 #11030844 未加载
评论 #11030594 未加载
mattdeboard超过 9 年前
Anyone have a link to a description of the firmware bug that caused the disk-mounting failure after power was restored?
评论 #11032265 未加载
TazeTSchnitzel超过 9 年前
&gt; We had inadvertently added a hard dependency on our Redis cluster being available within the boot path of our application code.<p>I seem to recall a recent post on here about how you shouldn&#x27;t have such hard dependencies. It&#x27;s good advice.<p>Incidentally, this type of dependency is unlikely to happen if you have a shared-nothing model (like PHP has, for instance), because in such a system each request is isolated and tries to connect on its own.
totally超过 9 年前
&gt; Because we have experience mitigating DDoS attacks, our response procedure is now habit and we are pleased we could act quickly and confidently without distracting other efforts to resolve the incident.<p>The thing that fixed the last problem doesn&#x27;t always fix the current problem.
评论 #11031310 未加载
swrobel超过 9 年前
Anyone got a good tl;dr version?
评论 #11030244 未加载
评论 #11030579 未加载
评论 #11030258 未加载
评论 #11032526 未加载
评论 #11030433 未加载
评论 #11031230 未加载
jargonless超过 9 年前
What is this &quot;HA&quot; jargon?<p>I would STFW, but searching for &quot;HA&quot; isn&#x27;t helpful.
评论 #11032431 未加载
评论 #11030439 未加载
评论 #11030446 未加载
评论 #11031300 未加载
评论 #11030965 未加载
评论 #11030943 未加载
评论 #11030640 未加载
评论 #11030457 未加载
julesbond007超过 9 年前
I seriously doubt this version of the story. While it&#x27;s possible for several hardware&#x2F;firmware to fail in all your datacenters, for them to fail at the same time is highly unlikely. This may just be a PR spin to think they&#x27;re not vulnerable to security attacks.<p>While this was happening at Github, I noticed several other companies facing that same issue at the same time. Atlassian was down for the most part. It could have been an issue with the service github uses, but they won&#x27;t admit that. Notice they never said what the firmware issue was instead blaming it on &#x27;hardware&#x27;.<p>I think they should be transparent with people about such vulnerability, but I suspect they would never say so because then they would lose revenue.<p>Here on my blog I talked about this issue: <a href="http:&#x2F;&#x2F;julesjaypaulynice.com&#x2F;simple-server-malicious-attacks&#x2F;" rel="nofollow">http:&#x2F;&#x2F;julesjaypaulynice.com&#x2F;simple-server-malicious-attacks...</a><p>I think it was some ddos campaign going on over the web.
评论 #11032275 未加载