Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

In the end - action matters. Somebody didn’t put the bolts back in.

Yes - zooming out it important and ultimately where actionable remediation can be applied - but blame is due where blame is due: somebody fucked up at work and it almost brought down a plane.



This is absolutely incorrect. It runs counter to every high functioning safety culture I've ever encountered.

The system allowed the human to take the incorrect action. If your intern destroys your prod database, it's because you failed to restrict access to the prod database. The remediation to "my intern is capable of destroying my prod database" is not "fire the intern" it's "restrict access to the prod db".

Even the best trained humans will make errors. They will make errors stochastically. Your systemic safety checks will guard against those errors becoming problems. If your safety culture requires all humans to be flawless 100% of the time, your safety culture sucks.

So no, this isn't a fault with a human. Because this was a possible error, it was inevitable that at some point a human would make that error. Because humans never operate without errors for extended periods of time.


Modern safety analysis acknowledges that humans are fallible, and they are generally acting in a good faith way to try and do their jobs correctly within a given system they are operating in.

That's why these reports tend to suggest corrective actions to the parts of the system that didn't work properly. Even in a perfectly functioning safety culture, an employee can make a mistake and forget to install the bolts. A functioning safety system has safeguards in place to ensure that mistake is found and corrected.


Super underrated point - and one that I am not sure the general public always keeps top-of-mind, as human imperfection should be the default assumption. The whole system of air travel is designed so that wherever possible, multiple f*ck-ups can occur and not result in a catastrophe. The success of people involved with anything touching on aviation safety is best measured as in "how many f*ck-ups can occur in the same episode and have everyone still walk away alive?" If you can get that number up to 3, 4 complete idiotic screw-ups one after the other, and the people still live, you've really achieved something great.


There's a reason why Murphy's Law is so commonly acknowledged, though. When you've got a process like this that gets repeated over and over by a bunch of different people, you simply must recognize that that, if it's possible for someone to fuck up, then somebody will fuck up.

And a relatively straightforward corollary of that reality is that, when somebody fucks up, putting too much personal blame on them is pointless. If it weren't them, it would have been somebody else.

In other words, this "blame is due where blame is due" framing is mostly useful as a cop-out excuse that helps incompetent managers who've been skimping on quality controls and failsafes to shift the blame away from where it really belongs.


> There's a reason why Murphy's Law is so commonly acknowledged, though.

In particular, the original formulation of Murphy's Law. The folk version has morphed into "anything that can go wrong, will go wrong". But the original was "If there are two or more ways to do something and one of those results in a catastrophe, then someone will do it that way".


Yes and, IMHO: docs, procedures, checklists, etc. strive to mitigate the challenge of assumed knowledge. It's a wicked hard problem.


In the end, identifying where you can usefully take action to reduce the chances of something similar happen in the future is far more useful than assigning blame.


Yes! It's basically better to take all screw-up(s) and make their recurrence the assumption. 'Given people will forget to replace bolts how can we best make it so the plane cannot exit the factory without the bolts in place?'


Assigning blame is often the antithesis of safety.

In aviation and other safety-critical fields, we use a just culture approach — not to avoid accountability, but to ensure that learning and prevention come first.

https://en.wikipedia.org/wiki/Just_culture


Others already said it but since I'm the person you responded to, I'll reiterate that my suggestion was only about reordering the sequence of that sentence for better clarity, not about placing blame on individuals over process. When a failure can cause serious consequences including killing people, proper system design should never even permit a single point of failure to exist, especially one relying on humans to always perform correctly and completely. Even well-trained, highly-conscientious humans can make a mistake. While these people should have received better training as well as comprehensive sequential checklists, a good system design will have critical failure points such as this each verified and signed off by a separate inspector.

The problem with a culture which prioritizes "blame is due where blame is due" is it can cause people to not report near-misses and other gaps as well as cover-up actual mistakes. The shift in the U.S. from blaming (and penalizing) occasional pilot lapses to a more 'blameless' default mode was controversial but has now clearly demonstrated that it nets better overall safety.


Have you read Donald Norman's Design of Everyday Things?




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: