Obviously, I'm fascinated by systems. Bruce Schneier recently posted a link to a very interesting white paper by Richard Cook of the University of Chicago, whose research interests include, "the study of human error, the role of technology in human expert performance, and patient safety."
Here's his list, which has some interesting parallels to my own book chapter on "System Design Principles" (you can watch a free video for that chapter here).
- Complex systems are intrinsically hazardous systems.
- Complex systems are heavily and successfully defended against failure.
- Catastrophe requires multiple failures –single point failures are not enough..
- Complex systems contain changing mixtures of failures latent within them.
- Complex systems run in degraded mode.
- Catastrophe is always just around the corner.
- Post-accident attribution accident to a ‘root cause’ is fundamentally wrong.
- Hindsight biases post-accident assessments of human performance.
- Human operators have dual roles: as producers & as defenders against failure.
- All practitioner actions are gambles.
- Actions at the sharp end resolve all ambiguity.
- Human practitioners are the adaptable element of complex systems.
- Human expertise in complex systems is constantly changing
- Change introduces new forms of failure.
- Views of ‘cause’ limit the effectiveness of defenses against future events.
- Safety is a characteristic of systems and not of their components
- People continuously create safety.
- Failure free operations require experience with failure.
I've often felt that systems take on a life of their own once assembled, and I found his paper, which is only a few pages long, a fascinating read. You can download it here.