Four Layers of Protection

The Swiss Cheese Model, conceptualized by James Reason, fundamentally defined modern risk management by illustrating how layered defenses interact with active and latent failures to prevent or enable adverse events. This framework underpins the Four Layers of Protection, a systematic approach to mitigating risks across industries. By integrating Reason’s Theory of Active and Latent Failures with modern adaptations like resilience engineering, organizations can create robust, adaptive systems.

The Swiss Cheese Model and Reason’s Theory: A Foundation for Layered Defenses

Reason’s Theory distinguishes between active failures (immediate errors by frontline personnel) and latent failures (systemic weaknesses in design, management, or culture). The Swiss Cheese Model visualizes these failures as holes in successive layers of defense. When holes align, hazards penetrate the system. For example:

  • In healthcare, a mislabeled specimen (active failure) might bypass defenses if staff are overworked (latent failure) and barcode scanners malfunction (technical failure).
  • In aviation, a pilot’s fatigue-induced error (active) could combine with inadequate simulator training (latent) and faulty sensors (technical) to cause a near-miss.

This model emphasizes that no single layer is foolproof; redundancy and diversity across layers are critical.

Four Layers of Protection:

While industries tailor layers to their risks, four core categories form the backbone of defense:

LayerKey PrinciplesIndustry Example
Inherent DesignEliminate hazards through intrinsic engineering (e.g., fail-safe mechanisms)Pharmaceutical isolators preventing human contact with sterile products
ProceduralAdministrative controls: protocols, training, and auditsISO 27001’s access management policies for data security
TechnicalAutomated systems, physical barriers, or real-time monitoringSafety Instrumented Systems (SIS) shutting down chemical reactors during leaks
OrganizationalCulture, leadership, and resource allocation sustaining qualityJust Culture frameworks encouraging transparent incident reporting

Industry Applications

1. Healthcare: Reducing Surgical Infections

  • Inherent: Antimicrobial-coated implants resist biofilm formation.
  • Procedural: WHO Surgical Safety Checklists standardize pre-operative verification.
  • Technical: UV-C robots disinfect operating rooms post-surgery.
  • Organizational: Hospital boards prioritizing infection prevention budgets.

2. Information Security: Aligning with ISO/IEC 27001

  • Inherent: Encryption embedded in software design (ISO 27001 Annex A.10).
  • Procedural: Regular penetration testing and access reviews (Annex A.12).
  • Technical: Intrusion detection systems (Annex A.13).
  • Organizational: Enterprise-wide risk assessments and governance (Annex A.5).

3. Biotech Manufacturing: Contamination Control

  • Inherent: Closed-system bioreactors with sterile welders.
  • Procedural: FDA-mandated Contamination Control Strategies (CCS).
  • Technical: Real-time viable particle monitoring with auto-alerts.
  • Organizational: Cross-functional teams analyzing trend data to preempt breaches.

Contamination Control and Layers of Controls Analysis (LOCA)

In contamination-critical industries, a Layers of Controls Analysis (LOCA) evaluates how failures in one layer impact others. For example:

  1. Procedural Failure: Skipping gowning steps in a cleanroom.
  2. Technical Compromise: HEPA filter leaks due to poor maintenance.
  3. Organizational Gap: Inadequate staff training on updated protocols.

LOCA reveals that latent organizational failures (e.g., insufficient training budgets) often undermine technical and procedural layers. LOCA ties contamination risks to systemic resource allocation, not just frontline errors.

Integration with ISO/IEC 27001

ISO/IEC 27001, the international standard for information security, exemplifies layered risk management:

ISO 27001 Control (Annex A)Corresponding LayerExample
A.8.3 (Information labeling)ProceduralClassifying data by sensitivity
A.9.4 (Network security)TechnicalFirewalls and VPNs
A.11.1 (Physical security)Inherent/TechnicalBiometric access to server rooms
A.5.1 (Policies for IS)OrganizationalBoard-level oversight of cyber risks

This alignment ensures that technical safeguards (e.g., encryption) are reinforced by procedural (e.g., audits) and organizational (e.g., governance) layers, mirroring the Swiss Cheese Model’s redundancy principle.

Resilience Engineering: Evolving the Layers

Resilience engineering moves beyond static defenses, focusing on a system’s capacity to anticipate, adapt, and recover from disruptions. It complements the Four Layers by adding dynamism:

Traditional LayerResilience Engineering ApproachExample
Inherent DesignBuild adaptive capacity (e.g., modular systems)Pharmaceutical plants with flexible cleanroom layouts
ProceduralDynamic procedures adjusted via real-time dataAI-driven prescribing systems updating dosage limits during shortages
TechnicalSelf-diagnosing systems with graceful degradationPower grids rerouting energy during cyberattacks
OrganizationalLearning cultures prioritizing near-miss reportingAviation safety databases sharing incident trends globally

Challenges and Future Directions

While the Swiss Cheese Model remains influential, critics argue it oversimplifies complex systems where layers interact unpredictably. For example, a malfunctioning algorithm (technical) could override procedural safeguards, necessitating organizational oversight of machine learning outputs.

Future applications will likely integrate:

  • Predictive Analytics: Leverages advanced algorithms, machine learning, and vast datasets to forecast future risks and opportunities, transforming risk management from a reactive to a proactive discipline. By analyzing historical and real-time data, predictive analytics identifies patterns and anomalies that signal potential threats—such as equipment failures or contamination events —enabling organizations to anticipate and mitigate risks before they escalate. The technology’s adaptability allows it to integrate internal and external data sources, providing dynamic, data-driven insights that support better decision-making, resource allocation, and compliance monitoring. As a result, predictive analytics not only enhances operational resilience and efficiency but also reduces costs associated with failures, recalls, or regulatory breaches, making it an indispensable tool for modern risk and quality management.
  • Human-Machine Teaming: Integrates human cognitive flexibility with machine precision to create collaborative systems that outperform isolated human or machine efforts. By framing machines as adaptive teammates rather than passive tools, HMT enables dynamic task allocation. Key benefits include accelerated decision-making through AI-driven data synthesis, reduced operational errors via automated safeguards, and enhanced resilience in complex environments. However, effective HMT requires addressing challenges such as establishing bidirectional trust through explainable AI, aligning ethical frameworks for accountability, and balancing autonomy levels through risk-categorized architectures. As HMT evolves, success hinges on designing systems that leverage human intuition and machine scalability while maintaining rigorous quality protocols.
  • Epistemic Governance: The processes through which actors collectively shape perceptions, validate knowledge, and steer decision-making in complex systems, particularly during crises. Rooted in the dynamic interplay between recognized reality (actors’ constructed understanding of a situation) and epistemic work (efforts to verify, apply, or challenge knowledge), this approach emphasizes adaptability over rigid frameworks. By appealing to norms like transparency and scientific rigor, epistemic governance bridges structural frameworks (e.g., ISO standards) and grassroots actions, enabling systems to address latent organizational weaknesses while fostering trust. It also confronts power dynamics in knowledge production, ensuring marginalized voices inform policies—a critical factor in sustainability and crisis management where equitable participation shapes outcomes. Ultimately, it transforms governance into a reflexive practice, balancing institutional mandates with the agility to navigate evolving threats.

Conclusion

The Four Layers of Protection, rooted in Reason’s Swiss Cheese Model, provide a versatile framework for managing risks—from data breaches to pharmaceutical contamination. By integrating standards and embracing resilience engineering, organizations can transform static defenses into adaptive systems capable of navigating modern complexities. As industries face evolving threats, the synergy between layered defenses and dynamic resilience will define the next era of risk management.

Applying a Layers of Controls Analysis to Contamination Control

Layers of Controls Analysis (LOCA)

Layers of Controls Analysis (LOCA) provides a comprehensive framework for evaluating multiple layers of protection to reduce and manage operational risks. By examining both preventive and mitigative control measures simultaneously, LOCA allows organizations to gain a holistic view of their risk management strategy. This approach is particularly valuable in complex operational environments where multiple safeguards and protective systems are in place.

One of the key strengths of LOCA is its ability to identify gaps in protection. By systematically analyzing each layer of control, from basic process design to emergency response procedures, LOCA can reveal areas where additional safeguards may be necessary. This insight is crucial for guiding decisions on implementing new risk reduction measures or enhancing existing ones. The analysis helps organizations prioritize their risk management efforts and allocate resources more effectively.

Furthermore, LOCA provides a structured way to document and justify risk reduction measures. This documentation is invaluable for regulatory compliance, internal audits, and continuous improvement initiatives. By clearly outlining the rationale behind each protective layer and its contribution to overall risk reduction, organizations can demonstrate due diligence in their safety and risk management practices.

Another significant advantage of LOCA is its promotion of a holistic view of risk control. Rather than evaluating individual safeguards in isolation, LOCA considers the cumulative effect of multiple protective layers. This approach recognizes that risk reduction is often achieved through the interaction of various control measures, ranging from engineered systems to administrative procedures and emergency response capabilities.

By building on other risk assessment techniques, such as Hazard and Operability (HAZOP) studies and Fault Tree Analysis, LOCA provides a more complete picture of protection systems. It allows organizations to assess the effectiveness of their entire risk management strategy, from prevention to mitigation, and ensures that risks are reduced to an acceptable level. This comprehensive approach is particularly valuable in high-hazard industries where the consequences of failures can be severe.

LOCA combines elements of two other methods – Layers of Protection Analysis (LOPA) and Layers of Mitigation Analysis (LOMA).

Layers of Protection Analysis

To execute a Layers of Protection Analysis (LOPA), follow these key steps:

Define the hazardous scenario and consequences:

  • Clearly identify the hazardous event being analyzed
  • Determine the potential consequences if all protection layers fail

Identify initiating events:

  • List events that could trigger the hazardous scenario
  • Estimate the frequency of each initiating event

Identify Independent Protection Layers (IPLs):

  • Determine existing safeguards that can prevent the scenario
  • Evaluate if each safeguard qualifies as an IPL (independent, auditable, effective)
  • Estimate the Probability of Failure on Demand (PFD) for each IPL

Identify Conditional Modifiers:

  • Determine factors that impact scenario probability (e.g. occupancy, ignition probability)
  • Estimate probability for each modifier

Calculate scenario frequency:

  • Multiply initiating event frequency by PFDs of IPLs and conditional modifiers

Compare to risk tolerance criteria:

  • Determine if calculated frequency meets acceptable risk level
  • If not, identify need for additional IPLs

Document results:

  • Record all assumptions, data sources, and calculations
  • Summarize findings and recommendations

Review and validate:

  • Have results reviewed by subject matter experts
  • Validate key assumptions and data inputs

Key aspects for successful LOPA execution

  • Use a multidisciplinary team
  • Ensure independence between IPLs
  • Be conservative in estimates
  • Focus on prevention rather than mitigation
  • Consider human factors in IPL reliability
  • Use consistent data sources and methods

Layers of Mitigation Analysis

LOMA focuses on analyzing reactionary or mitigative measures, as opposed to preventive measures.

A LOCA as part of Contamination Control

A Layers of Controls Analysis (LOCA) can be effectively applied to contamination control in biotech manufacturing by systematically evaluating multiple layers of protection against contamination risks.

To determine potential hazards when conducting a Layer of Controls Analysis (LOCA) for contamination control in biotech, follow these steps:

  1. Form a multidisciplinary team: Include members from manufacturing, quality control, microbiology, engineering, and environmental health & safety to gain diverse perspectives.
  2. Review existing processes and procedures: Examine standard operating procedures, experimental protocols, and equipment manuals to identify potential risks associated with each step.
  3. Consider different hazard types. Focus on categories like:
    • Biological hazards (e.g., microorganisms, cell lines)
    • Chemical hazards (e.g., toxic substances, flammable materials)
    • Physical hazards (e.g., equipment-related risks)
    • Radiological hazards (if applicable)
  4. Analyze specific contamination hazard types for biotech settings:
    • Mix-up: Materials used for the wrong product
    • Mechanical transfer: Cross-contamination via personnel, supplies, or equipment
    • Airborne transfer: Contaminant movement through air/HVAC systems
    • Retention: Inadequate removal of materials from surfaces
    • Proliferation: Potential growth of biological agents
  5. Conduct a process analysis: Break down each laboratory activity into steps and identify potential hazards at each stage.
  6. Consider human factors: Evaluate potential for human error, such as incorrect handling of materials or improper use of equipment.
  7. Assess facility and equipment: Examine the layout, containment measures, and equipment condition for potential hazards.
  8. Review past incidents and near-misses: Analyze previous safety incidents or close calls to identify recurring or potential hazards.
  9. Consult relevant guidelines and regulations: Reference industry standards, biosafety guidelines, and regulatory requirements to ensure comprehensive hazard identification.
  10. Use brainstorming techniques: Encourage team members to think creatively about potential hazards that may not be immediately obvious.
  11. Evaluate hazards at different scales: Consider how hazards might change as processes scale up from research to production levels.
  • Facility Design and Engineering Controls
    • Cleanroom design and classification
    • HVAC systems with HEPA filtration
    • Airlocks and pressure cascades
    • Segregated manufacturing areas
  • Equipment and Process Design
    • Closed processing systems
    • Single-use technologies
    • Sterilization and sanitization systems
    • In-line filtration
  • Operational Controls
    • Aseptic techniques and procedures
    • Environmental monitoring programs
    • Cleaning and disinfection protocols
    • Personnel gowning and hygiene practices
  • Quality Control Measures
    • In-process testing (e.g., bioburden, endotoxin)
    • Final product sterility testing
    • Environmental monitoring data review
    • Batch record review
  • Organizational Controls
    • Training programs
    • Standard operating procedures (SOPs)
    • Quality management systems
    • Change control processes
  1. Evaluate reliability and capability of each control:
    • Review historical performance data for each control measure
    • Assess the control’s ability to prevent or detect contamination
    • Consider the control’s consistency in different operating conditions
  2. Consider potential failure modes:
    • Conduct a Failure Mode and Effects Analysis (FMEA) for each control
    • Identify potential ways the control could fail or be compromised
    • Assess the likelihood and impact of each failure mode
  3. Evaluate human factors:
    • Assess the complexity and potential for human error in each control
    • Review training effectiveness and compliance with procedures
    • Consider ergonomics and usability of equipment and systems
  4. Analyze technology effectiveness:
    • Evaluate the performance of automated systems and equipment
    • Assess the reliability of monitoring and detection technologies
    • Consider the integration of different technological controls
  1. Quantify risk reduction:
    • Assign risk reduction factors to each layer based on its effectiveness
    • Use a consistent scale (e.g., 1-10) to rate each control’s risk reduction capability
    • Calculate the cumulative risk reduction across all layers
  2. Assess interdependencies between layers:
    • Identify any controls that rely on or affect other controls
    • Evaluate how failures in one layer might impact the effectiveness of others
    • Consider potential common mode failures across multiple layers
  3. Review control performance metrics:
    • Analyze trends in environmental monitoring data
    • Examine out-of-specification results and their root causes
    • Assess the frequency and severity of contamination events
  1. Determine acceptable risk levels:
    • Define your organization’s risk tolerance for contamination events
    • Compare current risk levels against these thresholds
  2. Identify gaps:
    • Highlight areas where current controls fall short of required protection
    • Note processes or areas with insufficient redundancy
  3. Propose improvements:
    • Suggest enhancements to existing controls
    • Recommend new control measures to address identified gaps
  4. Prioritize actions:
    • Rank proposed improvements based on risk reduction potential and feasibility
    • Consider cost-benefit analysis for major changes
  5. Seek expert input:
    • Consult with subject matter experts on proposed improvements
    • Consider third-party assessments for critical areas
  6. Plan for implementation:
    • Develop action plans for addressing identified gaps
    • Assign responsibilities and timelines for improvements
  1. Document and review:
  1. Implement continuous monitoring and review:
  2. Develop a holistic CCS document:
    • Describe overall contamination control approach
    • Detail how different controls work together
    • Include risk assessments and rationales
  3. Establish governance and oversight:
    • Create a cross-functional CCS team
    • Define roles and responsibilities
    • Implement a regular review process
  4. Integrate with quality systems:
    • Align CCS with existing quality management processes
    • Ensure change control procedures consider CCS impact
  5. Provide comprehensive training:
    • Train all personnel on CCS principles and practices
    • Implement contamination control ambassador program
  1. Implement regular review cycles:
    • Schedule periodic reviews of the LOCA (e.g., annually or bi-annually)
    • Involve a cross-functional team including quality, manufacturing, and engineering
  2. Analyze trends and data:
    • Review environmental monitoring data
    • Examine out-of-specification results and their root causes
    • Assess the frequency and severity of contamination events
  3. Identify improvement opportunities:
    • Use gap analysis to compare current controls against industry best practices
    • Evaluate new technologies and methodologies for contamination control
    • Consider feedback from contamination control ambassadors and staff
  4. Prioritize improvements:
    • Rank proposed enhancements based on risk reduction potential and feasibility
    • Consider cost-benefit analysis for major changes
  5. Implement changes:
    • Update standard operating procedures (SOPs) as needed
    • Provide training on new or modified control measures
    • Validate changes to ensure effectiveness
  6. Monitor and measure impact:
    • Establish key performance indicators (KPIs) for each layer of control
    • Track improvements in contamination rates and overall control effectiveness
  7. Foster a culture of continuous improvement:
    • Encourage proactive reporting of potential issues
    • Recognize and reward staff contributions to contamination control
  8. Stay updated on regulatory requirements:
    • Regularly review and incorporate changes in regulations (e.g., EU GMP Annex 1)
    • Attend industry conferences and workshops on contamination control
  9. Integrate with overall quality systems:
    • Ensure LOCA improvements align with the site’s Quality Management System
    • Update the Contamination Control Strategy (CCS) document as needed
  10. Leverage technology:
    • Implement digital solutions for environmental monitoring and data analysis
    • Consider advanced technologies like rapid microbial detection methods
  11. Conduct periodic audits:
    • Perform surprise audits to ensure adherence to protocols
    • Use findings to further refine the LOCA and control measures

Causal Factor

A causal factor is a significant contributor to an incident, event, or problem that, if eliminated or addressed, would have prevented the occurrence or reduced its severity or frequency. Here are the key points to understand about causal factors:

  1. Definition: A causal factor is a major unplanned, unintended contributor to an incident (a negative event or undesirable condition) that, if eliminated, would have either prevented the occurrence of the incident or reduced its severity or frequency.
  2. Distinction from root cause: While a causal factor contributes to an incident, it is not necessarily the primary driver. The root cause, on the other hand, is the fundamental reason for the occurrence of a problem or event. (Pay attention to the deficiencies of the model)
  3. Multiple contributors: An incident may have multiple causal factors, and eliminating one causal factor might not prevent the incident entirely but could reduce its likelihood or impact. Swiss-Cheese Model.
  4. Identification methods: Causal factors can be identified through various techniques, including: Root cause analysis (including such tools as fishbone diagrams (Ishikawa diagrams) or the Why-Why technique), Causal Learning Cycle(CLC) analysis, and Causal factor charting.
  5. Importance in problem-solving: Identifying causal factors is crucial for developing effective preventive measures and improving safety, quality, and efficiency.
  6. Characteristics: Causal factors must be mistakes, errors, or failures that directly lead to an incident or fail to mitigate its consequences. They should not contain other causal factors within them.
  7. Distinction from root causes: It’s important to note that root causes are not causal factors but rather lead to causal factors. Examples of root causes often mistaken for causal factors include inadequate procedures, improper training, or poor work culture.

Human Factors are not always Causal Factors, but can be!

Human factor and human error are related concepts but are not the same. A human error is always a causal factor, and the human factor explains why human errors can happen.

Human Error

Human error refers to an unintentional action or decision that fails to achieve the intended outcome. It encompasses mistakes, slips, lapses, and violations that can lead to accidents or incidents. There are two types:

  • Unintentional Errors include slips (attentional failures) and lapses (memory failures) caused by distractions, interruptions, fatigue, or stress.
  • Intentional Errors are violations in which an individual knowingly deviates from safe practices, procedures, or regulations. They are often categorized into routine, situational, or exceptional violations.

Human Factors

Human factors is a broader field that studies how humans interact with various system elements, including tools, machines, environments, and processes. It aims to optimize human well-being and overall system performance by understanding human capabilities, limitations, behaviors, and characteristics.

  • Physical Ergonomics focuses on human anatomical, anthropometric, physiological, and biomechanical characteristics.
  • Cognitive Ergonomics deals with mental processes such as perception, memory, reasoning, and motor response.
  • Organizational Ergonomics involves optimizing organizational structures, policies, and processes to improve overall system performance and worker well-being.

Relationship Between Human Factors and Human Error

  • Causal Relationship: Human factors delve into the underlying reasons why human errors occur. They consider the conditions and systems that contribute to errors, such as poor design, inadequate training, high workload, and environmental factors.
  • Error Prevention: By addressing human factors, organizations can design systems and processes that minimize the likelihood of human errors. This includes implementing error-proofing solutions, improving ergonomics, and enhancing training and supervision.

Key Differences

  • Focus:
    • Human Error: Focuses on the outcome of an action or decision that fails to achieve the intended result.
    • Human Factors: Focuses on the broader context and conditions that influence human performance and behavior.
  • Approach:
    • Human Error: Often addressed through training, disciplinary actions, and procedural changes.
    • Human Factors: Involves a multidisciplinary approach to design systems, environments, and processes that support optimal human performance and reduce the risk of errors.

Thinking of Swiss Cheese: Reason’s Theory of Active and Latent Failures

The Theory of Active and Latent Failures was proposed by James Reason in his book, Human Error. Reason stated accidents within most complex systems, such as health care, are caused by a breakdown or absence of safety barriers across four levels within a system. These levels can best be described as Unsafe Acts, Preconditions for Unsafe Acts, Supervisory Factors, and Organizational Influences. Reason used the term “active failures” to describe factors at the Unsafe Acts level, whereas “latent failures” was used to describe unsafe conditions higher up in the system.

This is represented as the Swiss Cheese model, and has become very popular in root cause analysis and risk management circles and widely applied beyond the safety world.

Swiss Cheese Model

In the Swiss Cheese model, the holes in the cheese depict the failure or absence of barriers within a system. Such occurrences represent failures that threaten the overall integrity of the system. If such failures never occurred within a system (i.e., if the system were perfect), then there would not be any holes in the cheese. We would have a nice Engelberg cheddar.

Not every hole that exists in a system will lead to an error. Sometimes holes may be inconsequential. Other times, holes in the cheese may be detected and corrected before something bad happens. This process of detecting and correcting errors occurs all the time.

The holes in the cheese are dynamic, not static. They open and close over time due to many factors, allowing the system to function appropriately without catastrophe. This is what human factors engineers call “resilience.” A resilient system is one that can adapt and adjust to changes or disturbances.

Holes in the cheese open and close at different rates. The rate at which holes pop up or disappear is determined by the type of failure the hole represents.

  1. Holes that occur at the Unsafe Acts level, and even some at the Preconditions level, represent active failures. Active failures usually occur during the activity of work and are directly linked to the bad outcome. Active failures change during the process of performing, opening, and closing over time as people make errors, catch their errors, and correct them.
  2. Latent failures occur higher up in the system, above the Unsafe Acts level — the Organizational, Supervisory, and Preconditions levels. These failures are referred to as “latent” because when they occur or open, they often go undetected. They can lie “dormant” or “latent” in the system for an extended period of time before they are recognized. Unlike active failures, latent failures do not close or disappear quickly.

Most events (harms) are associated with multiple active and latent failures. Unlike the typical Swiss Cheese diagram above, which shows an arrow flying through one hole at each level of the system, there can be a variety of failures at each level that interact to produce an event. In other words, there can be several failures at the Organizational, Supervisory, Preconditions, and Unsafe Acts levels that all lead to harm. The number of holes in the cheese associated with events are more frequent at the Unsafe Acts and Preconditions levels, but (usually) become fewer as one progresses upward through the Supervisory and Organizational levels.

Given the frequency and dynamic nature of activities, there are more opportunities for holes to open up at the Unsafe and Preconditions levels on a frequent basis and there are often more holes identified at these levels during root cause investigation and risk assessments.

The way the holes in the cheese interact across levels is important:

  • One-to-many mapping of causal factors is when a hole at a higher level (e.g., Preconditions) may result in several holes at a lower level (e.g. Unsafe acts)
  • Many-to-one mapping of causal factors when multiple holes at the higher level (e.g. preconditions) might interact to produce a single hole at the lower level (e.g. Unsafe Acts)

By understand the Swiss Cheese model, and Reason’s wider work in Active and Latent Failures, we can strengthen our approach to problem-solving.

Plus cheese is cool.

Swiss Cheese on a cheese board with knife