Boeing 737 MAX: A System-Level Analysis of MCAS, Organisational Design, and Safety Assumptions

sony dsc

Summary

The Boeing 737 MAX accidents weren’t caused by a single failure, or one obvious mistake.

They came from a mix of things interacting in ways that weren’t fully visible at the time—design assumptions, certification processes, organisational pressure, and how pilots actually interact with the system in real conditions.

What makes this case important is that nothing on its own looked completely broken.

But when everything came together, it created a failure mode that only really showed up at the system level—not at the component level.

This is a good example of how modern aviation incidents often work.


 

Event Overview

Accident Sequence Overview

There were two major accidents involving the Boeing 737 MAX 8:

Lion Air Flight 610 (October 2018)

Ethiopian Airlines Flight 302 (March 2019)

In both cases, the aircraft lost control after repeated nose-down trim commands.

These commands were triggered by the MCAS system, which was reacting to incorrect angle-of-attack sensor data.

From the outside, it can look like a simple “bad sensor” problem.

But that’s only a small part of the picture.


 

System Context: The MCAS Function

MCAS (Maneuvering Characteristics Augmentation System) was introduced to adjust how the aircraft handled in certain conditions—specifically at high angles of attack.

At a basic level, it worked like this:

It monitored angle-of-attack data

If certain conditions were met, it automatically trimmed the stabiliser nose-down

The goal was to make the aircraft feel and handle more like earlier 737 models

That last point is important.

MCAS wasn’t added as a clearly visible new system—it was added quietly in the background to preserve consistency.

A key issue here was that MCAS relied on a limited sensor setup.

In its original form, it didn’t have robust redundancy for the data it depended on.


 

System-Level Analysis

Design Assumptions and Architecture Constraints

The 737 MAX was developed with a strong constraint:

Keep it as similar as possible to previous 737 models.

This influenced a lot of decisions, including:

Avoiding major changes to pilot training

Keeping the same type rating

Positioning MCAS as a background system rather than something pilots actively manage

On paper, this makes sense—it simplifies operations and reduces cost.

But it also meant that MCAS wasn’t treated as a highly visible, safety-critical feature from a pilot’s perspective.

So you end up with a system that has significant authority…

but relatively low visibility.

Human–System Interaction

Now put yourself in the cockpit during these events.

You’re in a high-workload phase—takeoff or climb.

Suddenly:

The aircraft starts trimming nose-down

You correct it

It happens again

At the same time:

You may not fully understand why it’s happening

You don’t have a clear mental model of MCAS behaviour

You’re working under time pressure

This creates a feedback loop:

Automation pushes one way

Pilot corrects

Automation repeats

Over time, that interaction becomes harder to manage, not easier.

And that’s where control starts to break down.

Organisational and Development Pressures

This part is less technical, but just as important.

The 737 MAX was developed in a highly competitive environment.

There was strong pressure to:

Compete with the Airbus A320neo

Keep development timelines tight

Avoid costly changes to training and certification

None of these pressures are unusual in industry.

But they do shape decisions—sometimes in subtle ways.

In this case, they influenced how MCAS was implemented, how visible it was, and how it was presented within the overall system.

Regulatory Structure and Certification Pathways

The certification process also played a role.

Like many modern programs, parts of the certification work were delegated:

Some safety assessments were carried out by engineers working within the manufacturer

Software changes were evaluated within the context of an existing aircraft design

The focus was often on whether components met requirements—not always how the whole system behaved together

This works well when changes are small and well understood.

But when you introduce new layers of automation, things can slip through that don’t show up at the component level.

That’s where the gap starts to appear—between “certified” and “fully understood in operation.”


 

Why the System Failed

The accidents didn’t come from one failure.

They came from a combination of factors interacting:

A system relying on limited sensor input

Low visibility of that system to the pilots

Design decisions shaped by operational and commercial constraints

Certification processes that didn’t fully capture system-level behaviour

Each of these, on its own, might be manageable.

But together, they created a situation that wasn’t fully anticipated.


 

Key Lessons for System Safety

If you step back, a few things stand out:

Redundancy in critical inputs needs to be built in clearly—not assumed

Automation shouldn’t be “invisible” to the people relying on it

Certification needs to look beyond components and consider full system behaviour

Organisational pressures don’t just affect schedules—they affect system design

Modern risks often come from interactions, not isolated failures


 

Conclusion

The Boeing 737 MAX accidents are often discussed in terms of MCAS.

But MCAS on its own isn’t really the full story.

The real issue was how multiple decisions—technical, organisational, and regulatory—fit together.

Or more accurately, how they didn’t fully align when the system was under stress.

From a systems perspective, this wasn’t just a technical fault.

It was the result of tightly connected decisions across design, certification, and operation.

And that’s what makes it such an important case study.

It shows how a system can look acceptable in parts…

but still behave in unexpected ways when everything comes together.

Related Posts