Introduction
Stories about aviation often double as stories about identity. Two very different works — the anime Macross Plus (1994) and the Hollywood sequel Top Gun: Maverick (2022) — circle around the same anxiety: what happens when machines threaten to eclipse the pilot.
One frames the pilot as irreplaceable, a sunset ace whose instincts matter more than any drone. The other imagines a future where the machine doesn’t just replace the pilot, but exposes him — amplifying flaws as much as skill.
Taken together, they dramatise a tension that reaches far beyond cinema. They ask what happens to practice, apprenticeship, and judgment when machines begin to echo and surpass human performance. If the films are about spectacle, the professional lesson is about texture: the rituals, refusals, and archives that keep human judgment alive even as machines learn to imitate it.
Apprenticeship vs. Automation
- Maverick insists that instinct and experience are irreplaceable — the pilot as rock star, the jet as power fantasy.
- Macross Plus pre‑countered that claim a few decades earlier, suggesting that the machine doesn’t just replace us; it reveals us — if we let it. The nervous‑system link amplifies flaws and limitations as much as skill.
Together, they dramatise a professional dilemma: is mastery about being faster than the system, or about knowing what the system cannot know?
Echoes and Identity
- Machines echo human judgment, but echoes are not wisdom.
- Apprenticeship is about texture — the slow accumulation of craft, the rituals that resist reduction.
- When we mistake echoes for mastery, we risk hollowing out the very practices that make judgment resilient.
Objections and Answers
Automation has undeniably improved safety. But safety is not the same as stewardship. When you allow the AI to automate, 99.99% of the time it’s great. The system executes flawlessly, faster and more consistently than any human. But what happens in the 0.01%?
- When the model encounters an edge case it was never trained for.
- When the data reflects bias or blind spots.
- When the system’s logic collides with human values or context.
- When the prompt is ambiguous, and the system executes in ways that may be unsafe.
That is where apprenticeship matters. Not because humans are faster, but because they are stewards of exceptions. Apprenticeship teaches how to recognise anomalies, how to improvise when the system falters, and how to carry responsibility when the machine cannot.
Apprenticeship is inefficient. And that is precisely why it matters. The inefficiency is texture: the pauses, corrections, and rituals that build judgment. Strip those away, and you risk hollowing out the very practices that make human oversight credible.
Why This Matters Professionally
For leaders, engineers, and other stewards working at the edge of technology, the lesson is clear:
- Technology is not the villain.
- Identity is the stake.
- The danger is not obsolescence alone, but the erosion of apprenticeship — the rituals that teach us how to steward systems responsibly.
Closing
These stories remind us that machines will one day outgrow us. The question is whether we can keep alive the practices that make human judgment worth echoing.
In the professional world, the anxiety isn’t about speed — of thought or execution. It’s relevance. Machines may one day decide what matters, but for now, the burden of judgment is still ours.


Leave a Reply