The Context Gap Your Board Doesn't Know About
You are a middle market, CEO, so you've
seen your share of tech revolutions.
If something feels off about the
semantic layer hype, you're not behind.
You're paying attention.
Welcome to the Enduring Advantage podcast.
I'm your host, Zachary Alexander.
The analysts are right about the need.
We need to improve
decision making using AI.
The vendors are wrong about the solution.
If you fear that the semantic layer
is just another money pit designed
to extend the life of obsolete
solutions, you're not wrong.
Not only do you hear the drumbeat
for semantic layer ratcheting
up, your board members do too.
You know that entropy comes
for all markets and industries.
When AI capabilities double
every seven months, the cracks
show faster than anyone expects.
The headline for Altar here
is that there's a context
gap mucking up everything.
The reality is that you need
to be able to support better
decision making today and capture
institutional knowledge for tomorrow.
So many people hear context gap and they
think they've heard all they need to know.
However, you as a CEO have to
contend with context bound knowledge.
Context bound knowledge is
institutional knowledge that stays
trapped in the environment where it
grew up because nobody asks where
else it might be highly valued.
The instinct is to protect what you've
already built, isn't resistant to
change, it's the correct strategic read.
But protection doesn't mean preservation
in amber, it means unbinding what
you know from where it grew up.
So it can be deployed to landscapes
that don't even exist yet.
That's the goal.
Not better dashboards, not faster
queries, not a semantic layer that
optimizes decisions you're already making.
The goal is to free your institutional
knowledge from the context it's bound
to so that when the landscape shifts
and it will, that knowledge travels
with you instead of dying on the vine.
Unbinding institutional knowledge
isn't a technology projects,
it's a strategic posture.
When you uny institutional knowledge,
you're not migrating data or building
better reports, you're separating
what your organization knows from
where it happens to sit right now,
so it can be deployed to context
you haven't even encountered yet.
Think about what that means.
Every market you've navigated, every
customer problem you've solved,
every operational pattern your
team has refined over the years.
All of it is currently
expressed in one context.
Your industry, your org chart,
your competitive landscape.
Unbinding says that knowledge
isn't about this landscape.
It grew up here.
It doesn't have to stay here.
The semantic layer conversation
wants you to optimize how
knowledge serves today's decisions.
Unbinding asks a bigger question, where
else might that knowledge be deployed?
When today's landscape is unrecognizable?
Here's what that means for
the drone maker posture.
Most organization facing the
need to set out for new horizons
ask which platform do we buy?
Which vendor do we bet on?
That's an incumbent strategy.
You're picking sides in someone else's
war and hoping you choose, right?
The drone makeup posture asks a
different question, what do we know
that will survive this transfer?
You're not betting on platforms.
You're building an advantage layer.
That layer that makes every AI capability
you deploy more valuable because of
what your organization already knows.
And here's what your board needs to
understand, and you're not tier one.
So much of business literature is
targeted as large multinational
enterprises, traditional incumbents.
Their wants, their needs,
their priorities, and you
can't fault the writers.
They are simply adhering
to the 80 20 rule.
Big companies hire
consultants and they pay well.
That's not a disadvantage.
It's the game.
They don't understand
the drone maker posture.
In modern warfare, the drone maker
routinely lays waste to billion dollar
weapon systems with swarms of $500
hobby drones.
In business, we take the same posture to
defeat custom integrated solutions costing
tens of millions of dollars with consumer
great AI and adaptive intelligence.
If you take that posture, you're not
competing with the tier one budgets.
You're not trying to match
their enterprise stack.
You're deploying what you know, your
institutional knowledge, your domain
expertise, your years of pattern
matching through tools that cause a
fraction of what they're spending.
And because your advantage layer is
homegrown, not rented from a vendor.
It compounds every time you deploy it.
They're locked into platforms.
You're portable.
They're optimized based on
tier one priorities, not yours.
You're
Zachary Alexander: unbinding knowledge
That you already own for the next step.
They have budget, you have velocity.
The drone maker doesn't win by
building a better fighter jet.
The drone maker wins by making
the fighter jet a bad investment.
So what does this mean for your people?
Because that's the question underneath
everything else, not which platform
to buy, not which vendor to bet on.
What happens to your workforce?
Here's the answer.
Your people aren't becoming obsolete.
They're becoming operators.
There's a distinction that matters here.
Syntax expertise versus mission specialty.
Syntax is implementation.
Writing the code, generating
the report, executing the task.
AI handles that well,
that's not a prediction.
That's the current state of affairs.
The mission is everything.
Mission alignment, course correction,
strategic judgment, Knowing that the
AI's recommendation looks right on paper,
but doesn't account for what happened
with that customer three years ago.
Knowing which pattern match, which
ones are noise, knowing when to
override, they're now operators.
Same domain knowledge, same
institutional knowledge, different.
Operational identity.
Even the most sophisticated drone
swarms have operators, not because
the technology is inadequate,
because swarms need guidance,
mission alignment, course correction.
That's a human function.
It doesn't go away.
It becomes a differentiator.
It is odd that in a conversation,
about the context gap.
We now have to look at your job
descriptions What do they ask for
10 years of Salesforce experience,
five years of Python certified in
this platform, trained on that system.
That's a syntax checklist.
You're now hiring for implementation
skills that AI handles better
than any human you can afford.
Here's what an AI native
job subscription should ask.
What level of AI operator are you
100 foundational, 200 intermediate,
or are you 300 and above?
Try asking those types of questions.
The next time someone
brings up the context gap.
Everyone knows the blockbuster story.
They had the chance to buy
Netflix for $50 million.
They laughed it off foul
for bankruptcy in 2010.
It's the most popular
cautionary tale in business.
The company that couldn't see the future.
But here's a story that
doesn't get told enough.
While Blockbuster was clinging to
late fees and physical stores, Circuit
City, another company that eventually
went bankrupt, did something.
Blockbuster never could.
They asked the second question.
Both companies were facing the same thing.
The landscape they grew
up in was disappearing.
Consumer electronics retail was dying.
Video rental was dying.
Entropy was coming for both of them.
The high level use case is identical.
What do you do when the
landscape shifts beneath you?
Blockbuster's answer, optimize
harder, more stores, more inventory.
They even tried to buy Circuit City
for a billion dollars in 2008 to
double down on physical retail.
They looked at what they knew and
asked, how do we do more of this?
Circuit City's answer at least the answer
that Sharp and Ligon, the CEO and VP
of corporate Planning was different.
They asked where else might
what we do be highly valued.
They looked at their expertise in
pricing transparency, inventory
management at scale and customer trust
in high stakes purchases, and realized
none of it was about electronics.
It was about the buying experience.
So they built CarMax, Project X
at the time, also known internally
as honest Ricks used cars.
And here's what nobody tells you.
It nearly didn't work.
CarMax bled money for seven years.
Circuit City put up 170 million to get
it started, then eventually cut it off,
forced them to find their own capital.
They had to take CarMax Public as a
tracking stock just to fund the expansion.
They tried a no fees model.
Customers didn't care.
they expanded into new car
franchises, Chrysler, Toyota, Nissan.
That didn't stick either.
They didn't turn a profit
until the year 2000.
Barely 1.1
million on 2 billion in revenue.
Seven years of failed experiments,
but the knowledge transferred.
Price and transparency
survived the transfer.
Inventory management survived.
Customer trust architecture
survived the transfer.
The container was completely different.
Used cars instead of electronics.
The context were the same.
By 2002 Circuit City spun CarMax off the
spinoff, gave shareholders stock worth 1.2
billion.
Today, CarMax operates over
250 stores in 41 states.
Circuit City, filed for bankruptcy
in 2009 and Blockbuster, the company
that never asked the second question
filed for bankruptcy in 2010.
Two bankruptcies, one company
left behind a $20 billion legacy
because it unbound its knowledge
before the landscape collapsed.
The other left behind nothing
because it optimized for
landscape that was already gone.
That's the second question in action.
Not clean, not fast, not guaranteed.
But the only question that
builds something that outlasts
the landscape it grew up in.
While the world is debating
semantic layers, institutional
knowledge is depreciating.
Right now, dying on the vine.
At the end of every day, it walks out the
door and sometimes it never comes back.
Nobody tracks this.
Your CFO can tell you the
depreciation schedule on every
piece of equipment in the building.
But the knowledge your senior
operators carries out the door.
Why the process works that way?
What happened the last time
someone tried to change it?
Which customer relationships run on
trust that only one person holds?
That's not on a balance sheet?
It's not in any systems,
it's in someone's head.
And when they leave, it leaves too.
It's not just people.
Every time you migrate a system,
you capture the structure, the
data in the field, the workflows.
You don't capture the reasoning, the
workaround someone built because a
system couldn't handle an edge case.
The institutional memory
about why that field matters.
You modernize the platform and you lose
the knowledge that made the old one work.
So you've got knowledge walking out the
door every evening and knowledge getting
stripped out every time you upgrade.
And the world is debating semantic layers.
So how do you stop knowledge
from walking out the door?
You build a context graph, not a
dashboard, not a data warehouse, a living
map of what your organization actually
knows and where the knowledge compounds.
Think about it like this right now, you
should be building systems that help AI
build the right tools for the right task.
Conserving the number of tokens each
prompt consumes is basic economics.
The next stage requires
capturing decisions.
A context graph starts there,
but doesn't stop there.
Every time an operator makes a decision,
overrides an AI recommendation.
Grants an exception, applies knowledge.
The system couldn't.
That generates a Decision trace.
It's this institutional
knowledge that currently lives
in someone's head made visible.
The context graph captures it, not the
raw data, the reasoning why this operator
changes the recommendation, what they
knew that the system didn't Over time,
those decision traces start connecting.
A pricing pattern in the Western
region shows up in the eastern region.
A customer trust approach that works
in one division transfers to another.
The context graph tracks which patterns
survive the transfer and which ones don't.
It maps your institutional knowledge,
not as a static document, but as
an adaptive layer that gets richer
every time someone makes a decision.
Knowledge compounds.
That's the advantage layer.
It starts as navigation, helping
AI tools find the right system.
It compounds into something.
no competitor can replicate.
A map of what your organization
knows, where it transfers
and what it's worth to you.
Every decision trace makes the map richer.
Every deployment across a new
context tests what survives.
Every survival compounds the value.
Knowledge that was dying on the
vine is now knowledge that compounds
A semantic layer tells you
what your data means today.
A context graph tells you what
your organization knows, where
that knowledge compounds and
where it is yet to be harvested.
The semantic layer answers
the first question.
Does what we know make us
better at what we already do?
The context graph helps us
answer the second question,
where else might what we know?
Be highly valued?
So you've got the context
graph, capturing knowledge.
Now how do you deploy it?
Not one initiative, not one big bet,
not a 12 month implementation plan
that's obsolete before it launches.
You deploy it in swarms 40,
50 initiatives simultaneously.
Each one small enough to fail cheap
and fast, but enough to learn from.
Each one is a sensor
and a strategic action.
At the same time.
Testing where your institutional knowledge
survives a transfer while generating
intelligence about the landscape.
This is where the drone maker
posture becomes operational.
You're not launching one expensive
drone and hoping it hits the target.
you're launching 40 $5,500 drones and
watching which ones find something.
So 35 fail.
That's not waste.
That's adaptive intelligence.
At industrial scale, each failure tells
you what doesn't survive the transfer.
five breakthrough, and those
five may not have been predicted.
Through analysis, they emerge through
deployment two week cycles, not
quarterly reviews, not annual planning.
Every two weeks, you're adapting,
killing what isn't working and scaling.
What is launching the next wave based
on what you learn from the last one.
26 adaptive cycles a year versus your
competitor's one strategic planning cycle.
your competitors still analyzing
which single initiative to pursue.
You've already deployed 40, failed fast
at 30 on 35, identified five breakthroughs
and started your second wave.
By the time they've scheduled their
kickoff meeting, you're scaling proven
winners and launching your third
wave based on adaptive intelligence
from wave one and wave two.
And here's what makes this so effective.
Every swarm cycle enriches
the context graph.
Every deployment tests, which
patterns transfer every failure maps,
which knowledge doesn't survive.
Every success maps where it compounds
the context graph gets richer.
With every wave, the intelligence
compounds and a competitor starting from
zero can't replicate the map because the
map only exists because you deploy it.
That's the formula.
Volume times speed, times adaptive
intelligence equals enduring advantage.
Not a fortress.
Not a mott.
You build once and hope it holds.
A living system that gets stronger
every two weeks because you keep
deploying and keep learning.
So let's bring this back
to where we started.
Your board is going to hear context
gap and reach for semantic layer.
Now you know why that solves the surface
problem and not the structural one.
The structural problem is context bound
knowledge, institutional knowledge that's
trapped in the environment where it
grew up, because no one asked where else
it might be highly valued.
The goal isn't
optimization, it's unbinding
Freeing what your organization knows from
the context it's bound to, so it can be
deployed to landscapes that don't exist.
The posture is drone maker.
Not competing with tier one's budget.
Building an advantage layer with consumer
grade AI and adaptive intelligence that
makes their enterprise stack irrelevant.
Your people aren't becoming obsolete.
They're becoming operators.
Mission specialist whose years of
domain knowledge is the most valuable
asset in the building, and your
job descriptions need to catch up.
Circuit City asked where their
knowledge might be highly valued.
Built a $20 billion company.
Block Buster optimized for
a landscape that was already
gone, left behind nothing.
The context graph captures what's
currently walking out the door.
Decision traces, patterns, the
reasoning nobody writes down.
It maps where knowledge compounds and
where it's still dying on the vine, and
you deploy in swarms 40 initiatives, 50
initiatives and cycles that takes weeks.
Every wave enriches the map.
Every failure teaches what
doesn't survive the transfer.
Every success maps where knowledge
compounds, volume times speed,
times adaptive intelligence
equals enduring advantage, the map
only exists because you deployed And
no competitor can replicate that.
without putting skin in the game.
That's enduring advantage.
Not a fortress.
You build once a living system
that gets stronger every two weeks.
The question isn't whether
the context gap is real.
It is.
The question is whether you
unbind what you know and deploy
it before dies on the vine.