A field study published today gives empirical grounding to something Signal4i has been tracking across 287 signals and 15 categories. The binding constraint on AI transformation is not the model. It is the knowledge distance between the human and the task.
Researchers embedded at IG — a leading U.K. fintech — tested whether GenAI could enable professionals from different occupational backgrounds to perform tasks at the same level as domain specialists. They recruited three groups and gave them two sequential tasks: conceptualize a web article, then execute it.
Some participants had access to bespoke GenAI tools. Others did not. The experiment was designed to isolate one question: where does AI stop being able to bridge the expertise gap?
This is the GenAI wall effect: a threshold beyond which AI can no longer meaningfully bridge the expertise gap. It bridges adjacent distances. It cannot bridge distant ones.
The researchers identified why. The data scientist approached marketing content the way he'd approach technical documentation — and removed the hooks, calls-to-action, and narrative structure he didn't recognize as valuable. He made the AI output worse because he couldn't judge it. Marketing specialists could evaluate and refine. Distant outsiders had to trust the AI for both navigation and the destination — and that's where things collapsed.
The paper's conclusion: "The bottleneck isn't the idea's quality — it's the implementer's knowledge distance from the domain."
The Harvard/Stanford experiment doesn't just confirm the gap exists — it identifies the precise mechanism. Post-mortems on failed AI pilots return six root causes: lack of skills, high costs, inadequate tools, complex projects, data complexity, confidence gap. These are not six problems. They are one problem with six faces — and every face is knowledge distance. Lack of skills is knowledge distance named from HR. High costs are the rework bill when distant evaluators degrade output. Inadequate tools is what organizations reach for when the actual problem is the human can't judge what the current tool produces. Data complexity is invisible to anyone without domain proximity. The confidence gap is precisely what knowledge distance feels like from the inside. All six collapse into a single structural failure: the human holding the AI doesn't have the domain proximity to judge the output. So it stalls in review, or it gets degraded on the way to delivery.
This is why 94% of GenAI pilots are failing. It is not a technology story. It never was.
The three-layer transformation model — Technology, Organization, Human — maps directly onto what the experiment proved.
Pull any one strand and all three stop moving. The experiment ran a controlled test and confirmed the model.
This is not a strategy. It's a procurement cycle. The pattern in most organizations: evaluate the AI tool, budget it, integrate it, move on. The org design layer gets skipped entirely. Buying a more sophisticated model doesn't close knowledge distance. It produces more sophisticated output the org still can't evaluate — and now can't slow down.
Client/server. Web. Cloud. Each transition was evaluated as an infrastructure question: assess the technology, build the business case, integrate it, move on. The platform evolved. The org stayed the same. That pattern worked because the technology was a tool. Tools don't act without being acted upon. Agents do.
Most organizations approaching AI are still running the infrastructure playbook. They're asking "what AI tools should we deploy?" when the actual question is "what does our org look like when agents are doing the execution work?" These are not the same question. The gap between them is where transformations fail.
The knowledge distance paper measures failure on a single task with a human reviewing the output. The agentic deployment removes that review by design. Agents observe, decide, and execute across time horizons no human monitors in real time. Knowledge distance × agentic deployment = degraded decisions at autonomous scale, running unmonitored in production systems. The experiment showed you the wall. Agentic deployment is what happens when you hit it at 10× the speed with no human in the loop.
The data scientists in the experiment failed because they were far from the domain. IBM i practitioners with 30 years of business logic embedded in their heads are not far from anything. They are the domain expertise the experiment shows AI needs to function at the execution layer.
That is not legacy. That is the specific asset the experiment shows determines whether AI output gets elevated or degraded. The market invented a job title for the person who closes that distance.
The market invented a job for someone who deeply understands a domain, earns the trust of senior practitioners, and makes AI work in real environments — not demos. That person already exists in the IBM i community. They just don't know what they're worth yet.
The researchers noted the GenAI wall is not fixed — it will move as AI improves. That makes the strategic window narrow, not wide. Organizations encoding practitioner knowledge into governed, auditable systems now are building a compounding asset. The ones that wait face a gap that doesn't close.
Harvard and Stanford measured the gap. Signal4i has been tracking it since signal one.