The human role is left implicit by default

WEF Future of Jobs 2025 research identifies a persistent gap in how organisations deploy AI: the human role in AI-assisted workflows is rarely explicitly designed. The AI component is specified with precision: the model, the prompt, the output format, the integration point. The human component is left as "review" or "approve", without definition of what that review entails, what criteria govern the approval, how long it should take or what the person does when the AI output is wrong.

This is a design gap, not a trust gap. Organisations deploying AI are generally aware that human oversight matters. They specify it in principle and leave it undefined in practice. The result is that the human checkpoint becomes either a formality that adds time without adding value, or a bottleneck that eventually disappears under operational pressure.

Both outcomes are governance failures. A checkpoint that exists on paper but does not hold in practice is more dangerous than no checkpoint: it creates the appearance of oversight without the substance of it.

What explicit human role design requires

Effective human-AI collaboration requires the human role to be specified at the same level of precision as the AI component. Three questions must be answered before deployment.

What specifically does the human review? The entire output, or a defined subset? The factual claims, the tone, the decision implications or the compliance elements? A review brief that specifies exactly what the human is looking for produces consistent oversight. A review brief that says "check the output" produces whatever the reviewer chooses to check on that day.

What criteria determine whether the output passes? Without documented criteria, pass or fail becomes a judgment call that varies by reviewer, varies by day and varies by how much time the reviewer has. Documented criteria make the checkpoint consistent and auditable.

What action does the human take on each outcome? If the output passes, the workflow proceeds. If it fails, what happens? The reviewer edits it, rejects it and requests regeneration, escalates it or overrides it with their own output? Each action has different implications for the workflow design. None of them should be left to convention.

Organisations that sustain oversight design it explicitly

The organisations that sustain effective human oversight in AI workflows document the checkpoint at the same level of detail as the AI step. What does the person review? What criteria govern approval? How long should the review take? What action follows each outcome? Who handles escalation? These questions get explicit answers before deployment. A checkpoint that cannot be documented to this level has not been designed. It has been assumed. Assumptions hold in low-pressure conditions and collapse in high-pressure ones.

This matters in practice. When the human role is explicit, the people performing the checkpoint know what they are expected to do and checkpoints hold even when time is scarce. When the criteria are specific enough to define what matters, the review catches the errors that matter rather than whatever the reviewer happens to notice on the day. Generic oversight catches generic errors. Specific oversight catches the ones that change outcomes.

The organisations that make this investment find that their human checkpoints survive the shift from low-pressure design to high-pressure operation. Those that do not make this investment find that their checkpoints fade faster than the pressure increases.