The News Worth Taking Seriously
This week, DeepMind CEO Demis Hassabis stated publicly that current AI systems fail along three specific dimensions: continuous learning, long-term planning, and consistency. These are not minor engineering complaints. They are architectural limitations that define the boundary between narrow AI performance and what Hassabis considers genuine intelligence. The statement is notable not because it is surprising, but because of who is saying it and in what context. The person closest to the most advanced AI development program in the world is drawing a clear line between what these systems can do and what organizations are currently assuming they can do.
The gap between those two things is where I think the real organizational problem lives, and it is not the gap Hassabis named.
What Hassabis Got Right and What He Left Out
The three limitations Hassabis identified are real and well-documented. Current large language models do not update their weights during deployment, so they cannot learn from new information in the way a human professional does over the course of a career. They struggle to maintain coherent long-horizon plans across complex task sequences. And their output consistency is unreliable in ways that are difficult to predict in advance. These are legitimate constraints that matter for any organization making decisions about where to integrate AI systems into consequential workflows.
But there is a fourth problem that Hassabis did not name, because it is not an engineering problem. It is an organizational competence problem. Organizations are deploying systems with these limitations without having developed any structural understanding of what those limitations mean for coordination and decision-making. The failure is not in the AI. The failure is in the interpretive infrastructure surrounding it.
The Awareness-Capability Gap at the Organizational Level
Research on algorithmic literacy consistently finds that awareness of how algorithms function does not translate into improved performance or better decision-making (Kellogg, Valentine, and Christin, 2020; Gagarin, Naab, and Grub, 2024). Workers who know that an algorithm exists and even know something about how it operates do not automatically develop the competence to respond to it effectively. This gap between awareness and capability is one of the central puzzles in platform coordination research.
Hassabis's statement is a high-profile instance of awareness being generated at the leadership level. The DeepMind CEO has now publicly documented that current systems cannot plan long-term or maintain consistency. Leadership at major organizations will read this. Some will cite it in board presentations. Very few will translate it into revised protocols for how AI outputs are reviewed, escalated, or rejected. Knowing that AI systems lack consistency is categorically different from knowing how to build organizational processes that account for that inconsistency. This is the distinction Hatano and Inagaki (1986) drew between routine expertise and adaptive expertise, applied now at the level of institutional design rather than individual skill.
Consistency Failures Are Coordination Failures
The consistency problem Hassabis identified deserves particular attention from an organizational theory perspective. When a system produces inconsistent outputs, the downstream coordination problem is not simply one of error correction. It is a problem of schema mismatch. Human workers coordinating around AI outputs develop working models of what those outputs mean and how reliable they are. When outputs are inconsistent in ways that are unpredictable, those working models degrade. The folk theories that workers construct to make sense of algorithmic behavior, which Kellogg et al. (2020) identify as a common adaptive response, become progressively less accurate.
Sundar (2020) argues that machine agency introduces a fundamentally different communicative dynamic because human receivers attribute authority and neutrality to machine-generated outputs in ways they do not to human-generated ones. If that attribution is operating within organizations deploying current AI systems, and the evidence strongly suggests it is, then the consistency failures Hassabis described are not just quality problems. They are trust calibration problems that accumulate silently until a consequential decision fails.
What This Means for How Organizations Should Respond
The response I would predict from most organizations is procedural. They will update documentation, add disclaimers to AI outputs, and perhaps introduce a review checkpoint. This is the wrong response. Procedures address known failure modes in stable environments. The limitations Hassabis described are structural and variable, which means they require adaptive expertise rather than procedural compliance (Hatano and Inagaki, 1986). Organizations need workers who understand why current AI systems fail in the ways they do, not just that they sometimes fail.
Schema induction, rather than procedural training, is the appropriate intervention here. Workers who understand the architectural reasons for AI inconsistency, the absence of continuous learning, the absence of stable long-term planning, can generalize that understanding to novel failure contexts. Workers who have only been trained on documented failure cases will be unprepared when the system fails in a new way. Hassabis has done organizations a service by naming the gaps clearly. The harder work is building the organizational competence to actually use that information.
References
Gagarin, A., Naab, T. K., and Grub, J. (2024). Algorithmic media use and algorithm literacy. New Media and Society.
Hatano, G., and Inagaki, K. (1986). Two courses of expertise. In H. Stevenson, H. Azuma, and K. Hakuta (Eds.), Child development and education in Japan. Freeman.
Kellogg, K. C., Valentine, M. A., and Christin, A. (2020). Algorithms at work: The new contested terrain of control. Academy of Management Annals, 14(1), 366-410.
Sundar, S. S. (2020). Rise of machine agency: A framework for studying the psychology of human-AI interaction. Journal of Computer-Mediated Communication, 25(1), 74-88.
Roger Hunt