Core Primitive
Distribute cognitive work based on capacity and capability, not just availability. A team where one member is overwhelmed while others are underloaded is not using its collective capacity — it is wasting it.
The load that nobody measures
John Sweller developed cognitive load theory in the 1980s to explain why some instructional designs work and others fail. The theory identifies three types of cognitive load: intrinsic load (the inherent difficulty of the material), extraneous load (the unnecessary difficulty added by poor design), and germane load (the productive effort of building new understanding). Sweller demonstrated that learning fails when total cognitive load exceeds working memory capacity — not because the learner is incapable but because the system has asked too much of their cognitive resources at once (Sweller, 1988).
The theory was built for education, but its application to team work assignment is direct. Every task a team member undertakes imposes cognitive load — the mental effort of understanding the problem, holding the relevant context in working memory, generating and evaluating solutions, and coordinating with others. When the total cognitive load on a team member exceeds their capacity, their performance degrades in predictable ways: they make more errors, they miss important details, they take longer to complete work, and they lose the capacity for the creative and integrative thinking that their most important contributions require.
Teams routinely measure and manage many resources: time, budget, headcount, computing capacity. They almost never measure or manage cognitive load — the invisible resource that determines how well each team member thinks. The result is that cognitive load distribution happens by accident: it follows the path of least resistance, concentrating on the people who are most skilled (because they can handle it), most available (because they answer Slack fastest), or most willing (because they do not say no).
Why cognitive load concentrates
Left unmanaged, cognitive load in teams follows a power law distribution: a small number of team members carry a disproportionate share of the cognitive burden. Matthew Skelton and Manuel Pais, in Team Topologies, documented this pattern in software teams and identified the structural causes (Skelton & Pais, 2019):
Knowledge concentration. When only one person understands a critical system, all cognitive work related to that system flows to them — debugging, design decisions, code reviews, questions from other teams, incident response. The more they work on the system, the more knowledge they accumulate and the more load flows their way. Knowledge concentration is self-reinforcing and, without intervention, creates permanent imbalance.
Review bottlenecks. In many teams, code review, design review, and approval authority is concentrated in one or two senior engineers. Every piece of work must pass through these individuals, creating a cognitive bottleneck that slows the team and overloads the reviewers. The reviewers bear the cognitive load of understanding every change while the authors bear only the load of their own work.
Interrupt absorption. Some team members — often the most experienced or most approachable — absorb a disproportionate share of interruptions: questions from other teams, ad-hoc requests from product, "quick questions" from junior engineers. Each interruption is small. Their accumulation is crushing. Gloria Mark's research found that the average knowledge worker is interrupted every 11 minutes and requires 25 minutes to return to the original task — meaning that a person interrupted six times per day loses over two hours of productive cognitive capacity (Mark et al., 2008).
Invisible complexity. Some work looks simple from the outside but carries enormous cognitive load. Maintaining a legacy system that has no tests, no documentation, and no one else who understands it is less visible but more cognitively demanding than building a new feature with clear requirements and modern tooling. Teams that assign work based on apparent complexity rather than actual cognitive demand systematically overload the people who maintain the hardest systems.
Measuring cognitive load
Because cognitive load is invisible, teams must measure it deliberately. Three approaches complement each other:
Self-reported load surveys. At regular intervals (weekly or per-sprint), each team member rates their cognitive load on a simple scale: "How mentally demanding was your work this period?" and "How often did you feel overwhelmed by the complexity or volume of your cognitive tasks?" The self-report is subjective but informative — persistent high scores from the same individuals indicate structural load imbalance.
Work complexity mapping. For each task or project, estimate not just time but cognitive dimensions: How many different contexts must be held simultaneously? How much novel problem-solving is required versus routine execution? How many dependencies must be tracked? How much coordination with others is needed? A task that scores high on multiple dimensions carries more cognitive load than a task that scores high on only one, even if both take the same number of hours.
Interrupt tracking. For one week, each team member logs interruptions: the source, the duration, and the recovery time. The data reveals who absorbs the most interrupts and where the interrupts come from — information that enables structural interventions like designated interrupt handlers or improved self-service documentation that reduces the need for interruptions in the first place.
Structural interventions for load balancing
Primary and secondary ownership. Every critical system or domain has both a primary and secondary owner. The secondary participates in design decisions, reviews changes, and pairs with the primary regularly enough to maintain context. The dual ownership reduces the primary's exclusive burden and creates resilience against absence or departure. The investment in maintaining a secondary owner pays for itself the first time the primary is unavailable during an incident.
Rotating responsibilities. Review duties, on-call rotation, interrupt handling, and customer-facing tasks rotate across the team rather than defaulting to the same individuals. Rotation distributes load and builds collective capability: each team member gains exposure to different aspects of the system, and the team's transactive memory (The team's knowledge graph) becomes more robust.
Cognitive load budgets. Just as Team attention management proposed attention budgets for the team's collective focus, cognitive load budgets can be applied to individual team members. When assigning work for a sprint, the team explicitly considers each member's current load — not just their time availability but the cognitive demands of their existing commitments. A senior engineer who is already carrying a complex debugging investigation should not also be assigned to review three design documents and mentor a new hire in the same sprint.
Domain simplification. Sometimes the most effective intervention is not redistributing load but reducing it. A legacy system that imposes enormous cognitive load on its maintainer may benefit more from investment in documentation, testing, and simplification than from assigning a second person to share the suffering. Skelton and Pais argue that team cognitive load should be a first-class architectural consideration: if a system is too complex for the team that owns it to understand, the system should be simplified or the team should be restructured — not the people asked to work harder (Skelton & Pais, 2019).
The Third Brain
Your AI system can serve as a cognitive load reduction tool for the entire team. Many tasks that impose high cognitive load — understanding unfamiliar codebases, synthesizing information from multiple documents, tracing complex dependency chains — are tasks where the AI excels. When a team member faces a high-load task, their first step can be to share the context with the AI and ask for a structured summary, an analysis of the key decision points, or a walkthrough of the relevant code paths. The AI's analysis does not replace the engineer's judgment but reduces the cognitive overhead of building the context needed to exercise that judgment.
The AI can also help leaders monitor and manage load distribution. Share the team's sprint assignments with the AI along with estimates of cognitive complexity for each task, and ask: "How is cognitive load distributed across the team this sprint? Are any individuals carrying disproportionate load? What would a more balanced distribution look like?" The AI can flag imbalances that the leader might miss — particularly when load is hidden in tasks that look simple but carry high cognitive demand.
For ongoing team health monitoring, periodically share the team's self-reported load data with the AI and ask it to identify trends: "Which team members have reported consistently high load over the past three months? What types of work are correlated with high-load reports? Are there structural factors that could be changed to reduce load for the most overloaded members?"
From load to alignment
Cognitive load distribution ensures that the team's finite cognitive capacity is used efficiently. But efficiency depends on alignment — when team members hold conflicting mental models of the work, its goals, or its processes, the misalignment itself creates cognitive load as people navigate around incompatible assumptions.
The next lesson, Team schema alignment, examines team schema alignment — the practice of ensuring that the mental frameworks team members use to understand their work are compatible and mutually reinforcing.
Sources:
- Sweller, J. (1988). "Cognitive Load During Problem Solving: Effects on Learning." Cognitive Science, 12(2), 257-285.
- Skelton, M., & Pais, M. (2019). Team Topologies: Organizing Business and Technology Teams for Fast Flow. IT Revolution Press.
- Mark, G., Gudith, D., & Klocke, U. (2008). "The Cost of Interrupted Work: More Speed and Stress." Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, 107-110.
Frequently Asked Questions