Human Capabilities Assessment
A synthesis of the 2015 National Research Council report Measuring Human Capabilities: An Agenda for Basic Research on the Assessment of Individual and Group Performance Potential for Military Accession. The report was commissioned by the U.S. Army Research Institute to identify what additional constructs — beyond the existing ASVAB cognitive battery and TAPAS personality assessment — could sharpen how the Army predicts which recruits will perform well. Its framing is useful well beyond the military context: it is the most comprehensive recent audit of what basic research on individual-differences measurement can say about predicting real-world performance, and which measurement problems are ready for operational use versus still in basic-research territory. For the adjacent expertise-development literature see deliberate-practice-and-performance; for the personality side see personality-and-situation; for workplace measurement applications see leadership-frameworks.
What the Report Is
The National Research Council established the Committee on Measuring Human Capabilities: Performance Potential of Individuals and Collectives under its Board on Behavioral, Cognitive, and Sensory Sciences. The committee's charge was to develop a basic-research agenda aimed at supplementing — not replacing — the cognitive and personality measures already used in Army enlisted-soldier selection. The ASVAB (Armed Services Vocational Aptitude Battery) provides a cognitive baseline; the TAPAS (Tailored Adaptive Personality Assessment System) provides a personality baseline. The question the committee asked is: what other constructs are either under-measured or entirely missing from the current system that, if added, could improve accession decisions at the margin?
The framing matters because even small increases in predictive accuracy are valuable when applied at scale. Recruiting and training a soldier who subsequently underperforms or leaves is expensive enough that a predictor adding even a few percentage points of accuracy has large expected returns. The committee's filter was strict: constructs had to have (1) a theoretical foundation, (2) evidence of relevance to military outcomes, and (3) feasibility of mass administration to large candidate pools without specialized equipment. That filter excluded some promising areas (vocational interest measurement) as "operational, not basic research" and others as technologically premature.
The committee used Sackett and Lievens's (2008) taxonomy for how selection systems can be improved: (a) identifying new predictor constructs, (b) identifying new outcomes to predict, (c) improving measurement of existing constructs, or (d) identifying moderators of predictor-criterion relationships. The final research agenda falls into four clusters mapped onto that taxonomy.
New Predictor Constructs
Fluid intelligence, working memory capacity, executive attention, and inhibitory control. The committee treats these as related but distinguishable capacities. Fluid intelligence is the ability to reason about novel problems without relying on stored knowledge. Working memory capacity is the ability to hold and manipulate information in mind simultaneously. Executive attention is the ability to resist attentional capture by internal or external distractors. Inhibitory control is the ability to suppress prepotent but incorrect responses. All four load onto a shared factor but also carry unique variance, and the committee argues that measuring them separately — rather than absorbing them into the single "g" factor the ASVAB is optimized around — could improve prediction, particularly for tasks involving sustained attention under load. The link to self-control-and-willpower is direct: these are the cognitive substrates of self-regulation.
Cognitive biases. Defined in the report as reflexive thinking patterns that produce judgment errors or departures from normative rules. The committee treats susceptibility to specific biases (anchoring, availability, confirmation, framing) as an individual-differences dimension that has not been systematically incorporated into selection, despite evidence that people differ reliably in how much their judgments are distorted. See cognitive-biases-and-psychology for the broader research program this draws on.
Spatial abilities. The capacity to understand objects' spatial relationships and to manipulate multidimensional representations in mind. Spatial ability is a classic psychometric dimension but is under-weighted in most modern selection batteries relative to its predictive validity for technical and operational work. The committee flags this as a high-payoff area for better measurement.
New Outcomes to Predict
Teamwork behavior. The committee's most developed recommendation in this cluster. Existing selection focuses on individual taskwork — can this person perform the job? — and underweights teamwork, which concerns how individuals coordinate and cooperate as interdependent units. The report distinguishes "taskwork" (individual job-specific KSAOs) from "teamwork" (the coordinated processes that emerge when multiple people work together). It argues that measuring team-related individual differences pre-accession — roles members naturally take, communication tendencies, backup behavior, cross-training openness — could improve team-level outcomes more efficiently than trying to train these skills post-hoc. This connects to the esports-and-gaming research on communication as the #1 team performance problem, and to the Breuer et al. (2016) meta-analysis on trust in virtual teams in leadership-frameworks.
Hybrid Topics: Predictors and Outcomes
Hot cognition: defensive reactivity, emotional regulation, and performance under stress. "Hot cognition" refers to performance in situations that elicit strong emotion, as opposed to the "cold" cognition of standard testing conditions. The committee argues that individual differences in emotion regulation under stress are theoretically important and under-measured. Situational Tests of Emotional Management (MacCann & Roberts) and related instruments correlate meaningfully with well-being outcomes (r = .54 with eudaimonic well-being, r = .44 with net affect during work). Problem-focused coping predicts achievement outcomes after controlling for personality and cognitive ability. The report also notes a developmental caveat: emotion regulation improves through adolescence with frontal-lobe maturation, and the average Army recruit is twenty — at the upper end of that developmental window. See ed-batista-on-emotion-regulation for the coaching-side version, stress-mindset-science for the performance-under-stress cluster, and cofounder-conflict-physiology for the flooding-and-regulation applied case.
Adaptability and inventiveness. Adaptability is the ability to adjust to changing physical, interpersonal, cultural, and task environments. Inventiveness is the ability to generate novel, high-quality, task-appropriate ideas. Both map onto Big Five Openness-to-Experience but cut across multiple traits. The committee reports that predictor batteries combining specific cognitive abilities (divergent thinking, spatial ability) with personality variables (achievement motivation, creative personality) produce correlations for predicting innovative contributions as high as r ≈ .53. Adaptability is particularly interesting because modern operational environments increasingly demand it as a capacity separate from fixed competence in a specific task.
Methods and Methodology
Psychometrics and technology. The committee highlights measurement-technology advances that create new opportunities: computer-adaptive testing, real-time response tracking, latent-variable models that extract construct estimates from patterns of response behavior. Technology also enables faster iteration on test content and more sophisticated fraud resistance.
Situations and situational judgment tests (SJTs). SJTs present candidates with realistic scenarios and assess their judgment about what to do. They have become a standard part of many selection systems because they tap the interaction between person and situation in a way pure-trait measures cannot. The committee argues for a more systematic research program on what SJTs actually measure, how they can be constructed with better psychometric properties, and how they integrate with other predictors. The situational-versus-dispositional debate in personality-and-situation is directly relevant here: SJTs are one empirical response to the observation that behavior depends on both the person and the situation.
Assessment of individual differences through neuroscience measures. The committee is cautious but open. Neural measures (EEG, fMRI, and derivative indices) show systematic individual differences that correlate with cognitive and personality constructs. The question is whether those measures add incremental validity beyond what standard psychometric instruments already provide, and whether they can be administered feasibly at scale. The report treats this as a basic-research agenda — an area where measurement science might unlock something, but not an area ready for operational use.
Through-Line
Several themes recur across the committee's recommendations.
The person-situation interaction is load-bearing. The report repeatedly emphasizes that individual differences matter most when they interact with specific situational demands. SJTs, team-context measures, and hot-cognition assessments are all attempts to get closer to the person-in-situation unit that actually predicts real performance, rather than the trait-in-isolation that dominates classical testing. This is the same intuition that drives the situational-power findings in personality-and-situation.
Construct separation improves prediction. The classical measurement approach treats "intelligence" or "personality" as unitary and extracts composite scores. The committee argues that in many cases, separating finer-grained constructs (working memory vs. executive attention vs. inhibitory control; observing vs. non-reactivity; adaptability vs. inventiveness) and measuring them individually produces better prediction than collapsing them. The parallel to the mindfulness-science Five Facet work by Baer is instructive: measuring "mindfulness" as one thing hid the Observing-facet reversal visible only at the facet level.
Real-world performance is multi-determined. No single predictor does the heavy lifting. The most accurate prediction comes from combining cognitive ability, personality, situation-specific judgment, and stress-response measures. The report is a rebuttal to any framework that treats performance as reducible to a single trait (IQ, grit, conscientiousness), and an endorsement of multi-method batteries that respect the multiple causal paths to competent action.
Relevance for Jason's Work
The report's framing is useful for coaching in three ways.
First, it provides a vocabulary for talking about what is actually being measured when a client or organization runs a capability assessment. Many of Jason's clients — founders, executives, high-stakes athletes — operate in environments where both cognitive and affective performance under stress matter, and the report clarifies why standard "intelligence" or "personality" measures alone are insufficient predictors.
Second, it underwrites the case for individualized coaching. Because performance is multi-determined and construct-separation improves prediction, coaching that identifies specific dimensions where a client is strong or weak — rather than treating "performance" as a unitary trait to be pushed up — tends to produce more targeted interventions. The authentic-pride-patterns exercise is one version of this principle: find where the person is actually strong, build from there.
Third, the adaptability/inventiveness literature supports Jason's framing in expertise-as-river: in a changing environment, the capacity to adapt matters as much as the level of any fixed skill. The committee's research recommendations on measuring adaptability are effectively the academic counterpart to Jason's argument that identity should be tied to capacity for change rather than to current expertise.
Related Topics
- deliberate-practice-and-performance — The expertise side; how capacity becomes skill through training
- personality-and-situation — Individual differences plus situational power as the right unit of analysis
- self-control-and-willpower — Working memory, executive attention, and inhibitory control as self-regulation substrates
- cognitive-biases-and-psychology — The bias-susceptibility dimension as an individual-differences construct
- leadership-frameworks — Team effectiveness and Breuer et al. on trust in virtual teams
- stress-mindset-science — Performance under stress as a measurement target
- ed-batista-on-emotion-regulation — The coaching-side version of the hot-cognition cluster
- expertise-as-river — Adaptability as the AI-era core skill
- grit-ambition-and-achievement — The dispositional side of sustained performance
- brain-plasticity-and-cognition — The neuroscience-measures direction the committee flags as basic research