Selection Methods

Multi-Trait Multi-Method Design: Convergent and Discriminant Validity

By Editorial Team — reviewed for accuracy Published
Last reviewed:

The multi-trait multi-method matrix, originally articulated by Donald Campbell and Donald Fiske in their 1959 Psychological Bulletin paper, is the foundational framework for assessing whether a selection or psychological measurement instrument actually measures what it claims to measure. The MTMM matrix arrays multiple traits crossed with multiple methods, allowing the researcher to inspect correlations among same-trait-different-method pairings (convergent validity) versus different-trait-same-method pairings (discriminant validity). Where convergent correlations exceed discriminant correlations, the instrument is measuring traits more than methods; where they don’t, the instrument is partly or wholly method-bound.

This article walks through the MTMM logic, the convergent-discriminant decision rules, why the framework is foundational for selection-instrument design, the practical applications in modern selection-battery construction, and how AIEH applies MTMM logic to the multi-source composite underlying the Skills Passport.

Data Notice: Theoretical framework citations reflect peer-reviewed psychometric literature at time of writing. Specific MTMM-derived weighting decisions in the AIEH composite are documented in the scoring methodology and may evolve as calibration data accrues during launch.

The MTMM matrix structure

The MTMM matrix is a correlation table with traits as rows and methods as columns. Suppose three traits (cognitive ability, conscientiousness, agreeableness) are each measured by three methods (self-report questionnaire, peer rating, structured interview). The resulting matrix is 9x9, with each cell containing the correlation between one trait-method combination and another. Campbell and Fiske identified four diagnostic regions within the matrix:

  • Reliability diagonal. Same trait, same method, separate occasions (or separate forms). Coefficients here represent measurement reliability and are expected to be high.
  • Validity diagonal (convergent triangle). Same trait, different methods. Coefficients here represent convergent validity — the extent to which different methods of measuring the same trait yield consistent estimates. Strong convergent validity shows large coefficients in this region.
  • Heterotrait-monomethod triangles. Different traits, same method. Coefficients here represent method-shared variance — the extent to which scores from the same method correlate regardless of which trait is supposedly being measured.
  • Heterotrait-heteromethod triangles. Different traits, different methods. Coefficients here are the most stringent test of discriminant validity — if these are large, the instrument may be tapping some general construct rather than discriminating among the named traits.

Campbell and Fiske articulated four decision rules: convergent coefficients should be statistically significant and substantial, convergent coefficients should exceed heterotrait-heteromethod correlations in their row and column, convergent coefficients should exceed heterotrait-monomethod correlations involving the same trait, and the same pattern of trait correlations should hold across method blocks.

Why MTMM matters for selection design

The MTMM framework matters because selection instruments routinely fail under MTMM analysis in ways that look like trait measurement but are actually method measurement. The classic failure mode: an assessment center claims to measure seven managerial-competency dimensions, but factor-analytic decomposition shows that variance is dominated by exercise-level effects rather than dimension-level effects. Sackett and Dreher’s 1982 empirical work and subsequent replications documented this pattern across the assessment-center literature.

The implication is that a candidate’s “leadership” score in an in-basket exercise correlates more with the candidate’s “decision-making” score in the same in-basket exercise than with the candidate’s “leadership” score in a separate role-play. The exercise variance dominates the dimension variance — the methodology is partly measuring exercise-specific performance rather than the named dimensions.

This isn’t unique to assessment centers. The pattern recurs across selection-instrument families:

  • Self-report personality questionnaires show convergent validity across instruments measuring the same Big Five trait but also substantial method-shared variance from the self-report format.
  • Structured interviews show some convergent validity with other methods measuring the same dimensions but also notable interview-method variance.
  • Cognitive-ability tests show strong convergent validity across cognitive-test instruments because the construct is robust and the methods overlap.

For deeper coverage of the assessment-center construct-validity dispute that MTMM analysis catalyzed, see the assessment center validity article.

Convergent and discriminant validity in practice

Convergent validity is the easier criterion to establish. If two methods purport to measure the same trait — say, conscientiousness measured by self-report and conscientiousness measured by peer rating — strong convergent validity simply requires that the correlations between the two are substantial after correction for unreliability. Coefficients in the ~0.40-0.60 range corrected for unreliability are typical for well-designed convergent pairings.

Discriminant validity is the harder criterion. Two methods purporting to measure different traits but sharing a method (both self-report, say) tend to show inflated correlations from method-shared variance. The decision rule is that heterotrait-monomethod correlations should be lower than convergent (heterotrait, same trait via different method) correlations. When they’re not, the instrument is method-bound rather than trait-bound.

The contemporary literature has refined the original Campbell-Fiske decision rules through confirmatory-factor-analytic models — the correlated-uniqueness model, the correlated-traits-correlated-methods model, and multitrait-multimethod-multitime extensions — but the core logic remains: convergent must exceed discriminant, or the construct interpretation collapses.

For broader treatment of how construct interpretation ties to selection-validity research, see the validity generalization meta-analyses article.

Applications in selection-battery design

Practical implications of MTMM logic for selection-battery design:

  • Don’t double-count method-bound signal. A battery containing two self-report Big Five instruments and a self-report integrity test is partly measuring the self-report method rather than three distinct constructs. Method diversification matters: combining self-report with peer-rating, with behavioral observation, and with performance on objective tasks produces stronger composite validity than combining multiple instruments of the same method.
  • Convergent evidence builds confidence. When a candidate scores high on cognitive ability via a general-mental-ability test and high on the same underlying construct via job-knowledge performance, the convergent evidence supports the trait interpretation. When the two diverge substantially, one of the methods is producing method-bound variance.
  • Method blocks matter for predictor-criterion alignment. Sackett and Lievens (2008) note that selection methods predicting performance differ in the trait-vs-method composition of their variance. Methods with strong trait variance and weaker method variance tend to generalize better across organizational contexts.

For practical guidance on assembling a defensible selection battery, see hiring loop design and pre-employment screening evidence.

Common pitfalls

  • Treating method-bound covariance as construct evidence. When two self-report instruments correlate, the easy interpretation is that they’re measuring the same construct. The harder interpretation — and the more often correct one — is that they share a self-report method that contributes shared variance regardless of construct.
  • Ignoring exercise-level variance in multi-exercise designs. When a selection procedure uses multiple exercises (assessment center, multi-interview loop), exercise-level variance can dominate dimension-level variance. MTMM analysis would catch this; absent the analysis, the procedure is reported as measuring the named dimensions when it’s partly measuring the exercises.
  • Overinterpreting modest correlations. A convergent correlation of ~0.30 corrected isn’t strong support for shared-trait interpretation. The decision rules require convergent coefficients to be substantially higher than discriminant alternatives, not just statistically significant.

AIEH integration

The Skills Passport composite is built on MTMM-aware multi-source aggregation. The four pillars (cognitive, domain, AI fluency, communication) are deliberately constructed to combine evidence from different methods within each pillar, not to layer multiple same-method instruments into a single pillar. Examples:

  • The cognitive pillar combines general-mental-ability test evidence with job-knowledge test evidence with work-sample reasoning evidence — three different methods tapping related but distinct cognitive constructs.
  • The domain pillar combines work-sample evidence, job-knowledge test evidence, and structured technical-interview evidence — three different methods producing convergent or discriminant patterns the recruiter can inspect.
  • The AI fluency pillar combines AI Output Evaluation test evidence with AI Collaboration Literacy scenario evidence — two different methods within an emerging construct domain.
  • The communication pillar combines written scenario evidence with structured-interview communication ratings — different methods tapping the construct.

The composite logic is designed so that strong convergent evidence within a pillar produces a defensible high score, while purely method-bound high scores (high on one method only) appropriately weight lower in the composite. The scoring methodology documents how the weighting handles same-method-redundancy and cross-method convergence.

The candidate-owned framing means recruiters clicking through the Skills Passport at aieh.com/passport/{handle} see the per-pillar evidence breakdown and can inspect the multi-method composition rather than seeing only the composite. This supports informed hiring-decision-making at the hire workspace level. For the broader treatment of how multi-method evidence supports defensible selection, see skills-based hiring evidence.

Takeaway

Campbell and Fiske’s 1959 multi-trait multi-method matrix is the foundational framework for assessing whether selection instruments measure traits or methods. The four decision rules — substantial convergent correlations, convergent exceeding heterotrait-heteromethod, convergent exceeding heterotrait-monomethod, consistent trait pattern across methods — provide the diagnostic structure for separating construct measurement from method-bound covariance. The framework has shaped fifty-plus years of personnel-selection research, including the assessment-center construct-validity literature. AIEH applies MTMM logic to the Skills Passport composite by deliberately combining different methods within each pillar and by weighting cross-method convergence higher than same-method redundancy.

Sources

  • Schmidt, F. L., & Hunter, J. E. (1998). The validity and utility of selection methods in personnel psychology: Practical and theoretical implications of 85 years of research findings. Psychological Bulletin, 124(2), 262-274.
  • Sackett, P. R., & Lievens, F. (2008). Personnel selection. Annual Review of Psychology, 59, 419-450.
  • Campbell, D. T., & Fiske, D. W. (1959). Convergent and discriminant validation by the multitrait-multimethod matrix. Psychological Bulletin, 56(2), 81-105.
  • Sackett, P. R., & Dreher, G. F. (1982). Constructs and assessment center dimensions: Some troubling empirical findings. Journal of Applied Psychology, 67(4), 401-410.
  • Eid, M., Lischetzke, T., Nussbeck, F. W., & Trierweiler, L. I. (2003). Separating trait effects from trait-specific method effects in multitrait-multimethod models: A multiple-indicator CT-C(M-1) model. Psychological Methods, 8(1), 38-60.
  • Lance, C. E., Lambert, T. A., Gewin, A. G., Lievens, F., & Conway, J. M. (2004). Revised estimates of dimension and exercise variance components in assessment center postexercise dimension ratings. Journal of Applied Psychology, 89(2), 377-385.

About This Article

Researched and written by the AIEH editorial team using official sources. This article is for informational purposes only and does not constitute professional advice.

Last reviewed: · Editorial policy · Report an error