Leading universities
Academic research spanning motivation, engagement, psychometrics, and organizational behavior.
Evidence-grade measurement science for decisions you can defend—from research labs to real operations.
Explore the ToolkitAcademic research spanning motivation, engagement, psychometrics, and organizational behavior.
High-stakes assessment and workforce measurement in credentialing and public-interest contexts.
People analytics, selection science, and product-adjacent analytics with a bias for operational use.
Construct mapping, item writing, content validity, pilot testing, power planning, and revision protocols.
Dimensionality, cross-loadings, fit diagnostics, bifactor/higher-order models, and residual checks.
α/ω coefficients, hierarchical ω, test–retest, inter-rater models, and G-theory for facet variance.
1–3PL/GRM/GPCM, item banks, CAT readiness, information functions, scale linking, and score banding.
Convergent/discriminant patterns, criterion/predictive evidence, known-groups, sensitivity to change.
Configural/metric/scalar/strict checks, partial invariance protocols, subgroup stability reporting.
Uniform/non-uniform DIF, item-level flags, impact analyses, and mitigation plans with transparent trade-offs.
Parsimony + interpretability, protected-attribute handling, stability checks, and decision logs.
Scope, assumptions, intended use, limitations, and monitoring cadence—measurement as a governed asset.
Structured interviewing, multiple-hurdle designs, adverse-impact review, and score cutoffs tied to utility.
Driver analysis with validated scales, team-level norms, and action heuristics with practical effect sizes.
Competency models, mastery thresholds, assessment-for-learning loops, and longitudinal sensitivity.
Brief, validated instruments for fatigue, cognitive load, and risk awareness that survive operational constraints.
Score keys, norms, banding, and interpretation tables with uncertainty guidance.
Model specs, fit indices, item parameters, reliability & invariance evidence, and data-quality checks.
Plain-language “how to use” for leaders and practitioners; guardrails and change thresholds.
Score-based thresholds raised quality-of-hire indices at steady pass rates; zero flagged DIF after revisions.
Re-factoring improved reliability (ω↑) and clarified team-level variance components for action planning.
IRT-based forms detected meaningful change with fewer items, reducing burden and boosting completion.
Deeper write-ups and templates for factor analysis, reliability families, IRT models, DIF, invariance, and score engineering.
Validating a new construct, modernizing a legacy scale, or translating research into operator-ready tools? Let’s partner.