Martini, Peter J., Patrick Scagnelli, and Marisol E. Trujillo Medina. 2026. “From Prompt to Practice: Evaluating Unedited AI-Generated Psychometric Scales.” Survey Practice 20 (March). https://doi.org/10.29115/SP-2026-0002.
Download all (2)
  • Figure 1. Bland-Altman Analysis for self-esteem scales
  • Figure 2. Bland-Altman Analysis for RWA scale and subscales

Abstract

The current study assesses whether a non-expert researcher can take an unrefined, AI-generated scale for survey dissemination and obtain results comparable to those derived from conventional, psychometrically developed scales. A sample of 140 respondents completed both established measures (e.g., the Rosenberg Self-Esteem Scale and a 10-item variant of the Right-Wing Authoritarianism Scale) and unaltered AI-derived counterparts. Analyses included factor structure comparisons, reliability assessments, paired t-tests of standardized scores, and Bland-Altman plots of agreement. Results showed that AI-generated items often reproduced similar factor solutions and reliability levels as traditional scales. Agreement was strongest at the whole-scale level, though subscales revealed nuanced differences. These findings suggest that even without expert refinement, AI-generated measures can approximate established scales in many respects. For practitioners, this raises important considerations about efficiency and rigor in survey design: although AI can provide usable starting points, careful evaluation remains essential before deployment in applied research.

Accepted: January 21, 2026 EDT