AI Psychology

The Science of AI That Decides About People

AI psychology is the discipline that keeps AI systems valid, fair, and defensible when they make decisions about human beings. Prof. Llewellyn E. van Zyl (Ph.D) defines the field, and the AI-IARA framework operationalises it.

Prof. Llewellyn E. van Zyl (Ph.D). The leading voice in AI psychology.
AI Psychology

AI psychology is the science of designing, measuring, and assuring AI systems that make decisions about human beings. It treats people-impact AI as a measurement problem first and a technology problem second. The work asks four questions before a system is allowed near a hiring panel, a wellbeing programme, or a clinical workflow: does the system perceive the right context, interpret it correctly, act with proportionate authority, and preserve the user's autonomy and social agency. When any of those fail, the system harms people quietly and at scale. AI psychology names the failure modes, audits them, and turns the gaps into a defensible fix list.

AI psychology is not AI ethics, AI governance, or behavioural science. It is the measurement discipline behind all three when the decisions are about people.

The Method

AI-IARA. Six capacities every people-impact AI must demonstrate.

The AI-IARA framework (Awareness, Interpretation, Intention, Action, Relational Agency, Autonomy) is the canonical methodology this site teaches and applies. Each capacity surfaces a specific class of audit signal. Together they form the validity stack that determines whether an AI system that decides about people will hold up under scrutiny.

AI-IARAFramework
Awareness
Interpretation
Intention
Action
Relational Agency
Autonomy
The Validity Stack

Five layers of audit evidence

Every AI psychology audit produces evidence at five layers. A system that passes one layer but fails another is not deployable. The layers are sequential, not optional.

Step 01

Construct

Define the construct the system claims to measure or decide on, in language that an independent psychometrician can review. Wellbeing, engagement, fit, risk, and burnout are not interchangeable. The construct must be named, scoped, and tied to a published theoretical model. Without this layer the rest of the audit has nothing to anchor on.

Step 02

Calibration

Step 03

Cohort

Step 04

Drift

Step 05

Contestability

StepTitleDescription
01ConstructDefine the construct the system claims to measure or decide on, in language that an independent psychometrician can review. Wellbeing, engagement, fit, risk, and burnout are not interchangeable. The construct must be named, scoped, and tied to a published theoretical model. Without this layer the rest of the audit has nothing to anchor on.
02CalibrationTest whether the system's scores are equivalent across the populations who will be measured. A 70 in one demographic should mean what a 70 means in another, or the system is not measuring what it says. This is the layer where most commercial wellbeing and assessment tools fail and never get re-audited.
03CohortValidate the system in samples that match the deployment population, not just the convenience sample it was trained on. Differential validity, measurement invariance, and floor and ceiling effects all live here. If the deployment population has not been tested, the deployment is uncontrolled.
04DriftSpecify what proxy collapse, construct drift, and feedback-loop contamination look like for this system, with thresholds that trigger pause or rollback. Most people-impact AI degrades silently because no one is watching for it. The drift layer puts named owners and concrete signals on the watch.
05ContestabilitySpecify how a human subject of the AI system can see, question, and appeal a decision. Contestability is the audit layer that converts measurement validity into procedural fairness. Without it the system is not deployable in any high-stakes setting in any jurisdiction with a meaningful AI Act.
People Also Ask

Common questions about AI psychology

Proof Stack

The Authority Behind This Page

Every claim on this page is anchored in two or more independent proof types: peer-reviewed publications, third-party speaking engagements, formal standards, and named institutional roles.

Standards Cited

  • AERA, APA, and NCME Standards for Educational and Psychological Testing
  • ITC Guidelines on Psychological Testing
  • EU AI Act, high-risk people-impact provisions
  • ISO/IEC 42001 AI Management Systems
  • NIST AI Risk Management Framework

Institutions

  • Optentia Research Unit, North-West University
  • Centre for Behavioural Engineering and Insight, University of Twente
  • Frontiers in Psychology, Editorial Board
  • Psynalytics (Chief Solutions Architect)
  • Springer Nature, Editorial Affiliations

Subscribe to the AI Psychology newsletter

Lessons from peer-reviewed publications, deployed audits, and live case studies on why most people-impact AI fails, and how to build systems that hold up under scrutiny.