The Science of AI That Decides About People
AI psychology is the discipline that keeps AI systems valid, fair, and defensible when they make decisions about human beings. Prof. Llewellyn E. van Zyl (Ph.D) defines the field, and the AI-IARA framework operationalises it.

AI psychology is the science of designing, measuring, and assuring AI systems that make decisions about human beings. It treats people-impact AI as a measurement problem first and a technology problem second. The work asks four questions before a system is allowed near a hiring panel, a wellbeing programme, or a clinical workflow: does the system perceive the right context, interpret it correctly, act with proportionate authority, and preserve the user's autonomy and social agency. When any of those fail, the system harms people quietly and at scale. AI psychology names the failure modes, audits them, and turns the gaps into a defensible fix list.
AI psychology is not AI ethics, AI governance, or behavioural science. It is the measurement discipline behind all three when the decisions are about people.
AI-IARA. Six capacities every people-impact AI must demonstrate.
The AI-IARA framework (Awareness, Interpretation, Intention, Action, Relational Agency, Autonomy) is the canonical methodology this site teaches and applies. Each capacity surfaces a specific class of audit signal. Together they form the validity stack that determines whether an AI system that decides about people will hold up under scrutiny.
Five layers of audit evidence
Every AI psychology audit produces evidence at five layers. A system that passes one layer but fails another is not deployable. The layers are sequential, not optional.
Construct
Define the construct the system claims to measure or decide on, in language that an independent psychometrician can review. Wellbeing, engagement, fit, risk, and burnout are not interchangeable. The construct must be named, scoped, and tied to a published theoretical model. Without this layer the rest of the audit has nothing to anchor on.
Calibration
Cohort
Drift
Contestability
| Step | Title | Description |
|---|---|---|
| 01 | Construct | Define the construct the system claims to measure or decide on, in language that an independent psychometrician can review. Wellbeing, engagement, fit, risk, and burnout are not interchangeable. The construct must be named, scoped, and tied to a published theoretical model. Without this layer the rest of the audit has nothing to anchor on. |
| 02 | Calibration | Test whether the system's scores are equivalent across the populations who will be measured. A 70 in one demographic should mean what a 70 means in another, or the system is not measuring what it says. This is the layer where most commercial wellbeing and assessment tools fail and never get re-audited. |
| 03 | Cohort | Validate the system in samples that match the deployment population, not just the convenience sample it was trained on. Differential validity, measurement invariance, and floor and ceiling effects all live here. If the deployment population has not been tested, the deployment is uncontrolled. |
| 04 | Drift | Specify what proxy collapse, construct drift, and feedback-loop contamination look like for this system, with thresholds that trigger pause or rollback. Most people-impact AI degrades silently because no one is watching for it. The drift layer puts named owners and concrete signals on the watch. |
| 05 | Contestability | Specify how a human subject of the AI system can see, question, and appeal a decision. Contestability is the audit layer that converts measurement validity into procedural fairness. Without it the system is not deployable in any high-stakes setting in any jurisdiction with a meaningful AI Act. |
Common questions about AI psychology
The Authority Behind This Page
Every claim on this page is anchored in two or more independent proof types: peer-reviewed publications, third-party speaking engagements, formal standards, and named institutional roles.
Publications
Keynotes
Standards Cited
- AERA, APA, and NCME Standards for Educational and Psychological Testing
- ITC Guidelines on Psychological Testing
- EU AI Act, high-risk people-impact provisions
- ISO/IEC 42001 AI Management Systems
- NIST AI Risk Management Framework
Institutions
- Optentia Research Unit, North-West University
- Centre for Behavioural Engineering and Insight, University of Twente
- Frontiers in Psychology, Editorial Board
- Psynalytics (Chief Solutions Architect)
- Springer Nature, Editorial Affiliations
Related work and engagements
AI-Driven Assessments
Cornerstone Hub
The buyer-facing hub. What an AI assessment audit produces, with a worked example walking through all five Validity Stack layers.
The Malpractice of Good Intentions
IPPA AI Summit, 2026
Why good intentions are not enough when AI systems measure and influence human behaviour. A call for psychological product safety standards.
Psychologically Safe AI Infrastructure
6-12 week engagement
Build and assure AI systems that measure or influence human outcomes, with construct clarity, harm analysis, and lifecycle monitoring.
The AI System Psychologist's Blueprint
Article
Why talent assessment needs an architectural overhaul, and what an AI psychology audit actually produces.
The Last Generation to Define Wellbeing
IPPA AI Summit, 2026
Protecting human agency before AI optimisation removes our ability to define what flourishing means.
When AI Coaching Creates the Problems It Promises to Solve
Article
Algorithmic intimacy, parasocial dependence, and the relational-agency failure mode in AI coaching products.
Subscribe to the AI Psychology newsletter
Lessons from peer-reviewed publications, deployed audits, and live case studies on why most people-impact AI fails, and how to build systems that hold up under scrutiny.