Skip to main content

Lund University Publications

LUND UNIVERSITY LIBRARIES

Beyond rating scales : With targeted evaluation, large language models are poised for psychological assessment

Kjell, Oscar N.E. LU ; Kjell, Katarina LU and Schwartz, H. Andrew LU (2024) In Psychiatry Research 333.
Abstract

In this narrative review, we survey recent empirical evaluations of AI-based language assessments and present a case for the technology of large language models to be poised for changing standardized psychological assessment. Artificial intelligence has been undergoing a purported “paradigm shift” initiated by new machine learning models, large language models (e.g., BERT, LAMMA, and that behind ChatGPT). These models have led to unprecedented accuracy over most computerized language processing tasks, from web searches to automatic machine translation and question answering, while their dialogue-based forms, like ChatGPT have captured the interest of over a million users. The success of the large language model is mostly attributed to... (More)

In this narrative review, we survey recent empirical evaluations of AI-based language assessments and present a case for the technology of large language models to be poised for changing standardized psychological assessment. Artificial intelligence has been undergoing a purported “paradigm shift” initiated by new machine learning models, large language models (e.g., BERT, LAMMA, and that behind ChatGPT). These models have led to unprecedented accuracy over most computerized language processing tasks, from web searches to automatic machine translation and question answering, while their dialogue-based forms, like ChatGPT have captured the interest of over a million users. The success of the large language model is mostly attributed to its capability to numerically represent words in their context, long a weakness of previous attempts to automate psychological assessment from language. While potential applications for automated therapy are beginning to be studied on the heels of chatGPT's success, here we present evidence that suggests, with thorough validation of targeted deployment scenarios, that AI's newest technology can move mental health assessment away from rating scales and to instead use how people naturally communicate, in language.

(Less)
Please use this url to cite or link to this publication:
author
; and
organization
publishing date
type
Contribution to journal
publication status
published
subject
keywords
Artificial intelligence, Assessment, Large language models, Psychology, Transformers
in
Psychiatry Research
volume
333
article number
115667
publisher
Elsevier
external identifiers
  • pmid:38290286
  • scopus:85184248031
ISSN
0165-1781
DOI
10.1016/j.psychres.2023.115667
language
English
LU publication?
yes
id
d437678d-1bcb-484c-99ea-45fae37411ed
date added to LUP
2024-03-07 15:49:48
date last changed
2024-04-19 14:13:48
@article{d437678d-1bcb-484c-99ea-45fae37411ed,
  abstract     = {{<p>In this narrative review, we survey recent empirical evaluations of AI-based language assessments and present a case for the technology of large language models to be poised for changing standardized psychological assessment. Artificial intelligence has been undergoing a purported “paradigm shift” initiated by new machine learning models, large language models (e.g., BERT, LAMMA, and that behind ChatGPT). These models have led to unprecedented accuracy over most computerized language processing tasks, from web searches to automatic machine translation and question answering, while their dialogue-based forms, like ChatGPT have captured the interest of over a million users. The success of the large language model is mostly attributed to its capability to numerically represent words in their context, long a weakness of previous attempts to automate psychological assessment from language. While potential applications for automated therapy are beginning to be studied on the heels of chatGPT's success, here we present evidence that suggests, with thorough validation of targeted deployment scenarios, that AI's newest technology can move mental health assessment away from rating scales and to instead use how people naturally communicate, in language.</p>}},
  author       = {{Kjell, Oscar N.E. and Kjell, Katarina and Schwartz, H. Andrew}},
  issn         = {{0165-1781}},
  keywords     = {{Artificial intelligence; Assessment; Large language models; Psychology; Transformers}},
  language     = {{eng}},
  publisher    = {{Elsevier}},
  series       = {{Psychiatry Research}},
  title        = {{Beyond rating scales : With targeted evaluation, large language models are poised for psychological assessment}},
  url          = {{http://dx.doi.org/10.1016/j.psychres.2023.115667}},
  doi          = {{10.1016/j.psychres.2023.115667}},
  volume       = {{333}},
  year         = {{2024}},
}