Skip to main content

Lund University Publications

LUND UNIVERSITY LIBRARIES

Removing Biases in Communication of Severity Assessments of Intimate Partner Violence : Model Development and Evaluation

Sikström, Sverker LU orcid ; Dahl, Mats LU and Claesdotter-Knutsson, Emma LU (2023) In Journal of Medical Internet Research 25.
Abstract

Background: To support a victim of violence and establish the correct penalty for the perpetrator, it is crucial to correctly evaluate and communicate the severity of the violence. Recent data have shown these communications to be biased. However, computational language models provide opportunities for automated evaluation of the severity to mitigate the biases. Objective: We investigated whether these biases can be removed with computational algorithms trained to measure the severity of violence described. Methods: In phase 1 (P1), participants (N=71) were instructed to write some text and type 5 keywords describing an event where they experienced physical violence and 1 keyword describing an event where they experienced psychological... (More)

Background: To support a victim of violence and establish the correct penalty for the perpetrator, it is crucial to correctly evaluate and communicate the severity of the violence. Recent data have shown these communications to be biased. However, computational language models provide opportunities for automated evaluation of the severity to mitigate the biases. Objective: We investigated whether these biases can be removed with computational algorithms trained to measure the severity of violence described. Methods: In phase 1 (P1), participants (N=71) were instructed to write some text and type 5 keywords describing an event where they experienced physical violence and 1 keyword describing an event where they experienced psychological violence in an intimate partner relationship. They were also asked to rate the severity. In phase 2 (P2), another set of participants (N=40) read the texts and rated them for severity of violence on the same scale as in P1. We also quantified the text data to word embeddings. Machine learning was used to train a model to predict the severity ratings. Results: For physical violence, there was a greater accuracy bias for humans (r2=0.22) compared to the computational model (r2=0.31; t38=–2.37, P=.023). For psychological violence, the accuracy bias was greater for humans (r2=0.058) than for the computational model (r2=0.35; t38=–14.58, P<.001). Participants in P1 experienced psychological violence as more severe (mean 6.46, SD 1.69) than participants rating the same events in P2 (mean 5.84, SD 2.80; t86=–2.22, P=.029<.05), whereas no calibration bias was found for the computational model (t134=1.30, P=.195). However, no calibration bias was found for physical violence for humans between P1 (mean 6.59, SD 1.81) and P2 (mean 7.54, SD 2.62; t86=1.32, P=.19) or for the computational model (t134=0.62, P=.534). There was no difference in the severity ratings between psychological and physical violence in P1. However, the bias (ie, the ratings in P2 minus the ratings in P1) was highly negatively correlated with the severity ratings in P1 (r2=0.29) and in P2 (r2=0.37), whereas the ratings in P1 and P2 were somewhat less correlated (r2=0.11) using the psychological and physical data combined. Conclusions: The results show that the computational model mitigates accuracy bias and removes calibration biases. These results suggest that computational models can be used for debiasing the severity evaluations of violence. These findings may have application in a legal context, prioritizing resources in society and how violent events are presented in the media.

(Less)
Please use this url to cite or link to this publication:
author
; and
organization
publishing date
type
Contribution to journal
publication status
published
subject
keywords
debiasing, machine learning, natural language processing, physical, psychological, violence
in
Journal of Medical Internet Research
volume
25
article number
e43499
publisher
JMIR Publications Inc.
external identifiers
  • pmid:37115589
  • scopus:85156175757
ISSN
1438-8871
DOI
10.2196/43499
language
English
LU publication?
yes
additional info
Publisher Copyright: ©Sverker Sikstrom, Mats Dahl, Emma Claesdotter-Knutsson.
id
abf31e1f-3530-4bba-8623-147d8f9e1020
date added to LUP
2023-05-12 07:39:32
date last changed
2024-04-19 21:42:16
@article{abf31e1f-3530-4bba-8623-147d8f9e1020,
  abstract     = {{<p>Background: To support a victim of violence and establish the correct penalty for the perpetrator, it is crucial to correctly evaluate and communicate the severity of the violence. Recent data have shown these communications to be biased. However, computational language models provide opportunities for automated evaluation of the severity to mitigate the biases. Objective: We investigated whether these biases can be removed with computational algorithms trained to measure the severity of violence described. Methods: In phase 1 (P1), participants (N=71) were instructed to write some text and type 5 keywords describing an event where they experienced physical violence and 1 keyword describing an event where they experienced psychological violence in an intimate partner relationship. They were also asked to rate the severity. In phase 2 (P2), another set of participants (N=40) read the texts and rated them for severity of violence on the same scale as in P1. We also quantified the text data to word embeddings. Machine learning was used to train a model to predict the severity ratings. Results: For physical violence, there was a greater accuracy bias for humans (r<sup>2</sup>=0.22) compared to the computational model (r<sup>2</sup>=0.31; t<sub>38</sub>=–2.37, P=.023). For psychological violence, the accuracy bias was greater for humans (r<sup>2</sup>=0.058) than for the computational model (r<sup>2</sup>=0.35; t<sub>38</sub>=–14.58, P&lt;.001). Participants in P1 experienced psychological violence as more severe (mean 6.46, SD 1.69) than participants rating the same events in P2 (mean 5.84, SD 2.80; t<sub>86</sub>=–2.22, P=.029&lt;.05), whereas no calibration bias was found for the computational model (t<sub>134</sub>=1.30, P=.195). However, no calibration bias was found for physical violence for humans between P1 (mean 6.59, SD 1.81) and P2 (mean 7.54, SD 2.62; t<sub>86</sub>=1.32, P=.19) or for the computational model (t<sub>134</sub>=0.62, P=.534). There was no difference in the severity ratings between psychological and physical violence in P1. However, the bias (ie, the ratings in P2 minus the ratings in P1) was highly negatively correlated with the severity ratings in P1 (r<sup>2</sup>=0.29) and in P2 (r<sup>2</sup>=0.37), whereas the ratings in P1 and P2 were somewhat less correlated (r<sup>2</sup>=0.11) using the psychological and physical data combined. Conclusions: The results show that the computational model mitigates accuracy bias and removes calibration biases. These results suggest that computational models can be used for debiasing the severity evaluations of violence. These findings may have application in a legal context, prioritizing resources in society and how violent events are presented in the media.</p>}},
  author       = {{Sikström, Sverker and Dahl, Mats and Claesdotter-Knutsson, Emma}},
  issn         = {{1438-8871}},
  keywords     = {{debiasing; machine learning; natural language processing; physical; psychological; violence}},
  language     = {{eng}},
  publisher    = {{JMIR Publications Inc.}},
  series       = {{Journal of Medical Internet Research}},
  title        = {{Removing Biases in Communication of Severity Assessments of Intimate Partner Violence : Model Development and Evaluation}},
  url          = {{http://dx.doi.org/10.2196/43499}},
  doi          = {{10.2196/43499}},
  volume       = {{25}},
  year         = {{2023}},
}