﻿<?xml version="1.0" encoding="utf-8"?>
<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Publishing DTD v1.1 20151215//EN" "JATS-journalpublishing1.dtd">
<article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" article-type="systematic-review">
<front>
<journal-meta>
<journal-id journal-id-type="nlm-ta">Explor Med</journal-id>
<journal-id journal-id-type="publisher-id">EM</journal-id>
<journal-title-group>
<journal-title>Exploration of Medicine</journal-title>
</journal-title-group>
<issn pub-type="epub">2692-3106</issn>
<publisher>
<publisher-name>Open Exploration Publishing</publisher-name>
</publisher>
</journal-meta>
<article-meta>
<article-id pub-id-type="doi">10.37349/emed.2025.1001370</article-id>
<article-id pub-id-type="manuscript">1001370</article-id>
<article-categories>
<subj-group>
<subject>Systematic Review</subject>
</subj-group>
</article-categories>
<title-group>
<article-title>Artificial intelligence for pain assessment via facial expression recognition (2015–2025): a systematic review</article-title>
</title-group>
<contrib-group>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-5236-3132</contrib-id>
<name>
<surname>Cascella</surname>
<given-names>Marco</given-names>
</name>
<role content-type="https://credit.niso.org/contributor-roles/writing-original-draft/">Writing—original draft</role>
<role content-type="https://credit.niso.org/contributor-roles/formal-analysis/">Formal analysis</role>
<role content-type="https://credit.niso.org/contributor-roles/conceptualization/">Conceptualization</role>
<role content-type="https://credit.niso.org/contributor-roles/methodology/">Methodology</role>
<role content-type="https://credit.niso.org/contributor-roles/supervision/">Supervision</role>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing—review &amp; editing</role>
<xref ref-type="aff" rid="I1">
<sup>1</sup>
</xref>
<xref ref-type="corresp" rid="cor1">
<sup>*</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0009-0002-8889-0249</contrib-id>
<name>
<surname>Esposito</surname>
<given-names>Dalila</given-names>
</name>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing—review &amp; editing</role>
<xref ref-type="aff" rid="I1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0000-0002-8172-2325</contrib-id>
<name>
<surname>Muzio</surname>
<given-names>Maria Rosaria</given-names>
</name>
<role content-type="https://credit.niso.org/contributor-roles/writing-review-editing/">Writing—review &amp; editing</role>
<xref ref-type="aff" rid="I2">
<sup>2</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0009-0004-9377-1791</contrib-id>
<name>
<surname>Cascella</surname>
<given-names>Vincenzo</given-names>
</name>
<xref ref-type="aff" rid="I3">
<sup>3</sup>
</xref>
</contrib>
<contrib contrib-type="author">
<contrib-id contrib-id-type="orcid">https://orcid.org/0009-0006-5657-6959</contrib-id>
<name>
<surname>Cerrone</surname>
<given-names>Valentina</given-names>
</name>
<xref ref-type="aff" rid="I1">
<sup>1</sup>
</xref>
</contrib>
<contrib contrib-type="editor">
<name>
<surname>Chang</surname>
<given-names>Min Cheol</given-names>
</name>
<role>Academic Editor</role>
<aff>Yeungnam University, Republic of Korea</aff>
</contrib>
</contrib-group>
<aff id="I1">
<sup>1</sup>Department of Medicine, University of Salerno, 84081 Salerno, Italy</aff>
<aff id="I2">
<sup>2</sup>Division of Infantile Neuropsychiatry, UOMI-Maternal and Infant Health, Asl Napoli 3 Sud, Torre del Greco, 80059 Naples, Italy</aff>
<aff id="I3">
<sup>3</sup>School of Medicine, University of Pavia, 27100 Pavia, Italy</aff>
<author-notes>
<corresp id="cor1">
<bold>
<sup>*</sup>Correspondence:</bold> Marco Cascella, Department of Medicine, University of Salerno, 84081 Salerno, Italy. <email>mcascella@unisa.it</email></corresp>
</author-notes>
<pub-date pub-type="collection">
<year>2025</year>
</pub-date>
<pub-date pub-type="epub">
<day>12</day>
<month>11</month>
<year>2025</year>
</pub-date>
<volume>6</volume>
<elocation-id>1001370</elocation-id>
<history>
<date date-type="received">
<day>27</day>
<month>08</month>
<year>2025</year>
</date>
<date date-type="accepted">
<day>10</day>
<month>10</month>
<year>2025</year>
</date>
</history>
<permissions>
<copyright-statement>© The Author(s) 2025.</copyright-statement>
<license xlink:href="https://creativecommons.org/licenses/by/4.0/">
<license-p>This is an Open Access article licensed under a Creative Commons Attribution 4.0 International License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, sharing, adaptation, distribution and reproduction in any medium or format, for any purpose, even commercially, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.</license-p>
</license>
</permissions>
<abstract>
<sec>
<title>Background:</title>
<p id="absp-1">Although accurate pain assessment is crucial in clinical care, pain evaluation is traditionally based on self-report or observer-based scales. Artificial intelligence (AI) applied to facial expression recognition is promising for objective, automated, and real-time pain assessment.</p>
</sec>
<sec>
<title>Methods:</title>
<p id="absp-2">The study followed PRISMA guidelines. We searched PubMed/MEDLINE, Scopus, Web of Science, Cochrane Library, and the IEEE Xplore databases for the literature published between 2015 and 2025 on the applications of AI for pain assessment via facial expression analysis. Eligible studies included original articles in English applying different AI techniques. Exclusion criteria were neonatal/pediatric populations, non-facial approaches, reviews, case reports, letters, and editorials. Methodological quality was assessed using the RoB 2 tool (for RCTs) and adapted appraisal criteria for AI development studies. This systematic review was registered in PROSPERO (<ext-link xlink:href="https://doi.org/10.17605/OSF.IO/N9PZA" ext-link-type="uri">https://doi.org/10.17605/OSF.IO/N9PZA</ext-link>).</p>
</sec>
<sec>
<title>Results:</title>
<p id="absp-3">A total of 25 studies met the inclusion criteria. Sample sizes ranged from small experimental datasets (<italic>n</italic> &lt; 30) to larger clinical datasets (<italic>n</italic> &gt; 500). AI strategies included machine learning models, convolutional neural networks (CNNs), recurrent neural networks such as long short-term memory (LSTM), transformers, and multimodal fusion models. The accuracy in pain detection varied between ~70% and &gt; 90%, with higher performance observed in deep learning and multimodal frameworks. The risk of bias was overall moderate, with frequent concerns related to small datasets and lack of external validation. No meta-analysis was performed due to heterogeneity in datasets, methodologies, and outcome measures.</p>
</sec>
<sec>
<title>Discussion:</title>
<p id="absp-4">AI-based facial expression recognition shows promising accuracy for automated pain assessment, particularly in controlled settings and binary classification tasks. However, evidence remains limited by small sample sizes, methodological heterogeneity, and scarce external validation. Large-scale multicenter studies are required to confirm clinical applicability and to strengthen the certainty of evidence for use in diverse patient populations.</p>
</sec>
</abstract>
<kwd-group>
<kwd>artificial intelligence</kwd>
<kwd>facial expression recognition</kwd>
<kwd>pain assessment</kwd>
<kwd>action units</kwd>
</kwd-group>
</article-meta>
</front>
<body>
<sec id="s1">
<title>Introduction</title>
<p id="p-1">Pain assessment is a crucial aspect of healthcare. Since it traditionally relies on subjective assessments and observable symptoms, potential inaccuracies and delays in effective intervention can occur [<xref ref-type="bibr" rid="B1">1</xref>]. In recent years, automatic pain assessment (APA) has emerged as a research area of significant interest [<xref ref-type="bibr" rid="B1">1</xref>, <xref ref-type="bibr" rid="B2">2</xref>]. This complex set of approaches is aimed at objectively evaluating pain. Therefore, APA can enable individualized, patient-centered care, helping healthcare providers and caregivers to develop timely and appropriate interventions for improving pain management and quality of life [<xref ref-type="bibr" rid="B3">3</xref>].</p>
<p id="p-2">Interestingly, different strategies have been implemented for APA. They encompass biosignal-based investigations and behavior-based approaches. Recognition of facial expressions is the most investigated behavior for APA [<xref ref-type="bibr" rid="B4">4</xref>]. Given the significant advancements in the field of automatic facial image analysis through computer vision models, research on pain detection from facial expressions is encouraged [<xref ref-type="bibr" rid="B5">5</xref>]. Specifically, many approaches focus on the analysis of action units (AUs). They are the smallest visually observable facial muscle movements codified by the Facial Action Coding System (FACS). It is a standardized tool where each AU corresponds to the activation of one or more facial muscles [<xref ref-type="bibr" rid="B3">3</xref>–<xref ref-type="bibr" rid="B5">5</xref>]. Furthermore, in the context of multimodal strategies for APA research, AUs are often combined with other behaviors or biosignals, such as electrocardiography (ECG)-derived parameters, electrodermal activity (EDA), photoplethysmography (PPG), respiratory rate, and vocal features [<xref ref-type="bibr" rid="B6">6</xref>, <xref ref-type="bibr" rid="B7">7</xref>].</p>
<p id="p-3">Nevertheless, despite their interesting promises, the reliable application of these methods in pain evaluation is still an open challenge. These unresolved research questions concern the selection and standardization of datasets, the choice of artificial intelligence (AI) models and architectures, the design of processing pipelines, and the need for both internal and external validation in real-world conditions. Moreover, their applicability across different clinical contexts, such as acute versus chronic pain, and in diverse care settings, from emergency departments to palliative care, should be carefully investigated [<xref ref-type="bibr" rid="B8">8</xref>, <xref ref-type="bibr" rid="B9">9</xref>].</p>
<sec id="t1-1">
<title>Objective</title>
<p id="p-4">The objective was to summarize and critically evaluate the evidence of published studies between 2015 and 2025 on the application of AI for pain assessment through the analysis of facial expressions. The review addressed the following question: “What AI-based methods have been applied to detect or assess pain from facial expressions, and what is their accuracy and applicability in clinical or experimental settings?”.</p>
</sec>
</sec>
<sec id="s2">
<title>Materials and methods</title>
<sec id="t2-1">
<title>Literature search strategy</title>
<p id="p-5">The search strategy was developed in line with the PRISMA (Preferred Reporting Items for Systematic reviews and Meta-Analyses) guidelines [<xref ref-type="bibr" rid="B10">10</xref>].</p>
<p id="p-6">The systematic search was conducted in the main biomedical and computer databases, including PubMed/MEDLINE, Scopus, Cochrane Library, and Web of Science, integrated by IEEE Xplore for the literature in engineering and computer science. To ensure completeness, references of the selected articles and relevant citations were also manually checked. Only studies published in the period 1 January 2015–31 July 2025 were considered. The latest electronic search was completed in July 2025. Search records were managed in the Rayyan software [<xref ref-type="bibr" rid="B11">11</xref>].</p>
</sec>
<sec id="t2-2">
<title>Eligibility criteria</title>
<p id="p-7">We included articles addressing techniques for APA through facial expression analysis. Studies conducted on adult patients of different clinical conditions, without restrictions related to the care setting (e.g., hospital, intensive care, outpatient, or experimental setting), were considered eligible. A mandatory comparator was not required; however, when present, traditional pain assessment tools (e.g., visual analogue scale, numeric rating scale, clinical observation) were evaluated.</p>
<p id="p-8">Studies not related to the assessment of pain, opinion-only articles, editorials, conference abstracts without complete data, works that used non-AI-based approaches, or that employed AI in contexts unrelated to facial recognition were excluded.</p>
<p id="p-9">Only peer-reviewed studies published in English were included. Unpublished articles, not peer-reviewed manuscripts, technical reports, or grey literature were not considered.</p>
</sec>
<sec id="t2-3">
<title>Search strategy</title>
<p id="p-10">The search strategy was developed according to the PCC (Population, Concept, Context) framework and tailored for each database.</p>
<p id="p-11">The search terms combined controlled vocabulary (e.g., MeSH) and free-text words related to artificial intelligence (“artificial intelligence”, “machine learning”, “deep learning”, “neural networks”, “computer vision”), pain assessment (“pain assessment”, “pain detection”, “pain evaluation”, “pain monitoring”, “pain recognition”, “pain quantification”, “pain scoring”), and facial expressions (“facial expression recognition”, “facial expressions”, “face recognition”, “emotion recognition”, “facial coding”, “nonverbal communication”, “visual perception”).</p>
<p id="p-12">Filters were applied to restrict results to:</p>
<p id="p-13">
<list list-type="bullet">
<list-item>
<p>Years of publication: 2015–2025;</p>
</list-item>
<list-item>
<p>Language: English;</p>
</list-item>
<list-item>
<p>Study design: primary research articles (excluding reviews, case reports, letters, and editorials);</p>
</list-item>
<list-item>
<p>Population: adult humans (excluding neonatal and adolescent populations).</p>
</list-item>
</list>
</p>
</sec>
<sec id="t2-4">
<title>Selection process and data collection</title>
<p id="p-14">After removing duplicates, titles, and abstracts of all retrieved searches were independently screened by two reviewers (V Cerrone, MC) to assess if they met the inclusion criteria. Full texts of the included studies were then retrieved and reviewed for final selection. Disagreements were resolved through consultation with a third reviewer (DE). For each included study, two reviewers (V Cascella, MRM) independently extracted data using a standardized form developed for this review. Extracted information included study characteristics (authors, year, setting, population), methodological details, AI approaches applied, datasets used, and main outcomes. Discrepancies in data extraction were resolved through consensus.</p>
</sec>
<sec id="t2-5">
<title>Data items</title>
<p id="p-15">For each included study, we extracted information on study characteristics (authors, year of publication, country, and setting), study population (sample size, age range, clinical condition), methodological design, AI approach (machine learning or deep learning algorithm used), type of dataset employed (public or clinical), and main outcomes. Outcomes of interest included accuracy, sensitivity, specificity, F1-score, area under the receiver operating characteristic curve (AUROC), and other reported performance metrics. Additional variables collected were funding sources, validation strategy (e.g., cross-validation, external validation), and whether the study involved real-world clinical implementation.</p>
</sec>
<sec id="t2-6">
<title>Study risk of bias assessment</title>
<p id="p-16">Two reviewers (MRM, V Cascella) independently rated the methodological quality of all included studies using the Revised Cochrane risk-of-bias tool for randomized trials (RoB 2) [<xref ref-type="bibr" rid="B12">12</xref>]. As recommended by the guidelines, the risk of bias was described and assessed for all primary outcomes evaluated in each study. Any disagreements about the methodological quality were resolved through consensus.</p>
</sec>
<sec id="t2-7">
<title>Effect measures</title>
<p id="p-17">The primary effect measures considered were accuracy, sensitivity, specificity, F1-score, AUROC, and other key AI metrics, as reported in the original studies.</p>
</sec>
<sec id="t2-8">
<title>Synthesis methods</title>
<p id="p-18">Given the methodological heterogeneity across studies (e.g., different AI algorithms, datasets, and outcomes), a narrative synthesis was conducted. Studies were grouped by type of dataset (clinical vs. benchmark/public datasets) and by AI method applied (machine learning vs. deep learning). Results were tabulated to allow for structured comparison of study characteristics and main findings. No quantitative meta-analysis was conducted due to variability in outcome measures and study designs.</p>
</sec>
<sec id="t2-9">
<title>Reporting bias assessment</title>
<p id="p-19">Asymmetry or reporting bias was not formally assessed through statistical tests, since no meta-analysis was performed. However, we noted whether studies selectively reported only favorable performance metrics or failed to provide confidence intervals.</p>
</sec>
<sec id="t2-10">
<title>Certainty assessment</title>
<p id="p-20">Given the methodological nature of the review and the lack of homogeneity across outcomes, the certainty of the body of evidence was not graded using GRADE. Instead, emphasis was placed on highlighting recurring strengths and limitations of the included studies, as well as identifying areas of consistency versus heterogeneity.</p>
</sec>
</sec>
<sec id="s3">
<title>Results</title>
<sec id="t3-1">
<title>Study selection</title>
<p id="p-21">The literature search generated 79 references (Scopus = 16, PubMed = 31, Web of Science = 4, Cochrane Library = 2, IEEE = 26), of which 25 studies met our inclusion criteria. Numbers and reasons for exclusion at each stage are reported in <xref ref-type="fig" rid="fig1">Figure 1</xref>. The full search strategy for each database is reported in <xref ref-type="sec" rid="s-suppl">Table S1</xref>.</p>
<fig id="fig1" position="float">
<label>Figure 1</label>
<caption>
<p id="fig1-p-1">
<bold>PRISMA flow diagram.</bold> Adapted from [<xref ref-type="bibr" rid="B10">10</xref>]. © 2019 The Authors. Licensed under a Creative Commons Attribution (CC BY 4.0).</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="em-06-1001370-g001.tif" />
</fig>
<p id="p-22">During the full-text review, four studies were excluded despite initially appearing to meet the inclusion criteria. The study by Chan et al. [<xref ref-type="bibr" rid="B13">13</xref>] was excluded because it was only an abstract (conference proceeding) and did not provide sufficient methodological details for inclusion. Other manuscripts were excluded [<xref ref-type="bibr" rid="B4">4</xref>, <xref ref-type="bibr" rid="B14">14</xref>–<xref ref-type="bibr" rid="B16">16</xref>]. Specifically, they failed to focus on facial expression-based pain assessment [<xref ref-type="bibr" rid="B4">4</xref>, <xref ref-type="bibr" rid="B14">14</xref>, <xref ref-type="bibr" rid="B16">16</xref>], or we found a lack of AI strategies [<xref ref-type="bibr" rid="B15">15</xref>].</p>
</sec>
<sec id="t3-2">
<title>Study characteristics</title>
<p id="p-23">In this systematic review, a total of 25 studies published between 2015 and 2025 were included [<xref ref-type="bibr" rid="B17">17</xref>–<xref ref-type="bibr" rid="B41">41</xref>]. Most of this scientific output concerned experimental or feasibility studies. The number of participants varied widely, ranging from small experimental samples of fewer than 30 healthy volunteers exposed to controlled pain stimuli (e.g., cold pressor or heat pain tests) to large clinical datasets exceeding 500 postoperative patients.</p>
<p id="p-24">The population covered different contexts, such as healthy adults in laboratory conditions, surgical and perioperative patients, intensive care unit (ICU) patients, older adults with dementia, and oncology patients. While the common focus of these studies was the use of AI applied to facial expressions for the automatic detection or quantification of pain, some studies also integrated multimodal data (e.g., speech, audio, physiological signals), though facial analysis remained the primary input modality.</p>
<p id="p-25">About the AI methods, researchers employed different AI modalities. They ranged from classical machine learning algorithms (e.g., random forests and support vector machines) to deep learning approaches. Artificial neural network models included convolutional neural networks (CNNs), often combined with recurrent neural networks (RNNs) or long short-term memory (LSTM) units [<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B27">27</xref>, <xref ref-type="bibr" rid="B29">29</xref>, <xref ref-type="bibr" rid="B35">35</xref>, <xref ref-type="bibr" rid="B37">37</xref>]. More recent contributions introduced vision transformers (ViTs). These architectures leverage self-attention mechanisms to process images as sequences of patches, thereby improving robustness in different challenging conditions such as ICU scenarios [<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B31">31</xref>]. In parallel, attention-based deep learning architectures, such as enhanced residual attention-based subject-specific network (ErAS-Net), were specifically designed to enhance feature selection and improve classification performance across datasets [<xref ref-type="bibr" rid="B33">33</xref>]. A distinct subgroup of studies focused on multimodal approaches, integrating facial data with additional modalities such as speech, audio, or physiological signals [<xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B22">22</xref>, <xref ref-type="bibr" rid="B28">28</xref>, <xref ref-type="bibr" rid="B38">38</xref>]. These systems consistently outperformed unimodal facial analysis, particularly in challenging populations such as ICU patients with partial facial occlusion [<xref ref-type="bibr" rid="B40">40</xref>] or older adults with dementia [<xref ref-type="bibr" rid="B36">36</xref>], although at the expense of increased computational and methodological complexity. Other innovative approaches included the application of transfer entropy to landmark time-series analysis [<xref ref-type="bibr" rid="B25">25</xref>], surface electromyographic (sEMG) signals to capture subtle muscle activation [<xref ref-type="bibr" rid="B32">32</xref>], and binary classifiers trained on facial AUs derived from the FACS [<xref ref-type="bibr" rid="B21">21</xref>, <xref ref-type="bibr" rid="B26">26</xref>].</p>
<p id="p-26">The reported outcomes were generally expressed as accuracy, area under the curve (AUC), sensitivity, specificity, and F1-score, with many studies achieving high performance on benchmark datasets, such as Delaware [<xref ref-type="bibr" rid="B42">42</xref>], University of Northern British Columbia Pain Expression dataset (UNBC)-McMaster dataset [<xref ref-type="bibr" rid="B43">43</xref>], or on locally collected clinical datasets.</p>
<p id="p-27">The summary of the individual study characteristics, including setting and number of patients included, population, approach, and inclusion criteria, AI method, and reported outcomes, is presented in <xref ref-type="table" rid="t1">Table 1</xref>.</p>
<table-wrap id="t1">
<label>Table 1</label>
<caption>
<p id="t1-p-1">
<bold>Summary of the included studies.</bold>
</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th>Source/Year</th>
<th>Dataset/<italic>n</italic></th>
<th>Population</th>
<th>Approach/Inclusion criteria</th>
<th>AI method</th>
<th>Outcome (as reported)</th>
</tr>
</thead>
<tbody>
<tr>
<td>Bargshady et al. [<xref ref-type="bibr" rid="B17">17</xref>], 2024</td>
<td>Lab datasets (AI4PAIN: 51; BioVid: 87)</td>
<td>Adults</td>
<td>Acute Pain Datasets (video-based)</td>
<td>Video vision transformers (ViViTs)</td>
<td>Accuracy 66.9% (AI4PAIN), 79.9% (BioVid), outperforming ResNet baselines</td>
</tr>
<tr>
<td>Bargshady et al. [<xref ref-type="bibr" rid="B18">18</xref>], 2020</td>
<td>UNBC-McMaster, MIntPAIN</td>
<td>Adults</td>
<td>Benchmark datasets</td>
<td>Ensemble DL model (CNN + RNN hybrid, EDLM)</td>
<td>Accuracy &gt; 89%, ROC 0.93; robust vs. single-stream CNN</td>
</tr>
<tr>
<td>Bellal et al. [<xref ref-type="bibr" rid="B19">19</xref>], 2024</td>
<td>ICU, 30 patients</td>
<td>Critically ill, non-communicative adults</td>
<td>NEVVA<sup>®</sup> pilot device calibration</td>
<td>AI-based computer vision integrated in devices</td>
<td>Feasible, device calibrated against expert assessment</td>
</tr>
<tr>
<td>Benavent-Lledo et al. [<xref ref-type="bibr" rid="B20">20</xref>], 2023</td>
<td>UNBC, BioVid</td>
<td>Adults</td>
<td>Public pain expression datasets</td>
<td>Transformer-based computer vision</td>
<td>Accuracy &gt; 96% (UNBC), &gt; 94% (BioVid); high precision, recall</td>
</tr>
<tr>
<td>Cascella et al. [<xref ref-type="bibr" rid="B21">21</xref>], 2024</td>
<td>Oncology + public datasets (Delaware, UNBC)</td>
<td>Cancer patients, adults</td>
<td>Binary classifier using AUs</td>
<td>Neural network (17 AUs, OpenFace)</td>
<td>Accuracy ~94%; AUROC 0.98</td>
</tr>
<tr>
<td>Cascella et al. [<xref ref-type="bibr" rid="B22">22</xref>], 2024</td>
<td>Oncology</td>
<td>Adult cancer patients</td>
<td>Video + audio (facial + speech)</td>
<td>Multimodal AI (speech emotion + facial expression)</td>
<td>Feasibility shown; early accuracy promising</td>
</tr>
<tr>
<td>Cascella et al. [<xref ref-type="bibr" rid="B23">23</xref>], 2023</td>
<td>Clinical feasibility (real-time)</td>
<td>Adults</td>
<td>Real-time pain detection from facial videos</td>
<td>YOLOv8 object detection</td>
<td>Feasible, metrics reported with good accuracy (JPR)</td>
</tr>
<tr>
<td>Casti et al. [<xref ref-type="bibr" rid="B24">24</xref>], 2019</td>
<td>Clinical/Lab setting</td>
<td>Adults</td>
<td>Automatic pain detection calibration</td>
<td>DL-based system (CNN)</td>
<td>Benchmarked; addressed inter-/intra-observer variability</td>
</tr>
<tr>
<td>Casti et al. [<xref ref-type="bibr" rid="B25">25</xref>], 2021</td>
<td>Public dataset (video pain sequences)</td>
<td>Adults</td>
<td>Landmark time-series analysis</td>
<td>Transfer entropy (TE) + ML classifiers</td>
<td>TE-based approach improved accuracy, robust to noise</td>
</tr>
<tr>
<td>Chen et al. [<xref ref-type="bibr" rid="B26">26</xref>], 2022</td>
<td>UNBC + lung cancer dataset</td>
<td>Adults, including patients with lung cancer</td>
<td>Pain-related AUs</td>
<td>Weakly supervised MIL/MCIL</td>
<td>Accuracy 87%, AUC 0.94 (UNBC); validated also on clinical lung cancer data</td>
</tr>
<tr>
<td>Dutta and M [<xref ref-type="bibr" rid="B27">27</xref>], 2018</td>
<td>UNBC + live video</td>
<td>Adults</td>
<td>Real-time video-based pain recognition</td>
<td>Hybrid DL model</td>
<td>Validated in real-time; high accuracy reported</td>
</tr>
<tr>
<td>Ghosh et al. [<xref ref-type="bibr" rid="B28">28</xref>], 2025</td>
<td>UNBC, BioVid + VIVAE (audio)</td>
<td>Adults</td>
<td>Multimodal (facial + audio)</td>
<td>Ensemble DL with CNN + fusion</td>
<td>Accuracy up to 99.5% (3-class), 87.4% (5-class); audio peak 98%</td>
</tr>
<tr>
<td>Guo et al. [<xref ref-type="bibr" rid="B29">29</xref>], 2021</td>
<td>Cold pressor experiment; 29 subjects</td>
<td>Adults</td>
<td>Cold pain induction</td>
<td>CNN (Inception V3, VGG-LSTM, ConvLSTM)</td>
<td>F1 score 79.5% (personalized ConvLSTM)</td>
</tr>
<tr>
<td>Heintz et al. [<xref ref-type="bibr" rid="B30">30</xref>], 2025</td>
<td>Perioperative, multicenter (503 pts)</td>
<td>Adults perioperative</td>
<td>Computer vision nociception detection</td>
<td>CNN-based</td>
<td>Strong AUROC, external validation, and feasibility proven</td>
</tr>
<tr>
<td>Mao et al. [<xref ref-type="bibr" rid="B31">31</xref>], 2025</td>
<td>UNBC</td>
<td>Adults</td>
<td>Pain intensity estimation</td>
<td>Conv-Transformer (multi-task joint optimization)</td>
<td>Outperformed SOTA; improved regression + classification</td>
</tr>
<tr>
<td>Mieronkoski et al. [<xref ref-type="bibr" rid="B32">32</xref>], 2020</td>
<td>31 healthy volunteers, experimental</td>
<td>Adults</td>
<td>Pain induction + sEMG</td>
<td>ML (supervised on muscle activation)</td>
<td>Modest c-index 0.64; eyebrow/lip muscles most predictive</td>
</tr>
<tr>
<td>Morsali and Ghaffari [<xref ref-type="bibr" rid="B33">33</xref>], 2025</td>
<td>UNBC, BioVid</td>
<td>Adults</td>
<td>Public Pain Datasets</td>
<td>ErAS-Net (attention-based DL)</td>
<td>Accuracy 98.8% (binary, UNBC); 94.2% (4-class); cross-dataset BioVid 78%</td>
</tr>
<tr>
<td>Park et al. [<xref ref-type="bibr" rid="B34">34</xref>], 2024</td>
<td>155 pts post-gastrectomy</td>
<td>Postoperative adults</td>
<td>Clinical recordings</td>
<td>ML models (facial, ANI, vitals)</td>
<td>AUROC 0.93 (facial); better than ANI/vitals</td>
</tr>
<tr>
<td>Pikulkaew et al. [<xref ref-type="bibr" rid="B35">35</xref>], 2021</td>
<td>UNBC dataset</td>
<td>Adults</td>
<td>Sequential facial images</td>
<td>CNN (DL motion detection)</td>
<td>Precision: 99.7% (no pain), 92.9% (becoming pain), 95.1% (pain)</td>
</tr>
<tr>
<td>Rezaei et al. [<xref ref-type="bibr" rid="B36">36</xref>], 2021</td>
<td>Dementia patients, LTC setting</td>
<td>Older adults, dementia</td>
<td>Unobtrusive video dataset</td>
<td>Deep learning + pairwise/contrastive training</td>
<td>Outperformed baselines; validated on dementia cohort</td>
</tr>
<tr>
<td>Rodriguez et al. [<xref ref-type="bibr" rid="B37">37</xref>], 2022</td>
<td>UNBC + CK</td>
<td>Adults</td>
<td>Raw video frames</td>
<td>CNN + LSTM</td>
<td>Outperformed SOTA AUC (UNBC); competitive on CK</td>
</tr>
<tr>
<td>Semwal and Londhe [<xref ref-type="bibr" rid="B38">38</xref>], 2024</td>
<td>Multimodal dataset</td>
<td>Adults</td>
<td>Facial + multimodal integration</td>
<td>Multi-stream spatio-temporal network</td>
<td>Showed robust multiparametric pain assessment</td>
</tr>
<tr>
<td>Tan et al. [<xref ref-type="bibr" rid="B39">39</xref>], 2025</td>
<td>200 patients</td>
<td>Adults perioperative/interventional</td>
<td>Video recording (STA-LSTM)</td>
<td>STA-LSTM DL network</td>
<td>Accuracy, sensitivity, recall, F1 ≈ 0.92; clinical feasibility</td>
</tr>
<tr>
<td>Yuan et al. [<xref ref-type="bibr" rid="B40">40</xref>], 2024</td>
<td>ICU, public + 2 new datasets</td>
<td>Critically ill adults (ventilated)</td>
<td>Facial occlusion management</td>
<td>AU-guided CNN framework</td>
<td>Superior performance in binary, 4-class, regression tasks</td>
</tr>
<tr>
<td>Zhang et al. [<xref ref-type="bibr" rid="B41">41</xref>], 2025</td>
<td>503 postop patients + volunteers</td>
<td>Adults postoperative</td>
<td>Clinical Pain Dataset (CPD; 3,411 images) + Simulated Pain Dataset (CD)</td>
<td>VGG16 pretrained</td>
<td>AUROC 0.898 (CPD severe pain), 0.867 (CD); software prototype developed</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p id="t1-fn-1">AI: artificial intelligence; ResNet: Residual Network; UNBC: University of Northern British Columbia Pain Expression dataset; MIntPAIN: Multimodal International Pain dataset; DL: deep learning; CNN: convolutional neural network; RNN: recurrent neural network; EDLM: ensemble deep learning model; ROC: receiver operating characteristic; ICU: intensive care unit; NEVVA: Non-Verbal Visual Analog device; AUs: action units; AUROC: area under the receiver operating characteristic curve; YOLOv8: You Only Look Once version 8; JPR: Journal of Pain Research; ML: machine learning; AUC: area under the curve; MIL: multiple instance learning; MCIL: multiple clustered instance learning; VIVAE: Visual and Vocal Acute Expression dataset; VGG: visual geometry group; LSTM: long short-term memory; ConvLSTM: convolutional long short-term memory; SOTA: state-of-the-art; sEMG: surface electromyography; ErAS-Net: enhanced residual attention-based subject-specific network; ANI: analgesia nociception index; LTC: long-term care; CK: Cohn-Kanade dataset; STA-LSTM: Spatio-Temporal Attention Long Short-Term Memory; CD: Control Dataset.</p>
</fn>
</table-wrap-foot>
</table-wrap>
</sec>
<sec id="t3-3">
<title>Risk of bias in studies</title>
<p id="p-28">Since most included studies were observational, experimental, or methodological (not RCTs), the RoB 2 tool was adapted to the specific study designs. Specifically, the category “Low risk” was assigned when the methods, datasets, and validation were clearly reported; “Some concerns” was used when limitations such as small samples, lack of external validation, or simulation-only data were present (<xref ref-type="table" rid="t2">Table 2</xref>).</p>
<table-wrap id="t2">
<label>Table 2</label>
<caption>
<p id="t2-p-1">
<bold>Risk of bias of included studies.</bold>
</p>
</caption>
<table frame="hsides" rules="groups">
<thead>
<tr>
<th>Author/Year</th>
<th>Country</th>
<th>Intervention/AI approach</th>
<th>Timing</th>
<th>Outcomes measurement</th>
<th>Validation of tool (Y/N)</th>
<th>Quality assessment (RoB 2 overall)</th>
</tr>
</thead>
<tbody>
<tr>
<td>Bargshady et al. [<xref ref-type="bibr" rid="B17">17</xref>], 2024</td>
<td>Australia/USA</td>
<td>Vision transformer</td>
<td>Acute pain datasets</td>
<td>Accuracy, comparison with baselines</td>
<td>Y</td>
<td>Low risk (well-reported external datasets)</td>
</tr>
<tr>
<td>Bargshady et al. [<xref ref-type="bibr" rid="B18">18</xref>], 2020</td>
<td>Australia/Netherlands</td>
<td>Ensemble CNN + RNN</td>
<td>Lab datasets</td>
<td>Accuracy, ROC</td>
<td>Y</td>
<td>Some concerns (no external clinical validation)</td>
</tr>
<tr>
<td>Bellal et al. [<xref ref-type="bibr" rid="B19">19</xref>], 2024</td>
<td>France</td>
<td>NEVVA<sup>®</sup> device (AI facial)</td>
<td>ICU pilot</td>
<td>Device calibration vs. experts</td>
<td>Y</td>
<td>Some concerns (small sample, feasibility only)</td>
</tr>
<tr>
<td>Benavent-Lledo et al. [<xref ref-type="bibr" rid="B20">20</xref>], 2023</td>
<td>Spain</td>
<td>Transformer-based CV</td>
<td>Lab datasets</td>
<td>Accuracy, F1</td>
<td>Y</td>
<td>Low risk (robust datasets, transparent methods)</td>
</tr>
<tr>
<td>Cascella et al. [<xref ref-type="bibr" rid="B21">21</xref>], 2024</td>
<td>Italy</td>
<td>Binary AU-based classifier</td>
<td>Oncology outpatient</td>
<td>Accuracy, AUROC</td>
<td>Y</td>
<td>Some concerns (limited clinical cohort)</td>
</tr>
<tr>
<td>Cascella et al. [<xref ref-type="bibr" rid="B22">22</xref>], 2024</td>
<td>Italy</td>
<td>Multimodal (speech + facial)</td>
<td>Clinical trial NCT04726228</td>
<td>Classification accuracy</td>
<td>Y</td>
<td>Low risk (registered trial, multimodal)</td>
</tr>
<tr>
<td>Cascella et al. [<xref ref-type="bibr" rid="B23">23</xref>], 2023</td>
<td>Italy</td>
<td>YOLOv8</td>
<td>Lab/clinical feasibility</td>
<td>Detection metrics</td>
<td>Y</td>
<td>Some concerns (pilot, limited validation)</td>
</tr>
<tr>
<td>Casti et al. [<xref ref-type="bibr" rid="B24">24</xref>], 2019</td>
<td>Italy</td>
<td>DL pain intensity system</td>
<td>Lab</td>
<td>Accuracy, calibration</td>
<td>Y</td>
<td>Low risk (strong methodological rigor)</td>
</tr>
<tr>
<td>Casti et al. [<xref ref-type="bibr" rid="B25">25</xref>], 2021</td>
<td>Italy</td>
<td>Transfer entropy + ML</td>
<td>Lab</td>
<td>Accuracy, robustness</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Chen et al. [<xref ref-type="bibr" rid="B26">26</xref>], 2022</td>
<td>USA</td>
<td>AU combinations + MIL</td>
<td>Clinical + lab</td>
<td>Accuracy, AUC</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Dutta and M [<xref ref-type="bibr" rid="B27">27</xref>], 2018</td>
<td>India</td>
<td>Hybrid DL</td>
<td>Lab + simulated</td>
<td>Accuracy, computational metrics</td>
<td>Y</td>
<td>Some concerns (older methods, limited clinical data)</td>
</tr>
<tr>
<td>Ghosh et al. [<xref ref-type="bibr" rid="B28">28</xref>], 2025</td>
<td>India/Switzerland</td>
<td>Multimodal (facial + audio)</td>
<td>Lab datasets</td>
<td>Accuracy (2–5 classes)</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Guo et al. [<xref ref-type="bibr" rid="B29">29</xref>], 2021</td>
<td>China</td>
<td>CNN/LSTM</td>
<td>Cold pressor</td>
<td>F1 score</td>
<td>Y</td>
<td>Some concerns (small sample)</td>
</tr>
<tr>
<td>Heintz et al. [<xref ref-type="bibr" rid="B30">30</xref>], 2025</td>
<td>USA multicenter</td>
<td>CNN-based</td>
<td>Perioperative</td>
<td>AUROC, Brier score</td>
<td>Y</td>
<td>Low risk (robust clinical dataset)</td>
</tr>
<tr>
<td>Mao et al. [<xref ref-type="bibr" rid="B31">31</xref>], 2025</td>
<td>China</td>
<td>Conv-Transformer multitask</td>
<td>Lab</td>
<td>Regression + classification</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Mieronkoski et al. [<xref ref-type="bibr" rid="B32">32</xref>], 2020</td>
<td>Finland</td>
<td>sEMG + ML</td>
<td>Experimental pain</td>
<td>c-index, features</td>
<td>Y</td>
<td>Some concerns (small sample, modest accuracy)</td>
</tr>
<tr>
<td>Morsali and Ghaffari [<xref ref-type="bibr" rid="B33">33</xref>], 2025</td>
<td>Iran/UK</td>
<td>ErAS-Net</td>
<td>Lab datasets</td>
<td>Accuracy, cross-dataset</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Park et al. [<xref ref-type="bibr" rid="B34">34</xref>], 2024</td>
<td>Korea</td>
<td>ML (facial, ANI, vitals)</td>
<td>Postoperative</td>
<td>AUROC</td>
<td>Y</td>
<td>Low risk (clinical real-world)</td>
</tr>
<tr>
<td>Pikulkaew et al. [<xref ref-type="bibr" rid="B35">35</xref>], 2021</td>
<td>Thailand</td>
<td>CNN</td>
<td>Lab</td>
<td>Precision, accuracy</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Rezaei et al. [<xref ref-type="bibr" rid="B36">36</xref>], 2021</td>
<td>Canada</td>
<td>DL</td>
<td>Long-term care</td>
<td>Sensitivity, specificity</td>
<td>Y</td>
<td>Low risk (validated on target population)</td>
</tr>
<tr>
<td>Rodriguez et al. [<xref ref-type="bibr" rid="B37">37</xref>], 2022</td>
<td>Spain/Denmark</td>
<td>CNN + LSTM</td>
<td>Lab</td>
<td>AUC, accuracy</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Semwal and Londhe [<xref ref-type="bibr" rid="B38">38</xref>], 2024</td>
<td>India</td>
<td>Spatio-temporal network</td>
<td>Lab</td>
<td>Accuracy</td>
<td>Y</td>
<td>Some concerns (no external validation)</td>
</tr>
<tr>
<td>Tan et al. [<xref ref-type="bibr" rid="B39">39</xref>], 2025</td>
<td>Singapore</td>
<td>STA-LSTM</td>
<td>Clinical</td>
<td>Accuracy, F1</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Yuan et al. [<xref ref-type="bibr" rid="B40">40</xref>], 2024</td>
<td>China</td>
<td>AU-guided CNN</td>
<td>ICU, ventilated pts</td>
<td>Accuracy, regression</td>
<td>Y</td>
<td>Low risk</td>
</tr>
<tr>
<td>Zhang et al. [<xref ref-type="bibr" rid="B41">41</xref>], 2025</td>
<td>China</td>
<td>VGG16 pretrained</td>
<td>Postoperative</td>
<td>AUROC, F1</td>
<td>Y</td>
<td>Low risk</td>
</tr>
</tbody>
</table>
<table-wrap-foot>
<fn>
<p id="t2-fn-1">AI: artificial intelligence; CNN: convolutional neural network; RNN: recurrent neural network; ROC: receiver operating characteristic; NEVVA: Non-Verbal Visual Analog device; ICU: intensive care unit; CV: computer vision; AU: action unit; AUROC: area under the receiver operating characteristic curve; YOLOv8: You Only Look Once version 8; ML: machine learning; MIL: multiple instance learning; AUC: area under the curve; DL: deep learning; LSTM: long short-term memory; sEMG: surface electromyography; ErAS-Net: enhanced residual attention-based subject-specific network; ANI: analgesia nociception index; STA-LSTM: Spatio-Temporal Attention Long Short-Term Memory.</p>
</fn>
</table-wrap-foot>
</table-wrap>
<p id="p-29">Overall, we assigned low risk of bias to studies with robust datasets, transparent methodology, external validation, or those conducted within the framework of a registered clinical trial. Studies judged as having some concerns were typically characterized by small sample sizes, absence of external validation, or approaches tested only in laboratory-controlled settings, which can limit their generalizability. None of the studies included clearly fell into the category of high risk of bias.</p>
</sec>
<sec id="t3-4">
<title>Results of individual studies</title>
<p id="p-30">The included studies reported heterogeneous outcomes reflecting the performance of AI applied to facial expression recognition for pain assessment. Most studies evaluated models on either publicly available datasets [e.g., UNBC-McMaster Shoulder Pain, BioVid Heat Pain, Multimodal International Pain dataset (MIntPAIN)] or on original clinical cohorts (perioperative, postoperative, oncology, or ICU patients). Outcomes were mainly expressed as accuracy, AUROC, sensitivity, specificity, recall, and F1-scores. Confidence intervals were rarely reported.</p>
<p id="p-31">Experimental and dataset-based studies demonstrated very high performance. Bargshady et al. [<xref ref-type="bibr" rid="B18">18</xref>] proposed an ensemble deep learning framework integrating CNN and RNN, reaching &gt; 89% accuracy and AUC 0.93 on MIntPAIN and UNBC. Chen et al. [<xref ref-type="bibr" rid="B26">26</xref>] introduced a weakly supervised approach based on combinations of AUs, achieving 87% accuracy with AUC 0.94 on UNBC. Morsali and Ghaffari [<xref ref-type="bibr" rid="B33">33</xref>] developed ErAS-Net, an attention-based deep learning model, achieving 98.7% accuracy for binary and 94.2% for four-class pain classification, with cross-dataset validation on BioVid still showing robust results (78.1%). Rodriguez et al. [<xref ref-type="bibr" rid="B37">37</xref>] combined CNNs and LSTMs for temporal analysis, outperforming state-of-the-art methods on UNBC. Pikulkaew et al. [<xref ref-type="bibr" rid="B35">35</xref>] applied deep learning to 2D motion and expressions, achieving &gt; 95% accuracy across three pain classes.</p>
<p id="p-32">Transformer-based methods showed promise. Bargshady et al. [<xref ref-type="bibr" rid="B17">17</xref>] employed video ViTs (ViViTs) on AI4PAIN and BioVid datasets, with accuracies of 66.9% and 79.9%, outperforming CNN baselines. Mao et al. [<xref ref-type="bibr" rid="B31">31</xref>] refined Conv-Transformer architectures with multi-task learning, improving estimation of continuous pain intensities on UNBC.</p>
<p id="p-33">Clinical studies confirmed feasibility in real-world settings. Zhang et al. [<xref ref-type="bibr" rid="B41">41</xref>] applied VGG16 to &gt; 3,000 images from 503 postoperative patients, reporting AUROC 0.898 for severe pain detection and consistent F1-scores. Heintz et al. [<xref ref-type="bibr" rid="B30">30</xref>] validated CNN-based nociception recognition perioperatively, reporting robust AUROC and calibration (Brier score) across multicenter datasets. Park et al. [<xref ref-type="bibr" rid="B34">34</xref>] compared models in 155 gastrectomy patients, showing that facial-expression machine learning achieved an AUROC 0.93 for severe postoperative pain, outperforming analgesia nociception index (ANI) and vital signs. Bellal et al. [<xref ref-type="bibr" rid="B19">19</xref>] tested the NEVVA<sup>©</sup> device in ICU patients, showing the feasibility of automated detection in non-communicative patients.</p>
<p id="p-34">Other methodological innovations included electromyography of facial muscles [<xref ref-type="bibr" rid="B32">32</xref>], showing modest predictive capacity (c-index 0.64), and transfer entropy applied to facial landmarks [<xref ref-type="bibr" rid="B25">25</xref>], which demonstrated robustness to uncertainty. Earlier, Casti et al. [<xref ref-type="bibr" rid="B24">24</xref>] benchmarked multi-expert calibration, confirming reproducibility of automated systems. Moreover, Dutta and M [<xref ref-type="bibr" rid="B27">27</xref>] provided proof-of-concept evidence for real-time video analysis.</p>
<p id="p-35">Special populations were investigated. Rezaei et al. [<xref ref-type="bibr" rid="B36">36</xref>] validated a deep learning system in older adults with dementia, showing reliable detection where self-report is not possible. Yuan et al. [<xref ref-type="bibr" rid="B40">40</xref>] developed an AU-guided CNN for ICU patients with occluded faces (ventilation), reporting strong performance across binary, multiclass, and regression tasks. On the other hand, oncology-related contributions are limited. In 2023, Cascella et al. [<xref ref-type="bibr" rid="B23">23</xref>] developed a binary classifier on cancer patients’ videos, achieving ~94% accuracy and an AUROC of 0.98. Later, the same authors integrated facial and vocal features in oncologic pain monitoring, confirming feasibility in clinical use [<xref ref-type="bibr" rid="B21">21</xref>] and subsequently also tested the You Only Look Once version 8 (YOLOv8) architecture for real-time pain detection, reporting strong detection performance [<xref ref-type="bibr" rid="B22">22</xref>].</p>
<p id="p-36">Several multimodal approaches combined facial expression with additional modalities. Benavent-Lledo et al. [<xref ref-type="bibr" rid="B20">20</xref>] used computer vision and multimodal signals, achieving &gt; 96% accuracy and &gt; 94% F1-score on UNBC and BioVid. Ghosh et al. [<xref ref-type="bibr" rid="B28">28</xref>] proposed an IoT-enabled system integrating facial and audio data, reaching near-perfect accuracy (&gt; 99%) in binary and three-class classification. Semwal and Londhe [<xref ref-type="bibr" rid="B38">38</xref>] designed a spatio-temporal behavioral system with multimodal integration, confirming the added value of combining sources.</p>
<p id="p-37">In summary, across the 25 included studies [<xref ref-type="bibr" rid="B17">17</xref>–<xref ref-type="bibr" rid="B41">41</xref>], AI-based facial expression recognition for pain consistently demonstrated high performance in controlled datasets and increasing feasibility in clinical populations (postoperative, perioperative, oncology, ICU, dementia). The overall result ranged from modest (with an accuracy c-index of 0.64) [<xref ref-type="bibr" rid="B32">32</xref>] to excellent (with an accuracy &gt; 98%) [<xref ref-type="bibr" rid="B33">33</xref>]. <xref ref-type="fig" rid="fig2">Figure 2</xref> illustrates performance (i.e., accuracy) ranges across different AI methods.</p>
<fig id="fig2" position="float">
<label>Figure 2</label>
<caption>
<p id="fig2-p-1">
<bold>AI methods used for facial expression-based pain assessment (2015–2025) and their reported performance ranges (i.e., accuracies).</bold> Traditional machine learning (ML) approaches showed moderate performance (65–85%). Convolutional neural networks (CNNs) and RNN/hybrid (recurrent neural network) models achieved higher accuracy (70–95% and 75–93%, respectively). Transformer-based models reached accuracies ranging from 66% to 98%. Multimodal approaches (facial + audio and/or physiological signals) consistently outperformed unimodal systems, achieving accuracies up to 99.5%.</p>
</caption>
<graphic xmlns:xlink="http://www.w3.org/1999/xlink" xlink:href="em-06-1001370-g002.tif" />
</fig>
</sec>
</sec>
<sec id="s4">
<title>Discussion</title>
<p id="p-38">This systematic review highlights both the promise and the limitations of AI applied to facial expression analysis for APA. Across 25 studies published between 2015 and 2025, AI-based systems demonstrated consistently strong performance in controlled settings, with reported accuracies often exceeding 90% and AUROC values ranging between 0.87 and 0.98 in clinical studies [<xref ref-type="bibr" rid="B21">21</xref>, <xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B34">34</xref>, <xref ref-type="bibr" rid="B41">41</xref>]. Collectively, these results support the feasibility of APA recognition and underscore the potential of AI to provide objective, real-time assessments in contexts where traditional self-report is unreliable or unavailable. On the other hand, it is important to note that not all approaches yielded successful outcomes. Some studies, for instance, reported modest or even poor performance. Specifically, performance varied markedly depending on study design, population, and dataset source. For example, models trained and validated exclusively on benchmark datasets (e.g., UNBC-McMaster, BioVid, MIntPAIN) achieved very high accuracies, sometimes exceeding 95% [<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B26">26</xref>, <xref ref-type="bibr" rid="B33">33</xref>, <xref ref-type="bibr" rid="B35">35</xref>, <xref ref-type="bibr" rid="B37">37</xref>], although their generalizability to clinical cohorts was limited. In contrast, studies conducted in real-world clinical populations, including perioperative patients [<xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B39">39</xref>], oncology cohorts [<xref ref-type="bibr" rid="B21">21</xref>–<xref ref-type="bibr" rid="B23">23</xref>], ICU patients [<xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B40">40</xref>], and older adults with dementia [<xref ref-type="bibr" rid="B36">36</xref>], reported slightly lower but more clinically relevant performance. Importantly, this gap is more evident for analyses relying on small datasets, limited facial visibility (e.g., ICU settings), or when applying methods such as sEMG-based models that achieved only moderate predictive value (c-index 0.64) [<xref ref-type="bibr" rid="B32">32</xref>]. These negative or suboptimal results highlight the fragility of certain approaches and underscore the need for robust, diverse, and clinically validated datasets. Moreover, when comparing studies using benchmark datasets with those relying on clinical populations, it emerges that while benchmark-based models frequently reported very high accuracies (&gt; 90%) [<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B26">26</xref>, <xref ref-type="bibr" rid="B33">33</xref>, <xref ref-type="bibr" rid="B35">35</xref>, <xref ref-type="bibr" rid="B37">37</xref>], their clinical generalizability was often limited. Conversely, clinical studies, although achieving slightly lower performance, provided more realistic insights into real-world feasibility and robustness [<xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B21">21</xref>–<xref ref-type="bibr" rid="B23">23</xref>, <xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B34">34</xref>, <xref ref-type="bibr" rid="B36">36</xref>, <xref ref-type="bibr" rid="B40">40</xref>, <xref ref-type="bibr" rid="B41">41</xref>].</p>
<p id="p-39">Multimodal approaches integrating facial expressions with speech, audio, or physiological signals generally outperformed unimodal systems [<xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B22">22</xref>, <xref ref-type="bibr" rid="B28">28</xref>, <xref ref-type="bibr" rid="B38">38</xref>], particularly in challenging scenarios such as ICU patients [<xref ref-type="bibr" rid="B40">40</xref>] or dementia patients [<xref ref-type="bibr" rid="B36">36</xref>]. Nevertheless, these models also introduced greater complexity, raising issues of computational burden, implementation feasibility, and interpretability in clinical practice.</p>
<p id="p-40">Concerning the AI-based strategies, methodological evolution across the last decade reflects a shift from classical CNN- and RNN-based pipelines to more sophisticated approaches involving transformer-based and multimodal systems. Specifically, compared to CNNs, ViTs, and hybrid Conv-Transformer models have demonstrated superior ability to capture long-range dependencies and subtle spatio-temporal dynamics in facial expressions [<xref ref-type="bibr" rid="B17">17</xref>, <xref ref-type="bibr" rid="B31">31</xref>]. Similarly, attention-based networks such as ErAS-Net provide efficient feature selection and cross-dataset robustness, achieving accuracies close to 99% in binary pain classification [<xref ref-type="bibr" rid="B33">33</xref>]. In parallel, multimodal fusion architectures that integrate facial cues with audio, speech emotion recognition, or physiological signals have consistently improved F1-scores and AUROC values compared with unimodal models [<xref ref-type="bibr" rid="B20">20</xref>, <xref ref-type="bibr" rid="B22">22</xref>, <xref ref-type="bibr" rid="B28">28</xref>, <xref ref-type="bibr" rid="B38">38</xref>]. These advances highlight a trend toward increasingly complex frameworks capable of addressing challenging clinical conditions, such as facial occlusion in ICU patients [<xref ref-type="bibr" rid="B40">40</xref>] or atypical expressions in dementia [<xref ref-type="bibr" rid="B36">36</xref>]. Furthermore, the integration of explainable AI (XAI) techniques and standardized multimodal datasets may represent key steps to improve transparency, scalability, and cross-institutional generalizability.</p>
<p id="p-41">The risk of bias was overall moderate. Studies based on larger, multicentre clinical cohorts with transparent methodology and external validation were considered low risk [<xref ref-type="bibr" rid="B21">21</xref>–<xref ref-type="bibr" rid="B23">23</xref>, <xref ref-type="bibr" rid="B30">30</xref>, <xref ref-type="bibr" rid="B34">34</xref>, <xref ref-type="bibr" rid="B41">41</xref>], while smaller experimental works without external validation raised concerns [<xref ref-type="bibr" rid="B27">27</xref>, <xref ref-type="bibr" rid="B32">32</xref>]. Importantly, no study was deemed consistently at high risk of bias. Furthermore, selective reporting remains an issue as many experimental studies emphasized accuracy, F1-score, or AUROC [<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B26">26</xref>, <xref ref-type="bibr" rid="B29">29</xref>, <xref ref-type="bibr" rid="B33">33</xref>, <xref ref-type="bibr" rid="B37">37</xref>], while failing to report misclassification rates, calibration statistics, or subgroup-specific results. The lack of registered protocols in most works further increases the risk of reporting bias and reduces comparability across studies [<xref ref-type="bibr" rid="B22">22</xref>, <xref ref-type="bibr" rid="B30">30</xref>]. Moreover, a recurrent limitation is the reliance on small, non-representative datasets and the frequent absence of confidence intervals in reporting. These factors increase the risk of overestimating performance and reduce the robustness of the reported findings. From the analysis, other key issues emerged. For example, we underline that while systematic use of calibration metrics and external validation is crucial for clinical translation, their absence in most studies represents a major barrier to clinical adoption and should be prioritized in future research.</p>
<p id="p-42">Multimodal models integrating facial data with physiological or audio signals outperformed unimodal approaches, particularly in challenging clinical scenarios. Moreover, the certainty of the evidence is best described as moderate for binary classification tasks (pain vs. no pain), where results are consistent across different datasets and study designs [<xref ref-type="bibr" rid="B18">18</xref>, <xref ref-type="bibr" rid="B26">26</xref>, <xref ref-type="bibr" rid="B33">33</xref>, <xref ref-type="bibr" rid="B36">36</xref>], but low for more advanced applications such as pain intensity estimation [<xref ref-type="bibr" rid="B29">29</xref>, <xref ref-type="bibr" rid="B31">31</xref>, <xref ref-type="bibr" rid="B32">32</xref>] or deployment in fragile populations [<xref ref-type="bibr" rid="B19">19</xref>, <xref ref-type="bibr" rid="B36">36</xref>, <xref ref-type="bibr" rid="B41">41</xref>]. This downgrading is mainly due to small sample sizes, reliance on controlled datasets, and the absence of precision measures such as confidence intervals. Consequently, dataset-related limitations strongly affected model performance. Thus, restricted sample sizes, lack of demographic diversity, and heterogeneous annotation protocols limited reproducibility and generalizability, especially across populations with diverse ethnic and clinical backgrounds.</p>
<p id="p-43">Despite these interesting results, several challenges should be addressed. The ethical implications of AI-based pain assessment cannot be overlooked. Patient privacy in facial video datasets, the absence of clear data sharing policies, and the limited interpretability of most deep learning models raise concerns regarding fairness, transparency, and responsible use in clinical care.</p>
<p id="p-44">In conclusion, the findings confirm that AI can complement traditional pain assessment, particularly in patients unable to self-report. Nevertheless, key challenges remain before large-scale clinical adoption can be realized. These include the need for standardized datasets reflecting real-world heterogeneity, transparent reporting practices, and multicentre trials for evaluating AI performance across diverse populations and settings.</p>
</sec>
</body>
<back>
<glossary>
<title>Abbreviations</title>
<def-list>
<def-item>
<term>AI</term>
<def>
<p>artificial intelligence</p>
</def>
</def-item>
<def-item>
<term>ANI</term>
<def>
<p>analgesia nociception index</p>
</def>
</def-item>
<def-item>
<term>APA</term>
<def>
<p>automatic pain assessment</p>
</def>
</def-item>
<def-item>
<term>AUC</term>
<def>
<p>area under the curve</p>
</def>
</def-item>
<def-item>
<term>AUROC</term>
<def>
<p>area under the receiver operating characteristic curve</p>
</def>
</def-item>
<def-item>
<term>AUs</term>
<def>
<p>action units</p>
</def>
</def-item>
<def-item>
<term>CNNs</term>
<def>
<p>convolutional neural networks</p>
</def>
</def-item>
<def-item>
<term>ErAS-Net</term>
<def>
<p>enhanced residual attention-based subject-specific network</p>
</def>
</def-item>
<def-item>
<term>FACS</term>
<def>
<p>Facial Action Coding System</p>
</def>
</def-item>
<def-item>
<term>ICU</term>
<def>
<p>intensive care unit</p>
</def>
</def-item>
<def-item>
<term>LSTM</term>
<def>
<p>long short-term memory</p>
</def>
</def-item>
<def-item>
<term>MIntPAIN</term>
<def>
<p>Multimodal International Pain dataset</p>
</def>
</def-item>
<def-item>
<term>RNN</term>
<def>
<p>recurrent neural network</p>
</def>
</def-item>
<def-item>
<term>sEMG</term>
<def>
<p>surface electromyography</p>
</def>
</def-item>
<def-item>
<term>UNBC</term>
<def>
<p>University of Northern British Columbia Pain Expression dataset</p>
</def>
</def-item>
<def-item>
<term>ViTs</term>
<def>
<p>vision transformers</p>
</def>
</def-item>
<def-item>
<term>ViViTs</term>
<def>
<p>video vision transformers</p>
</def>
</def-item>
</def-list>
</glossary>
<sec id="s-suppl" sec-type="supplementary-material">
<title>Supplementary materials</title>
<p>The supplementary table for this article is available at: <uri xlink:href="https://www.explorationpub.com/uploads/Article/file/1001370_sup_1.pdf">https://www.explorationpub.com/uploads/Article/file/1001370_sup_1.pdf</uri>.</p>
<supplementary-material id="SD1" content-type="local-data">
<media xlink:href="1001370_sup_1.pdf" mimetype="application" mime-subtype="pdf"></media>
</supplementary-material>
</sec>
<sec id="s6">
<title>Declarations</title>
<sec id="t-6-1">
<title>Author contributions</title>
<p>MC: Writing—original draft, Formal analysis, Conceptualization, Methodology, Supervision, Writing—review &amp; editing. DE: Writing—review &amp; editing. MRM: Writing—review &amp; editing. V Cascella: Writing—review &amp; editing. V Cerrone: Writing—review &amp; editing, Conceptualization, Methodology, Supervision. All authors read and approved the submitted version.</p>
</sec>
<sec id="t-6-2" sec-type="COI-statement">
<title>Conflicts of interest</title>
<p>Marco Cascella, who is the Editorial Board Member and Guest Editor of Exploration of Medicine, had no involvement in the decision-making or the review process of this manuscript. The other authors declare no conflicts of interest.</p>
</sec>
<sec id="t-6-3">
<title>Ethical approval</title>
<p>Not applicable.</p>
</sec>
<sec id="t-6-4">
<title>Consent to participate</title>
<p>Not applicable.</p>
</sec>
<sec id="t-6-5">
<title>Consent to publication</title>
<p>Not applicable.</p>
</sec>
<sec id="t-6-6" sec-type="data-availability">
<title>Availability of data and materials</title>
<p>The primary data for this review were sourced online from databases listed in the methods. Referenced articles are accessible on PubMed/MEDLINE, Scopus, Cochrane Library, Web of Science, and IEEE Xplore. All data are available from the corresponding author on reasonable request.</p>
</sec>
<sec id="t-6-7">
<title>Funding</title>
<p>No external funding was received for this study.</p>
</sec>
<sec id="t-6-8">
<title>Copyright</title>
<p>© The Author(s) 2025.</p>
</sec>
</sec>
<sec id="s7">
<title>Publisher’s note</title>
<p>Open Exploration maintains a neutral stance on jurisdictional claims in published institutional affiliations and maps. All opinions expressed in this article are the personal views of the author(s) and do not represent the stance of the editorial team or the publisher.</p>
</sec>
<ref-list>
<ref id="B1">
<label>1</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Albahdal</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Aljebreen</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Ibrahim</surname>
<given-names>DM</given-names>
</name>
</person-group>
<article-title>PainMeter: Automatic Assessment of Pain Intensity Levels From Multiple Physiological Signals Using Machine Learning</article-title>
<source>IEEE Access</source>
<year iso-8601-date="2024">2024</year>
<volume>12</volume>
<fpage>48349</fpage>
<lpage>65</lpage>
<pub-id pub-id-type="doi">10.1109/access.2024.3384359</pub-id>
</element-citation>
</ref>
<ref id="B2">
<label>2</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cascella</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Leoni</surname>
<given-names>MLG</given-names>
</name>
<name>
<surname>Shariff</surname>
<given-names>MN</given-names>
</name>
<name>
<surname>Varrassi</surname>
<given-names>G</given-names>
</name>
</person-group>
<article-title>Artificial Intelligence-Driven Diagnostic Processes and Comprehensive Multimodal Models in Pain Medicine</article-title>
<source>J Pers Med</source>
<year iso-8601-date="2024">2024</year>
<volume>14</volume>
<elocation-id>983</elocation-id>
<pub-id pub-id-type="doi">10.3390/jpm14090983</pub-id>
<pub-id pub-id-type="pmid">39338237</pub-id>
<pub-id pub-id-type="pmcid">PMC11432921</pub-id>
</element-citation>
</ref>
<ref id="B3">
<label>3</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cao</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Fu</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Wermter</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Liu</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>H</given-names>
</name>
</person-group>
<article-title>Pain recognition and pain empathy from a human-centered AI perspective</article-title>
<source>iScience</source>
<year iso-8601-date="2024">2024</year>
<volume>27</volume>
<elocation-id>110570</elocation-id>
<pub-id pub-id-type="doi">10.1016/j.isci.2024.110570</pub-id>
<pub-id pub-id-type="pmid">39211548</pub-id>
<pub-id pub-id-type="pmcid">PMC11357883</pub-id>
</element-citation>
</ref>
<ref id="B4">
<label>4</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Miao</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Han</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Zhu</surname>
<given-names>Y</given-names>
</name>
</person-group>
<article-title>Recognizing Facial Expressions Using a Shallow Convolutional Neural Network</article-title>
<source>IEEE Access</source>
<year iso-8601-date="2019">2019</year>
<volume>7</volume>
<fpage>78000</fpage>
<lpage>11</lpage>
<pub-id pub-id-type="doi">10.1109/access.2019.2921220</pub-id>
</element-citation>
</ref>
<ref id="B5">
<label>5</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Hassan</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Seus</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Wollenberg</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Weitz</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Kunz</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Lautenbacher</surname>
<given-names>S</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Automatic Detection of Pain from Facial Expressions: A Survey</article-title>
<source>IEEE Trans Pattern Anal Mach Intell</source>
<year iso-8601-date="2021">2021</year>
<volume>43</volume>
<fpage>1815</fpage>
<lpage>31</lpage>
<pub-id pub-id-type="doi">10.1109/TPAMI.2019.2958341</pub-id>
<pub-id pub-id-type="pmid">31825861</pub-id>
</element-citation>
</ref>
<ref id="B6">
<label>6</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zen</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Porzi</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Sangineto</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Ricci</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Sebe</surname>
<given-names>N</given-names>
</name>
</person-group>
<article-title>Learning Personalized Models for Facial Expression Analysis and Gesture Recognition</article-title>
<source>IEEE Trans. Multimedia</source>
<year iso-8601-date="2016">2016</year>
<volume>18</volume>
<fpage>775</fpage>
<lpage>88</lpage>
<pub-id pub-id-type="doi">10.1109/tmm.2016.2523421</pub-id>
</element-citation>
</ref>
<ref id="B7">
<label>7</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Gkikas</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Tsiknakis</surname>
<given-names>M</given-names>
</name>
</person-group>
<article-title>Automatic assessment of pain based on deep learning methods: A systematic review</article-title>
<source>Comput Methods Programs Biomed</source>
<year iso-8601-date="2023">2023</year>
<volume>231</volume>
<elocation-id>107365</elocation-id>
<pub-id pub-id-type="doi">10.1016/j.cmpb.2023.107365</pub-id>
<pub-id pub-id-type="pmid">36764062</pub-id>
</element-citation>
</ref>
<ref id="B8">
<label>8</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cascella</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Shariff</surname>
<given-names>MN</given-names>
</name>
</person-group>
<article-title>Why is applying artificial intelligence to pain so challenging?</article-title>
<source>Curr Med Res Opin</source>
<year iso-8601-date="2024">2024</year>
<volume>40</volume>
<fpage>2021</fpage>
<lpage>4</lpage>
<pub-id pub-id-type="doi">10.1080/03007995.2024.2434078</pub-id>
<pub-id pub-id-type="pmid">39579196</pub-id>
</element-citation>
</ref>
<ref id="B9">
<label>9</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cascella</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Ponsiglione</surname>
<given-names>AM</given-names>
</name>
<name>
<surname>Santoriello</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Romano</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Cerrone</surname>
<given-names>V</given-names>
</name>
<name>
<surname>Esposito</surname>
<given-names>D</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Expert consensus on feasibility and application of automatic pain assessment in routine clinical use</article-title>
<source>J Anesth Analg Crit Care</source>
<year iso-8601-date="2025">2025</year>
<volume>5</volume>
<elocation-id>29</elocation-id>
<pub-id pub-id-type="doi">10.1186/s44158-025-00249-8</pub-id>
<pub-id pub-id-type="pmid">40457422</pub-id>
<pub-id pub-id-type="pmcid">PMC12131339</pub-id>
</element-citation>
</ref>
<ref id="B10">
<label>10</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Page</surname>
<given-names>MJ</given-names>
</name>
<name>
<surname>McKenzie</surname>
<given-names>JE</given-names>
</name>
<name>
<surname>Bossuyt</surname>
<given-names>PM</given-names>
</name>
<name>
<surname>Boutron</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Hoffmann</surname>
<given-names>TC</given-names>
</name>
<name>
<surname>Mulrow</surname>
<given-names>CD</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>The PRISMA 2020 statement: an updated guideline for reporting systematic reviews</article-title>
<source>BMJ</source>
<year iso-8601-date="2021">2021</year>
<volume>372</volume>
<elocation-id>n71</elocation-id>
<pub-id pub-id-type="doi">10.1136/bmj.n71</pub-id>
<pub-id pub-id-type="pmid">33782057</pub-id>
<pub-id pub-id-type="pmcid">PMC8005924</pub-id>
</element-citation>
</ref>
<ref id="B11">
<label>11</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ouzzani</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Hammady</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Fedorowicz</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Elmagarmid</surname>
<given-names>A</given-names>
</name>
</person-group>
<article-title>Rayyan—a web and mobile app for systematic reviews</article-title>
<source>Syst Rev</source>
<year iso-8601-date="2016">2016</year>
<volume>5</volume>
<elocation-id>210</elocation-id>
<pub-id pub-id-type="doi">10.1186/s13643-016-0384-4</pub-id>
<pub-id pub-id-type="pmid">27919275</pub-id>
<pub-id pub-id-type="pmcid">PMC5139140</pub-id>
</element-citation>
</ref>
<ref id="B12">
<label>12</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Minozzi</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Cinquini</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Gianola</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Gonzalez-Lorenzo</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Banzi</surname>
<given-names>R</given-names>
</name>
</person-group>
<article-title>The revised Cochrane risk of bias tool for randomized trials (RoB 2) showed low interrater reliability and challenges in its application</article-title>
<source>J Clin Epidemiol</source>
<year iso-8601-date="2020">2020</year>
<volume>126</volume>
<fpage>37</fpage>
<lpage>44</lpage>
<pub-id pub-id-type="doi">10.1016/j.jclinepi.2020.06.015</pub-id>
<pub-id pub-id-type="pmid">32562833</pub-id>
</element-citation>
</ref>
<ref id="B13">
<label>13</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chan</surname>
<given-names>DXH</given-names>
</name>
<name>
<surname>Tan</surname>
<given-names>CW</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Teo</surname>
<given-names>JC</given-names>
</name>
<name>
<surname>Wong</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Tan</surname>
<given-names>YR</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>LP001 Automated pain detection via facial expression for adult patients using artificial intelligence</article-title>
<source>Reg Anesth Pain Med</source>
<year iso-8601-date="2024">2024</year>
<volume>49</volume>
<elocation-id>A410</elocation-id>
<pub-id pub-id-type="doi">10.1136/rapm-2024-esra.586</pub-id>
</element-citation>
</ref>
<ref id="B14">
<label>14</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>López</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Palacios-Arias</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Romeu</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Jofre-Roca</surname>
<given-names>L</given-names>
</name>
</person-group>
<article-title>3D Pain Face Expression Recognition Using a ML-MIMO Radar Profiler</article-title>
<source>IEEE Access</source>
<year iso-8601-date="2024">2024</year>
<volume>12</volume>
<fpage>48266</fpage>
<lpage>76</lpage>
<pub-id pub-id-type="doi">10.1109/access.2024.3383143</pub-id>
</element-citation>
</ref>
<ref id="B15">
<label>15</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pu</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Coppieters</surname>
<given-names>MW</given-names>
</name>
<name>
<surname>Smalbrugge</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Jones</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Byrnes</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Todorovic</surname>
<given-names>M</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Associations between facial expressions and observational pain in residents with dementia and chronic pain</article-title>
<source>J Adv Nurs</source>
<year iso-8601-date="2024">2024</year>
<volume>80</volume>
<fpage>3846</fpage>
<lpage>55</lpage>
<pub-id pub-id-type="doi">10.1111/jan.16063</pub-id>
<pub-id pub-id-type="pmid">38334268</pub-id>
</element-citation>
</ref>
<ref id="B16">
<label>16</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Uddin</surname>
<given-names>MT</given-names>
</name>
<name>
<surname>Zamzmi</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Canavan</surname>
<given-names>S</given-names>
</name>
</person-group>
<article-title>Cooperative Learning for Personalized Context-Aware Pain Assessment From Wearable Data</article-title>
<source>IEEE J Biomed Health Inform</source>
<year iso-8601-date="2023">2023</year>
<volume>27</volume>
<fpage>5260</fpage>
<lpage>71</lpage>
<pub-id pub-id-type="doi">10.1109/JBHI.2023.3294903</pub-id>
<pub-id pub-id-type="pmid">37440405</pub-id>
</element-citation>
</ref>
<ref id="B17">
<label>17</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bargshady</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Joseph</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Hirachan</surname>
<given-names>N</given-names>
</name>
<name>
<surname>Goecke</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Rojas</surname>
<given-names>RF</given-names>
</name>
</person-group>
<article-title>Acute Pain Recognition from Facial Expression Videos using Vision Transformers</article-title>
<source>Annu Int Conf IEEE Eng Med Biol Soc</source>
<year iso-8601-date="2024">2024</year>
<volume>2024</volume>
<fpage>1</fpage>
<lpage>4</lpage>
<pub-id pub-id-type="doi">10.1109/EMBC53108.2024.10781616</pub-id>
<pub-id pub-id-type="pmid">40039359</pub-id>
</element-citation>
</ref>
<ref id="B18">
<label>18</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bargshady</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Zhou</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Deo</surname>
<given-names>RC</given-names>
</name>
<name>
<surname>Soar</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Whittaker</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>H</given-names>
</name>
</person-group>
<article-title>Ensemble neural network approach detecting pain intensity from facial expressions</article-title>
<source>Artif Intell Med</source>
<year iso-8601-date="2020">2020</year>
<volume>109</volume>
<elocation-id>101954</elocation-id>
<pub-id pub-id-type="doi">10.1016/j.artmed.2020.101954</pub-id>
<pub-id pub-id-type="pmid">34756219</pub-id>
</element-citation>
</ref>
<ref id="B19">
<label>19</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Bellal</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Lelandais</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Chabin</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Heudron</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Gourmelon</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Bauduin</surname>
<given-names>P</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Calibration trial of an innovative medical device (<italic>NEVVA<sup>©</sup></italic>) for the evaluation of pain in non-communicating patients in the intensive care unit</article-title>
<source>Front Med (Lausanne)</source>
<year iso-8601-date="2024">2024</year>
<volume>11</volume>
<elocation-id>1309720</elocation-id>
<pub-id pub-id-type="doi">10.3389/fmed.2024.1309720</pub-id>
<pub-id pub-id-type="pmid">38994344</pub-id>
<pub-id pub-id-type="pmcid">PMC11236545</pub-id>
</element-citation>
</ref>
<ref id="B20">
<label>20</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Benavent-Lledo</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Mulero-Pérez</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Ortiz-Perez</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Rodriguez-Juan</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Berenguer-Agullo</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Psarrou</surname>
<given-names>A</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>A Comprehensive Study on Pain Assessment from Multimodal Sensor Data</article-title>
<source>Sensors (Basel)</source>
<year iso-8601-date="2023">2023</year>
<volume>23</volume>
<elocation-id>9675</elocation-id>
<pub-id pub-id-type="doi">10.3390/s23249675</pub-id>
<pub-id pub-id-type="pmid">38139521</pub-id>
<pub-id pub-id-type="pmcid">PMC10747670</pub-id>
</element-citation>
</ref>
<ref id="B21">
<label>21</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cascella</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Cutugno</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Mariani</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Vitale</surname>
<given-names>VN</given-names>
</name>
<name>
<surname>Iuorio</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Cuomo</surname>
<given-names>A</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>AI-based cancer pain assessment through speech emotion recognition and video facial expressions classification</article-title>
<source>Signa Vitae</source>
<year iso-8601-date="2024">2024</year>
<volume>20;28–38</volume>
<pub-id pub-id-type="doi">10.22514/sv.2024.153</pub-id>
</element-citation>
</ref>
<ref id="B22">
<label>22</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cascella</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Shariff</surname>
<given-names>MN</given-names>
</name>
<name>
<surname>Bianco</surname>
<given-names>GL</given-names>
</name>
<name>
<surname>Monaco</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Gargano</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Simonini</surname>
<given-names>A</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Employing the Artificial Intelligence Object Detection Tool YOLOv8 for Real-Time Pain Detection: A Feasibility Study</article-title>
<source>J Pain Res</source>
<year iso-8601-date="2024">2024</year>
<volume>17</volume>
<fpage>3681</fpage>
<lpage>96</lpage>
<pub-id pub-id-type="doi">10.2147/JPR.S491574</pub-id>
<pub-id pub-id-type="pmid">39540033</pub-id>
<pub-id pub-id-type="pmcid">PMC11559421</pub-id>
</element-citation>
</ref>
<ref id="B23">
<label>23</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Cascella</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Vitale</surname>
<given-names>VN</given-names>
</name>
<name>
<surname>Mariani</surname>
<given-names>F</given-names>
</name>
<name>
<surname>Iuorio</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Cutugno</surname>
<given-names>F</given-names>
</name>
</person-group>
<article-title>Development of a binary classifier model from extended facial codes toward video-based pain recognition in cancer patients</article-title>
<source>Scand J Pain</source>
<year iso-8601-date="2023">2023</year>
<volume>23</volume>
<fpage>638</fpage>
<lpage>45</lpage>
<pub-id pub-id-type="doi">10.1515/sjpain-2023-0011</pub-id>
<pub-id pub-id-type="pmid">37665749</pub-id>
</element-citation>
</ref>
<ref id="B24">
<label>24</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Casti</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Mencattini</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Comes</surname>
<given-names>MC</given-names>
</name>
<name>
<surname>Callari</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Di</surname>
<given-names>Giuseppe D</given-names>
</name>
<name>
<surname>Natoli</surname>
<given-names>S</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Calibration of Vision-Based Measurement of Pain Intensity With Multiple Expert Observers</article-title>
<source>IEEE Trans Instrum. Meas</source>
<year iso-8601-date="2019">2019</year>
<volume>68</volume>
<fpage>2442</fpage>
<lpage>50</lpage>
<pub-id pub-id-type="doi">10.1109/tim.2019.2909603</pub-id>
</element-citation>
</ref>
<ref id="B25">
<label>25</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Casti</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Mencattini</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Filippi</surname>
<given-names>J</given-names>
</name>
<name>
<surname>DʼOrazio</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Comes</surname>
<given-names>MC</given-names>
</name>
<name>
<surname>Giuseppe</surname>
<given-names>DD</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Metrological Characterization of a Pain Detection System Based on Transfer Entropy of Facial Landmarks</article-title>
<source>IEEE Trans Instrum Meas</source>
<year iso-8601-date="2021">2021</year>
<volume>70</volume>
<fpage>1</fpage>
<lpage>8</lpage>
<pub-id pub-id-type="doi">10.1109/tim.2021.3067611</pub-id>
</element-citation>
</ref>
<ref id="B26">
<label>26</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Chen</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Ansari</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Wilkie</surname>
<given-names>DJ</given-names>
</name>
</person-group>
<article-title>Learning Pain from Action Unit Combinations: A Weakly Supervised Approach via Multiple Instance Learning</article-title>
<source>IEEE Trans Affect Comput</source>
<year iso-8601-date="2022">2022</year>
<volume>13</volume>
<fpage>135</fpage>
<lpage>46</lpage>
<pub-id pub-id-type="doi">10.1109/taffc.2019.2949314</pub-id>
<pub-id pub-id-type="pmid">35242282</pub-id>
<pub-id pub-id-type="pmcid">PMC8890070</pub-id>
</element-citation>
</ref>
<ref id="B27">
<label>27</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Dutta</surname>
<given-names>P</given-names>
</name>
<name>
<surname>M</surname>
<given-names>N</given-names>
</name>
</person-group>
<article-title>Facial Pain Expression Recognition in Real-Time Videos</article-title>
<source>J Healthc Eng</source>
<year iso-8601-date="2018">2018</year>
<volume>2018</volume>
<elocation-id>7961427</elocation-id>
<pub-id pub-id-type="doi">10.1155/2018/7961427</pub-id>
<pub-id pub-id-type="pmid">30510672</pub-id>
<pub-id pub-id-type="pmcid">PMC6231392</pub-id>
</element-citation>
</ref>
<ref id="B28">
<label>28</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Ghosh</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Umer</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Dhara</surname>
<given-names>BC</given-names>
</name>
<name>
<surname>Ali</surname>
<given-names>GGMN</given-names>
</name>
</person-group>
<article-title>A Multimodal Pain Sentiment Analysis System Using Ensembled Deep Learning Approaches for IoT-Enabled Healthcare Framework</article-title>
<source>Sensors (Basel)</source>
<year iso-8601-date="2025">2025</year>
<volume>25</volume>
<elocation-id>1223</elocation-id>
<pub-id pub-id-type="doi">10.3390/s25041223</pub-id>
<pub-id pub-id-type="pmid">40006452</pub-id>
<pub-id pub-id-type="pmcid">PMC11860027</pub-id>
</element-citation>
</ref>
<ref id="B29">
<label>29</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Guo</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Wang</surname>
<given-names>L</given-names>
</name>
<name>
<surname>Xiao</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>Y</given-names>
</name>
</person-group>
<article-title>A Personalized Spatial-Temporal Cold Pain Intensity Estimation Model Based on Facial Expression</article-title>
<source>IEEE J Transl Eng Health Med</source>
<year iso-8601-date="2021">2021</year>
<volume>9</volume>
<elocation-id>4901008</elocation-id>
<pub-id pub-id-type="doi">10.1109/JTEHM.2021.3116867</pub-id>
<pub-id pub-id-type="pmid">34650836</pub-id>
<pub-id pub-id-type="pmcid">PMC8500272</pub-id>
</element-citation>
</ref>
<ref id="B30">
<label>30</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Heintz</surname>
<given-names>TA</given-names>
</name>
<name>
<surname>Badathala</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Wooten</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Cu</surname>
<given-names>CW</given-names>
</name>
<name>
<surname>Wallace</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Pham</surname>
<given-names>B</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Preliminary Development and Validation of Automated Nociception Recognition Using Computer Vision in Perioperative Patients</article-title>
<source>Anesthesiology</source>
<year iso-8601-date="2025">2025</year>
<volume>142</volume>
<fpage>726</fpage>
<lpage>37</lpage>
<pub-id pub-id-type="doi">10.1097/ALN.0000000000005370</pub-id>
<pub-id pub-id-type="pmid">39804295</pub-id>
</element-citation>
</ref>
<ref id="B31">
<label>31</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mao</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Luo</surname>
<given-names>Y</given-names>
</name>
<name>
<surname>Gou</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Qi</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Li</surname>
<given-names>T</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Multi-Task Hybrid Conv-Transformer With Emotional Localized Ambiguity Exploration for Facial Pain Assessment</article-title>
<source>IEEE J Biomed Health Inform</source>
<year iso-8601-date="2025">2025</year>
<volume>29</volume>
<fpage>5133</fpage>
<lpage>45</lpage>
<pub-id pub-id-type="doi">10.1109/JBHI.2025.3533308</pub-id>
<pub-id pub-id-type="pmid">40031827</pub-id>
</element-citation>
</ref>
<ref id="B32">
<label>32</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mieronkoski</surname>
<given-names>R</given-names>
</name>
<name>
<surname>Syrjälä</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Jiang</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Rahmani</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Pahikkala</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Liljeberg</surname>
<given-names>P</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Developing a pain intensity prediction model using facial expression: A feasibility study with electromyography</article-title>
<source>PLoS One</source>
<year iso-8601-date="2020">2020</year>
<volume>15</volume>
<elocation-id>e0235545</elocation-id>
<pub-id pub-id-type="doi">10.1371/journal.pone.0235545</pub-id>
<pub-id pub-id-type="pmid">32645045</pub-id>
<pub-id pub-id-type="pmcid">PMC7347182</pub-id>
</element-citation>
</ref>
<ref id="B33">
<label>33</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Morsali</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Ghaffari</surname>
<given-names>A</given-names>
</name>
</person-group>
<article-title>Enhanced residual attention-based subject-specific network (ErAS-Net): facial expression-based pain classification with multiple attention mechanisms</article-title>
<source>Sci Rep</source>
<year iso-8601-date="2025">2025</year>
<volume>15</volume>
<elocation-id>19425</elocation-id>
<pub-id pub-id-type="doi">10.1038/s41598-025-04552-w</pub-id>
<pub-id pub-id-type="pmid">40461564</pub-id>
<pub-id pub-id-type="pmcid">PMC12134349</pub-id>
</element-citation>
</ref>
<ref id="B34">
<label>34</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Park</surname>
<given-names>I</given-names>
</name>
<name>
<surname>Park</surname>
<given-names>JH</given-names>
</name>
<name>
<surname>Yoon</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Na</surname>
<given-names>H</given-names>
</name>
<name>
<surname>Oh</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Ryu</surname>
<given-names>J</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Machine learning model of facial expression outperforms models using analgesia nociception index and vital signs to predict postoperative pain intensity: a pilot study</article-title>
<source>Korean J Anesthesiol</source>
<year iso-8601-date="2024">2024</year>
<volume>77</volume>
<fpage>195</fpage>
<lpage>204</lpage>
<pub-id pub-id-type="doi">10.4097/kja.23583</pub-id>
<pub-id pub-id-type="pmid">38176698</pub-id>
<pub-id pub-id-type="pmcid">PMC10982524</pub-id>
</element-citation>
</ref>
<ref id="B35">
<label>35</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Pikulkaew</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Boonchieng</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Boonchieng</surname>
<given-names>E</given-names>
</name>
<name>
<surname>Chouvatut</surname>
<given-names>V</given-names>
</name>
</person-group>
<article-title>2D Facial Expression and Movement of Motion for Pain Identification With Deep Learning Methods</article-title>
<source>IEEE Access</source>
<year iso-8601-date="2021">2021</year>
<volume>9</volume>
<fpage>109903</fpage>
<lpage>14</lpage>
<pub-id pub-id-type="doi">10.1109/access.2021.3101396</pub-id>
</element-citation>
</ref>
<ref id="B36">
<label>36</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rezaei</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Moturu</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Prkachin</surname>
<given-names>KM</given-names>
</name>
<name>
<surname>Hadjistavropoulos</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Taati</surname>
<given-names>B</given-names>
</name>
</person-group>
<article-title>Unobtrusive Pain Monitoring in Older Adults With Dementia Using Pairwise and Contrastive Training</article-title>
<source>IEEE J Biomed Health Inform</source>
<year iso-8601-date="2021">2021</year>
<volume>25</volume>
<fpage>1450</fpage>
<lpage>62</lpage>
<pub-id pub-id-type="doi">10.1109/JBHI.2020.3045743</pub-id>
<pub-id pub-id-type="pmid">33338024</pub-id>
</element-citation>
</ref>
<ref id="B37">
<label>37</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Rodriguez</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Cucurull</surname>
<given-names>G</given-names>
</name>
<name>
<surname>Gonzalez</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Gonfaus</surname>
<given-names>JM</given-names>
</name>
<name>
<surname>Nasrollahi</surname>
<given-names>K</given-names>
</name>
<name>
<surname>Moeslund</surname>
<given-names>TB</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Deep Pain: Exploiting Long Short-Term Memory Networks for Facial Expression Classification</article-title>
<source>IEEE Trans Cybern</source>
<year iso-8601-date="2022">2022</year>
<volume>52</volume>
<fpage>3314</fpage>
<lpage>24</lpage>
<pub-id pub-id-type="doi">10.1109/TCYB.2017.2662199</pub-id>
<pub-id pub-id-type="pmid">28207407</pub-id>
</element-citation>
</ref>
<ref id="B38">
<label>38</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Semwal</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Londhe</surname>
<given-names>ND</given-names>
</name>
</person-group>
<article-title>A multi-stream spatio-temporal network based behavioural multiparametric pain assessment system</article-title>
<source>Biomed Signal Process Control</source>
<year iso-8601-date="2024">2024</year>
<volume>90</volume>
<elocation-id>105820</elocation-id>
<pub-id pub-id-type="doi">10.1016/j.bspc.2023.105820</pub-id>
</element-citation>
</ref>
<ref id="B39">
<label>39</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Tan</surname>
<given-names>CW</given-names>
</name>
<name>
<surname>Du</surname>
<given-names>T</given-names>
</name>
<name>
<surname>Teo</surname>
<given-names>JC</given-names>
</name>
<name>
<surname>Chan</surname>
<given-names>DXH</given-names>
</name>
<name>
<surname>Kong</surname>
<given-names>WM</given-names>
</name>
<name>
<surname>Sng</surname>
<given-names>BL</given-names>
</name>
</person-group>
<article-title>Automated pain detection using facial expression in adult patients with a customized spatial temporal attention long short-term memory (STA-LSTM) network</article-title>
<source>Sci Rep</source>
<year iso-8601-date="2025">2025</year>
<volume>15</volume>
<elocation-id>13429</elocation-id>
<pub-id pub-id-type="doi">10.1038/s41598-025-97885-5</pub-id>
<pub-id pub-id-type="pmid">40251301</pub-id>
<pub-id pub-id-type="pmcid">PMC12008390</pub-id>
</element-citation>
</ref>
<ref id="B40">
<label>40</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Yuan</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Cui</surname>
<given-names>Z</given-names>
</name>
<name>
<surname>Xu</surname>
<given-names>D</given-names>
</name>
<name>
<surname>Zhang</surname>
<given-names>S</given-names>
</name>
<name>
<surname>Zhao</surname>
<given-names>C</given-names>
</name>
<name>
<surname>Wu</surname>
<given-names>X</given-names>
</name>
<etal>et al.</etal>
</person-group>
<article-title>Occluded Facial Pain Assessment in the ICU using Action Units Guided Network</article-title>
<source>IEEE J Biomed Health Inform</source>
<year iso-8601-date="2024">2024</year>
<volume>28</volume>
<fpage>438</fpage>
<lpage>49</lpage>
<pub-id pub-id-type="doi">10.1109/JBHI.2023.3336157</pub-id>
<pub-id pub-id-type="pmid">37995171</pub-id>
</element-citation>
</ref>
<ref id="B41">
<label>41</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Zhang</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Hu</surname>
<given-names>X</given-names>
</name>
<name>
<surname>Duan</surname>
<given-names>W</given-names>
</name>
<name>
<surname>Ji</surname>
<given-names>M</given-names>
</name>
<name>
<surname>Yang</surname>
<given-names>J</given-names>
</name>
</person-group>
<article-title>Application of deep learning-based facial pain recognition model for postoperative pain assessment</article-title>
<source>J Clin Anesth</source>
<year iso-8601-date="2025">2025</year>
<volume>105</volume>
<elocation-id>111898</elocation-id>
<pub-id pub-id-type="doi">10.1016/j.jclinane.2025.111898</pub-id>
<pub-id pub-id-type="pmid">40516197</pub-id>
</element-citation>
</ref>
<ref id="B42">
<label>42</label>
<element-citation publication-type="journal">
<person-group person-group-type="author">
<name>
<surname>Mende-Siedlecki</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Qu-Lee</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Lin</surname>
<given-names>J</given-names>
</name>
<name>
<surname>Drain</surname>
<given-names>A</given-names>
</name>
<name>
<surname>Goharzad</surname>
<given-names>A</given-names>
</name>
</person-group>
<article-title>The Delaware Pain Database: a set of painful expressions and corresponding norming data</article-title>
<source>Pain Rep</source>
<year iso-8601-date="2020">2020</year>
<volume>5</volume>
<elocation-id>e853</elocation-id>
<pub-id pub-id-type="doi">10.1097/PR9.0000000000000853</pub-id>
<pub-id pub-id-type="pmid">33134750</pub-id>
<pub-id pub-id-type="pmcid">PMC7587421</pub-id>
</element-citation>
</ref>
<ref id="B43">
<label>43</label>
<element-citation publication-type="confproc">
<person-group person-group-type="author">
<name>
<surname>Lucey</surname>
<given-names>P</given-names>
</name>
<name>
<surname>Cohn</surname>
<given-names>JF</given-names>
</name>
<name>
<surname>Prkachin</surname>
<given-names>KM</given-names>
</name>
<name>
<surname>Solomon</surname>
<given-names>PE</given-names>
</name>
<name>
<surname>Matthews</surname>
<given-names>I</given-names>
</name>
</person-group>
<comment>Painful data: The UNBC-McMaster shoulder pain expression archive database. In: Proceedings of 2011 IEEE International Conference on Automatic Face &amp; Gesture Recognition (FG); 2011 Mar 21-25; Santa Barbara, CA, USA. IEEE; 2011. pp. 57–64.</comment>
<pub-id pub-id-type="doi">10.1109/fg.2011.5771462</pub-id>
</element-citation>
</ref>
</ref-list>
</back>
</article>