Transformer Meets Gated Residual Networks to Enhance PICU’s PPG Artifact Detection Informed by Mutual Information Neural Estimation

  • Thanh Dung Le
  • , Clara Macabiau
  • , Kevin Albert
  • , Symeon Chatzinotas
  • , Philippe Jouvet
  • , Rita Noumeir

Résultats de recherche: Contribution à un journalArticle publié dans une revue, révisé par les pairsRevue par des pairs

Résumé

This study delves into the effectiveness of various learning methods in improving Transformer models, focusing mainly on the Gated Residual Network (GRN) Transformer in the context of pediatric intensive care units (PICUs) with limited data availability. Our findings indicate that Transformers trained via supervised learning are less effective than MLP, CNN, and LSTM networks in such environments. Yet, leveraging unsupervised and self-supervised learning (SSL) on unannotated data, with subsequent fine-tuning on annotated data, notably enhances Transformer performance, although not to the level of the GRN–Transformer. Central to our research is analyzing different activation functions for the gated linear unit (GLU), a crucial element of the GRN structure. We also employ Mutual Information Neural Estimation (MINE) to evaluate the GRN’s contribution. Additionally, the study examines the effects of integrating GRN within the Transformer’s attention mechanism versus using it as a separate intermediary layer. Our results highlight that GLU with sigmoid activation stands out, achieving 0.98 accuracy, 0.91 precision, 0.96 recall, and 0.94~F1 -score. The MINE analysis supports the hypothesis that GRN enhances the mutual information (MI) between the hidden representations and the output. Moreover, using GRN as an intermediate filter layer proves more beneficial than incorporating it within the Attention mechanism. This study clarifies how GRN boosters GRN–Transformer’s performance surpasses other techniques. These findings offer a promising avenue for adopting sophisticated models like Transformers in data-constrained environments, such as PPG artifact detection in PICU settings.

langue originaleAnglais
journalIEEE Transactions on Neural Networks and Learning Systems
Les DOIs
étatAccepté/Sous presse - 2026
Modification externeOui

Empreinte digitale

Voici les principaux termes ou expressions associés à « Transformer Meets Gated Residual Networks to Enhance PICU’s PPG Artifact Detection Informed by Mutual Information Neural Estimation ». Ces libellés thématiques sont générés à partir du titre et du résumé de la publication. Ensemble, ils forment une empreinte digitale unique.

Contient cette citation