Papers
Topics
Authors
Recent
Search
2000 character limit reached

Transformer Meets Gated Residual Networks To Enhance Photoplethysmogram Artifact Detection Informed by Mutual Information Neural Estimation

Published 25 May 2024 in eess.SP | (2405.16177v2)

Abstract: This study delves into the effectiveness of various learning methods in improving Transformer models, focusing particularly on the Gated Residual Network Transformer (GRN-Transformer) in the context of pediatric intensive care units (PICU) with limited data availability. Our findings indicate that Transformers trained via supervised learning are less effective compared to MLP, CNN, and LSTM networks in such environments. Yet, leveraging unsupervised and self-supervised learning on unannotated data, with subsequent fine-tuning on annotated data, notably enhances Transformer performance, although not to the level of the GRN-Transformer. Central to our research is the analysis of different activation functions for the Gated Linear Unit (GLU), a crucial element of the GRN structure. We also employ Mutual Information Neural Estimation (MINE) to evaluate the GRN's contribution. Additionally, the study examines the effects of integrating GRN within the Transformer's Attention mechanism versus using it as a separate intermediary layer. Our results highlight that GLU with sigmoid activation stands out, achieving 0.98 accuracy, 0.91 precision, 0.96 recall, and 0.94 F1 score. The MINE analysis supports the hypothesis that GRN enhances the mutual information between the hidden representations and the output. Moreover, the use of GRN as an intermediate filter layer proves more beneficial than incorporating it within the Attention mechanism. In summary, this research clarifies how GRN bolsters GRN-Transformer's performance, surpassing other learning techniques. These findings offer a promising avenue for adopting sophisticated models like Transformers in data-constrained environments, such as PPG artifact detection in PICU settings.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (27)
  1. D. Brossier, R. El Taani, M. Sauthier, N. Roumeliotis, G. Emeriaud, and P. Jouvet, “Creating a high-frequency electronic database in the picu: the perpetual patient,” Pediatr. Crit. Care Med., vol. 19, no. 4, pp. e189–e198, 2018.
  2. N. Roumeliotis, G. Parisien, S. Charette, E. Arpin, F. Brunet, and P. Jouvet, “Reorganizing care with the implementation of electronic medical records: a time-motion study in the picu,” Pediatr. Crit. Care Med., vol. 19, no. 4, pp. e172–e179, 2018.
  3. A. Mathieu and et. al., “Validation process of a high-resolution database in a pediatric intensive care unit—describing the perpetual patient’s validation,” Journal of Evaluation in Clinical Practice, vol. 27, no. 2, pp. 316–324, 2021.
  4. A. C. Dziorny and et. al., “Clinical decision support in the picu: Implications for design and evaluation,” Pediatr. Crit. Care Med., vol. 23, no. 8, pp. e392–e396, 2022.
  5. T.-D. Le and et. al., “Detecting of a patient’s condition from clinical narratives using natural language representation,” IEEE Open J. Eng. Med. Biol., vol. 3, pp. 142–149, 2022.
  6. M. Sauthier, G. Tuli, P. A. Jouvet, J. S. Brownstein, and A. G. Randolph, “Estimated pao2: A continuous and noninvasive method to estimate pao2 and oxygenation index,” Critical care explorations, vol. 3, no. 10, 2021.
  7. G. Emeriaud, Y. M. López-Fernández, N. P. Iyer, M. M. Bembea, A. Agulnik, R. P. Barbaro, F. Baudin, A. Bhalla, W. B. De Carvalho, C. L. Carroll, et al., “Executive summary of the second international guidelines for the diagnosis and management of pediatric acute respiratory distress syndrome (palicc-2),” Pediatr. Crit. Care Med., vol. 24, no. 2, p. 143, 2023.
  8. P. Jouvet and et. al., “A pilot prospective study on closed loop controlled ventilation and oxygenation in ventilated children during the weaning phase,” Critical Care, vol. 16, no. 3, pp. 1–9, 2012.
  9. M. Wysocki, P. Jouvet, and S. Jaber, “Closed loop mechanical ventilation,” J. Clin. Monit. Comput., vol. 28, pp. 49–56, 2014.
  10. B. L. Hill and et. al., “Imputation of the continuous arterial line blood pressure waveform from non-invasive measurements using deep learning,” Scientific reports, vol. 11, no. 1, p. 15755, 2021.
  11. F. Fan and et. al., “Estimating spo 2 via time-efficient high-resolution harmonics analysis and maximum likelihood tracking,” IEEE J. Biomed. Health Inform., vol. 22, no. 4, pp. 1075–1086, 2017.
  12. C. Macabiau, T.-D. Le, K. Albert, P. Jouvet, and R. Noumeir, “Label propagation techniques for artifact detection in imbalanced classes using photoplethysmogram signals,” arXiv preprint arXiv:2308.08480, 2023.
  13. T.-D. Le, C. Macabiau, K. Albert, P. Jouvet, and R. Noumeir, “Grn-transformer: Enhancing motion artifact detection in picu photoplethysmogram signals,” arXiv preprint arXiv:2308.03722, 2023.
  14. T.-D. Le, “Boosting transformer’s robustness and efficacy in PPG signal artifact detection with self-supervised learning,” arXiv preprint arXiv:2401.01013, 2024.
  15. M. I. Belghazi, A. Baratin, S. Rajeshwar, S. Ozair, Y. Bengio, A. Courville, and D. Hjelm, “Mutual information neural estimation,” in International conference on machine learning.   PMLR, 2018, pp. 531–540.
  16. T.-D. Le and et. al., “A small-scale switch transformer and nlp-based model for clinical narratives classification,” arXiv preprint arXiv:2303.12892, 2023.
  17. M. Clara and et al., “Label propagation techniques for artifact detection in imbalanced classes using photoplethysmograph signals,” Submitted to IEEE for possible publication, 2023.
  18. S. H. Lee and et. al., “Vision transformer for small-size datasets,” arXiv preprint arXiv:2112.13492, 2021.
  19. R. Shao and X.-J. Bi, “Transformers meet small datasets,” IEEE Access, vol. 10, pp. 118 454–118 464, 2022.
  20. M. Hahn, “Theoretical limitations of self-attention in neural sequence models,” Trans. Assoc. Comput. Linguist., vol. 8, pp. 156–171, 2020.
  21. T. Sattler and et. al., “Understanding the limitations of cnn-based absolute camera pose regression,” in IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 3302–3312.
  22. Y. N. Dauphin and et. al., “Language modeling with gated convolutional networks,” in International Conference on ML, 2017, pp. 933–941.
  23. M. Liu and et. al., “Gated transformer networks for multivariate time series classification,” arXiv preprint arXiv:2103.14438, 2021.
  24. B. Lim, S. Ö. Arık, N. Loeff, and T. Pfister, “Temporal fusion transformers for interpretable multi-horizon time series forecasting,” International Journal of Forecasting, vol. 37, no. 4, pp. 1748–1764, 2021.
  25. P. Savarese and D. Figueiredo, “Residual gates: A simple mechanism for improved network optimization,” in Int. Conf. Learn. Represent., 2017.
  26. A. D. Rasamoelina, F. Adjailia, and P. Sinčák, “A review of activation function for artificial neural network,” in 2020 IEEE 18th World Symposium on Applied Machine Intelligence and Informatics (SAMI).   IEEE, 2020, pp. 281–286.
  27. J. Lederer, “Activation functions in artificial neural networks: A systematic overview,” arXiv preprint arXiv:2101.09957, 2021.
Citations (1)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.