Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Temporal Cross-Attention for Dynamic Embedding and Tokenization of Multimodal Electronic Health Records (2403.04012v2)

Published 6 Mar 2024 in cs.LG

Abstract: The breadth, scale, and temporal granularity of modern electronic health records (EHR) systems offers great potential for estimating personalized and contextual patient health trajectories using sequential deep learning. However, learning useful representations of EHR data is challenging due to its high dimensionality, sparsity, multimodality, irregular and variable-specific recording frequency, and timestamp duplication when multiple measurements are recorded simultaneously. Although recent efforts to fuse structured EHR and unstructured clinical notes suggest the potential for more accurate prediction of clinical outcomes, less focus has been placed on EHR embedding approaches that directly address temporal EHR challenges by learning time-aware representations from multimodal patient time series. In this paper, we introduce a dynamic embedding and tokenization framework for precise representation of multimodal clinical time series that combines novel methods for encoding time and sequential position with temporal cross-attention. Our embedding and tokenization framework, when integrated into a multitask transformer classifier with sliding window attention, outperformed baseline approaches on the exemplar task of predicting the occurrence of nine postoperative complications of more than 120,000 major inpatient surgeries using multimodal data from three hospitals and two academic health centers in the United States.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (41)
  1. A deep learning–based dynamic model for predicting acute kidney injury risk severity in postoperative patients. Surgery, 2023.
  2. Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150, 2020.
  3. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555, 2014.
  4. Position information in transformers: An overview. Computational Linguistics, 48(3):733–763, 2022.
  5. On the importance of clinical notes in multi-modal learning for ehr data. arXiv preprint arXiv:2212.03044, 2022.
  6. Analysis of free text in electronic health records for identification of cancer patient trajectories. Scientific reports, 7(1):46226, 2017.
  7. Time2vec: Learning a vector representation of time. arXiv preprint arXiv:1907.05321, 2019.
  8. 1d convolutional neural networks and applications: A survey. Mechanical systems and signal processing, 151:107398, 2021.
  9. Multimodal data fusion: an overview of methods, challenges, and prospects. Proceedings of the IEEE, 103(9):1449–1477, 2015.
  10. Behrt: transformer for electronic health records. Scientific reports, 10(1):7155, 2020.
  11. Hi-behrt: Hierarchical transformer-based model for accurate prediction of clinical events using multimodal longitudinal electronic health records. IEEE journal of biomedical and health informatics, 27(2):1106–1117, 2022a.
  12. Clinical-longformer and clinical-bigbird: Transformers for long clinical sequences. arXiv preprint arXiv:2201.11838, 2022b.
  13. Learn from relational correlations and periodic events for temporal knowledge graph reasoning. In Proceedings of the 46th International ACM SIGIR Conference on Research and Development in Information Retrieval, pp.  1559–1568, 2023.
  14. High-modality multimodal transformer: Quantifying modality & interaction heterogeneity for high-modality representation learning. Transactions on Machine Learning Research, 2022.
  15. Attention-based multimodal fusion with contrast for robust clinical prediction in the face of missing modalities. Journal of Biomedical Informatics, 145:104466, 2023.
  16. Dynamic sepsis prediction for intensive care unit patients using xgboost-based model with novel time-dependent features. IEEE Journal of Biomedical and Health Informatics, 26(8):4258–4269, 2022.
  17. Vilbert: Pretraining task-agnostic visiolinguistic representations for vision-and-language tasks. Advances in neural information processing systems, 32, 2019.
  18. A multimodal transformer: Fusing clinical notes with structured ehr data for interpretable in-hospital mortality prediction. In AMIA Annual Symposium Proceedings, volume 2022, pp.  719. American Medical Informatics Association, 2022.
  19. How noisy is too noisy? the impact of data noise on multimodal recognition of confusion and conflict during collaborative learning. In Proceedings of the 25th International Conference on Multimodal Interaction, pp.  326–335, 2023.
  20. Long Short-Term Memory. Long short-term memory. Neural computation, 9(8):1735–1780, 2010.
  21. Foundation models for generalist medical artificial intelligence. Nature, 616(7956):259–265, 2023.
  22. Time series prediction using deep learning methods in healthcare. ACM Transactions on Management Information Systems, 14(1):1–29, 2023.
  23. Big data analytics in healthcare: promise and potential. Health information science and systems, 2:1–10, 2014.
  24. Self-attention with relative position representations. arXiv preprint arXiv:1803.02155, 2018.
  25. Deep dynamic imputation of clinical time series for mortality prediction. Information Sciences, 579:607–622, 2021.
  26. The dawn of multimodal artificial intelligence in nephrology. Nature Reviews Nephrology, pp.  1–2, 2023.
  27. Multi-dimensional patient acuity estimation with longitudinal ehr tokenization and flexible transformer networks. Frontiers in Digital Health, 4:1029191, 2022.
  28. Dynamic predictions of postoperative complications from explainable, uncertainty-aware, and multi-task deep neural networks. Scientific Reports, 13(1):1224, 2023.
  29. Data-gru: Dual-attention time-aware gated recurrent unit for irregular multivariate time series. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 34, pp.  930–937, 2020.
  30. Self-supervised transformer for sparse and irregularly sampled multivariate clinical time-series. ACM Transactions on Knowledge Discovery from Data (TKDD), 16(6):1–17, 2022.
  31. Utilizing imbalanced electronic health records to predict acute kidney injury by ensemble learning and time series model. BMC Medical Informatics and Decision Making, 20(1):1–13, 2020a.
  32. Transmodality: An end2end fusion method with transformer for multimodal sentiment analysis. In Proceedings of The Web Conference 2020, pp.  2514–2520, 2020b.
  33. Position-aware self-attention based neural sequence labeling. Pattern Recognition, 110:107636, 2021.
  34. Transformers in medical image segmentation: A review. Biomedical Signal Processing and Control, 84:104791, 2023.
  35. Multimodal learning with transformers: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  36. Big bird: Transformers for longer sequences. Advances in neural information processing systems, 33:17283–17297, 2020.
  37. Are transformers effective for time series forecasting? In Proceedings of the AAAI conference on artificial intelligence, volume 37, pp.  11121–11128, 2023.
  38. Time-aware transformer-based network for clinical notes series prediction. In Machine learning for healthcare conference, pp.  566–588. PMLR, 2020.
  39. Graph-guided network for irregularly sampled multivariate time series. arXiv preprint arXiv:2110.05357, 2021.
  40. Improving medical predictions by irregular multimodal electronic health records modeling. In International Conference on Machine Learning, pp.  41300–41313. PMLR, 2023.
  41. Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. In The Eleventh International Conference on Learning Representations, 2022.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (13)
  1. Yingbo Ma (20 papers)
  2. Suraj Kolla (3 papers)
  3. Dhruv Kaliraman (2 papers)
  4. Victoria Nolan (2 papers)
  5. Zhenhong Hu (9 papers)
  6. Ziyuan Guan (20 papers)
  7. Yuanfang Ren (24 papers)
  8. Brooke Armfield (5 papers)
  9. Tezcan Ozrazgat-Baslanti (32 papers)
  10. Tyler J. Loftus (15 papers)
  11. Parisa Rashidi (59 papers)
  12. Azra Bihorac (51 papers)
  13. Benjamin Shickel (24 papers)
Citations (1)
X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets