Papers
Topics
Authors
Recent
Search
2000 character limit reached

Optimizing Medication Decisions for Patients with Atrial Fibrillation through Path Development Network

Published 18 Jan 2024 in cs.LG and eess.SP | (2401.10014v1)

Abstract: Atrial fibrillation (AF) is a common cardiac arrhythmia characterized by rapid and irregular contractions of the atria. It significantly elevates the risk of strokes due to slowed blood flow in the atria, especially in the left atrial appendage, which is prone to blood clot formation. Such clots can migrate into cerebral arteries, leading to ischemic stroke. To assess whether AF patients should be prescribed anticoagulants, doctors often use the CHA2DS2-VASc scoring system. However, anticoagulant use must be approached with caution as it can impact clotting functions. This study introduces a machine learning algorithm that predicts whether patients with AF should be recommended anticoagulant therapy using 12-lead ECG data. In this model, we use STOME to enhance time-series data and then process it through a Convolutional Neural Network (CNN). By incorporating a path development layer, the model achieves a specificity of 30.6% under the condition of an NPV of 1. In contrast, LSTM algorithms without path development yield a specificity of only 2.7% under the same NPV condition.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (110)
  1. Mark E. Silverman. From rebellious palpitations to the discovery of auricular fibrillation: Contributions of mackenzie, lewis and einthoven. The American Journal of Cardiology, 73(5):384–389, 1994.
  2. Preventing stroke in patients with atrial fibrillation. JAMA, 281(19):1830–1835, 1999.
  3. Atrial fibrillation as an independent risk factor for stroke: the framingham study. Stroke, 22(8):983–988, 1991.
  4. Predictive value of cha2ds2-vasc scores regarding the risk of stroke and all-cause mortality in patients with atrial fibrillation (consort compliant). Medicine (Baltimore), 98(31):e16560, Aug 2019.
  5. Adverse outcomes and predictors of underuse of antithrombotic therapy in medicare beneficiaries with chronic atrial fibrillation. Stroke, 31(4):822–827, 2000.
  6. A convolutional neural network for ecg annotation as the basis for classification of cardiac rhythms. Physiological Measurement, 39(10):104005, Oct 2018.
  7. Multi-branch fusion network for myocardial infarction screening from 12-lead ecg images. Computer Methods and Programs in Biomedicine, 184:105286, 2020.
  8. G. A. Tadesse et al. Cardiovascular disease diagnosis using cross-domain transfer learning. In 2019 41st Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), pages 4262–4265, Berlin, Germany, 2019.
  9. Heart arrhythmia detection using support vector machines. Intelligent Automation & Soft Computing, 19(1):1–9, 2013.
  10. An adaptive backpropagation neural network for real-time ischemia episodes detection: development and performance analysis using the european st-t database. IEEE Transactions on Biomedical Engineering, 45(7):805–813, July 1998.
  11. Ecg pattern recognition and classification using non-linear transformations and neural networks: A review. International Journal of Medical Informatics, 52(1–3):191–208, 1998.
  12. An effective lstm recurrent network to detect arrhythmia on imbalanced ecg dataset. Journal of Healthcare Engineering, 2019:1–10, 2019.
  13. Automated atrial fibrillation detection using a hybrid cnn-lstm network on imbalanced ecg datasets. Biomedical Signal Processing and Control, 63:102194, 2021.
  14. Imagenet classification with deep convolutional neural networks. In F. Pereira, C.J. Burges, L. Bottou, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 25. Curran Associates, Inc., 2012.
  15. F. Rosenblatt. The Perceptron, a Perceiving and Recognizing Automaton Project Para. Report: Cornell Aeronautical Laboratory. Cornell Aeronautical Laboratory, 1957.
  16. M. Minsky and S. Papert. Perceptrons; an Introduction to Computational Geometry. MIT Press, 1969.
  17. Learning representations by back-propagating errors. Nature, 323:533–536, 1986.
  18. Long short-term memory. Neural Computation, 9(8):1735–1780, 11 1997.
  19. Gradient-based learning applied to document recognition. Proceedings of the IEEE, 86(11):2278–2324, 11 1998.
  20. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 770–778, 2016.
  21. Generative adversarial nets. In Advances in Neural Information Processing Systems (NIPS), volume 27, 2014.
  22. Robustness of neural networks: A probabilistic and practical approach. In 2019 IEEE/ACM 41st International Conference on Software Engineering: New Ideas and Emerging Results (ICSE-NIER), pages 93–96. IEEE, 2019.
  23. Path development network with finite-dimensional lie group representation, 2022.
  24. George V. Cybenko. Approximation by superpositions of a sigmoidal function. Mathematics of Control, Signals and Systems, 2:303–314, 1989.
  25. Multilayer feedforward networks are universal approximators. Neural Networks, 2(5):359–366, 1989.
  26. A Pinkus. Approximation theory of the mlp model in neural networks. Acta Numerica, 8:143–195, 1999.
  27. State-of-the-art in artificial neural network applications: A survey. Heliyon, 4:e00938, 2018.
  28. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics. JMLR Workshop and Conference Proceedings, 2010.
  29. Matus Telgarsky. benefits of depth in neural networks. In Vitaly Feldman, Alexander Rakhlin, and Ohad Shamir, editors, 29th Annual Conference on Learning Theory, volume 49 of Proceedings of Machine Learning Research, pages 1517–1539, Columbia University, New York, New York, USA, 23–26 Jun 2016. PMLR.
  30. On the expressive power of deep neural networks, 2017.
  31. The expressive power of neural networks: A view from the width. In NIPS, 2017.
  32. Deep learning. Nature, 521(7553):436–444, 2015.
  33. Visualizing and understanding convolutional networks. In David Fleet, Tomas Pajdla, Bernt Schiele, and Tinne Tuytelaars, editors, Computer Vision – ECCV 2014, pages 818–833, Cham, 2014. Springer International Publishing.
  34. Visualizing higher-layer features of a deep network. University of Montreal, 1341(3):1, 2009.
  35. What is the best multi-stage architecture for object recognition? In 2009 IEEE 12th international conference on computer vision, pages 2146–2153. IEEE, 2009.
  36. Very deep convolutional networks for large-scale image recognition, 2015.
  37. U-net: Convolutional networks for biomedical image segmentation. In Nassir Navab, Joachim Hornegger, William M. Wells, and Alejandro F. Frangi, editors, Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015, pages 234–241, Cham, 2015. Springer International Publishing.
  38. Violence detection in video by using 3d convolutional neural networks. In Advances in Visual Computing: 10th International Symposium, ISVC 2014, Las Vegas, NV, USA, December 8-10, 2014, Proceedings, Part II, pages 551–558. Springer International Publishing, 2014.
  39. Not all pixels are equal: Difficulty-aware semantic segmentation via deep layer cascade. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 6459–6468. IEEE, 2017.
  40. Tsung-Yi Lin et al. Feature pyramid networks for object detection. In 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 936–944. IEEE, 2017.
  41. Understanding the effective receptive field in deep convolutional neural networks. Advances in neural information processing systems, 29, 2016.
  42. Receptive field block net for accurate and fast object detection. In Vittorio Ferrari, Martial Hebert, Cristian Sminchisescu, and Yair Weiss, editors, Computer Vision – ECCV 2018, pages 404–419, Cham, 2018. Springer International Publishing.
  43. Time series classification with multivariate convolutional neural network. IEEE Transactions on Industrial Electronics, 66(6):4788–4797, 2018.
  44. Temporal convolutional autoencoder for unsupervised anomaly detection in time series. Applied Soft Computing, 112:107751, 2021.
  45. Multi-scale convolutional neural networks for time series classification, 2016.
  46. Yoav Goldberg. A primer on neural network models for natural language processing. Journal of Artificial Intelligence Research, 57:345–420, 2016.
  47. Andrei Ignatov. Real-time human activity recognition from accelerometer data using convolutional neural networks. Applied Soft Computing, 62:915–922, 2018.
  48. Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. IEEE Signal Processing Magazine, 29(6):82–97, 2012.
  49. Time series classification from scratch with deep neural networks: A strong baseline. 2017 International Joint Conference on Neural Networks (IJCNN), pages 1578–1585, 2016.
  50. Deep learning for time series classification: a review. Data Mining and Knowledge Discovery, 33, 7 2019.
  51. Deep learning for time series classification: a review. Data Mining and Knowledge Discovery, 33(4):917–963, 07 2019.
  52. The great time series classification bake off: a review and experimental evaluation of recent algorithmic advances. Data Mining and Knowledge Discovery, 31:606–660, 2017.
  53. Jeffrey L. Elman. Finding structure in time. Cognitive Science, 14(2):179–211, 1990.
  54. Michael Auli and et al. Joint language and translation modeling with recurrent neural networks. In Proceedings of EMNLP, 2013.
  55. Sequence to sequence learning with neural networks. In Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 2, NIPS’14, pages 3104–3112, Cambridge, MA, USA, 2014. MIT Press.
  56. Gradient Flow in Recurrent Nets: The Difficulty of Learning Long-Term Dependencies, pages 237–243. IEEE, 2001.
  57. Lstm: A search space odyssey. IEEE Transactions on Neural Networks and Learning Systems, 28:2222–2232, 2015.
  58. An empirical exploration of recurrent network architectures. In International Conference on Machine Learning, 2015.
  59. Ma-lstm: A multi-attention based lstm for complex pattern extraction. 2020 25th International Conference on Pattern Recognition (ICPR), pages 3605–3611, 2021.
  60. A review of recurrent neural networks: Lstm cells and network architectures. Neural Computation, 31:1235–1270, 2019.
  61. Brian C. Hall. An elementary introduction to groups and representations, 2000.
  62. Silvére Bonnabel. Stochastic gradient descent on riemannian manifolds. IEEE Transactions on Automatic Control, 58:2217–2229, 2011.
  63. Riemannian adaptive optimization methods. ArXiv, abs/1810.00760, 2018.
  64. Mario Lezcano-Casado. Trivializations for Gradient-Based Optimization on Manifolds. Curran Associates Inc., Red Hook, NY, USA, 2019.
  65. M. Hunacek. Lie groups: An introduction through linear groups, by wulf rossmann. pp. 265. £29.95 (pbk). 2006. isbn 0 19 920251 6 (oxford university press). The Mathematical Gazette, 92(524):380–382, 2008.
  66. Mario Lezcano Casado. Adaptive and momentum methods on manifolds through trivializations. ArXiv, abs/2010.04617, 2020.
  67. A. M. Davie. Differential equations driven by rough paths: an approach via discrete approximation, 2007.
  68. A primer on the signature method in machine learning, 2016.
  69. The signature of a rough path: Uniqueness. Advances in Mathematics, 293:720–737, 2016.
  70. Kuo-Tsai Chen. Integration of paths–a faithful representation of paths by noncommutative formal power series. Transactions of the American Mathematical Society, 89(2):395–407, 1958.
  71. Uniqueness for the signature of a path of bounded variation and the reduced path group. Annals of Mathematics, 171(1):109–167, 2010.
  72. Stationary wavelet transform based ecg signal denoising method. ISA Transactions, 114:251–262, 2021.
  73. An empirical study of the behavior of classifiers on imbalanced and overlapped data sets. In Luis Rueda, Domingo Mery, and Josef Kittler, editors, Progress in Pattern Recognition, Image Analysis and Applications, pages 397–406, Berlin, Heidelberg, 2007. Springer Berlin Heidelberg.
  74. Focal loss for dense object detection. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(2):318–327, February 2020.
  75. Smote: Synthetic minority over-sampling technique. J. Artif. Int. Res., 16(1):321–357, June 2002.
  76. Model Selection and Validation, pages 114–123. Cambridge University Press, Cambridge, 2014.
  77. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning - Volume 37, ICML’15, pages 448–456. JMLR.org, 2015.
  78. Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on International Conference on Machine Learning, ICML’10, page 807–814, Madison, WI, USA, 2010. Omnipress.
  79. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In 2015 IEEE International Conference on Computer Vision (ICCV), pages 1026–1034, Los Alamitos, CA, USA, dec 2015. IEEE Computer Society.
  80. Deep sparse rectifier neural networks. In Geoffrey Gordon, David Dunson, and Miroslav Dudík, editors, Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, volume 15 of Proceedings of Machine Learning Research, pages 315–323, Fort Lauderdale, FL, USA, 11–13 Apr 2011. PMLR.
  81. Fast and accurate deep network learning by exponential linear units (elus), 2016.
  82. Random search for hyper-parameter optimization. Journal of Machine Learning Research, 13:281–305, Feb 2012.
  83. Practical bayesian optimization of machine learning algorithms. In F. Pereira, C.J. Burges, L. Bottou, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems, volume 25, 2012.
  84. Adam: A method for stochastic optimization, 2017.
  85. Attention is all you need. In I. Guyon, U. Von Luxburg, S. Bengio, H. Wallach, R. Fergus, S. Vishwanathan, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 30, 2017.
  86. Show, attend and tell: Neural image caption generation with visual attention, 2016.
  87. Sequential attention: A context-aware alignment function for machine reading. In Proceedings of the 2nd Workshop on Representation Learning for NLP, pages 75–80, Vancouver, Canada, August 2017. Association for Computational Linguistics.
  88. Attend and diagnose: Clinical time series analysis using attention models, 2017.
  89. Unsupervised representation learning with deep convolutional generative adversarial networks, 2016.
  90. Data augmentation using generative adversarial networks for robust speech recognition. Speech Commun., 114(C):1–9, nov 2019.
  91. Large scale GAN training for high fidelity natural image synthesis. In 7th International Conference on Learning Representations, ICLR 2019, New Orleans, LA, USA, May 6-9, 2019. OpenReview.net, 2019.
  92. Li Yang and Abdallah Shami. On hyperparameter optimization of machine learning algorithms: Theory and practice. Neurocomputing, 415:295–316, 2020.
  93. Hyper-parameter optimization: A review of algorithms and applications, 2020.
  94. Algorithms for hyper-parameter optimization. In NIPS, 2011.
  95. The relationship between precision-recall and roc curves. In Proceedings of the 23rd International Conference on Machine Learning, ICML ’06, pages 233–240, New York, NY, USA, 2006. Association for Computing Machinery.
  96. Machine learning and deep learning. Electron Markets, 31:685–695, 2021.
  97. Overfitting in neural nets: Backpropagation, conjugate gradient, and early stopping. In T. Leen, T. Dietterich, and V. Tresp, editors, Advances in Neural Information Processing Systems, volume 13. MIT Press, 2000.
  98. Regularization and Stability, pages 137–149. Cambridge University Press, 2014.
  99. David H. Wolpert. The Lack of A Priori Distinctions Between Learning Algorithms. Neural Computation, 8(7):1341–1390, 10 1996.
  100. No free lunch theorems for optimization. IEEE Transactions on Evolutionary Computation, 1(1):67–82, 1997.
  101. The Bias-Complexity Trade-off, pages 36–42. Cambridge University Press, 2014.
  102. Robert Tibshirani. Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society. Series B (Methodological), 58(1):267–288, 1996.
  103. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics, 42(1):80–86, 2000.
  104. A survey of cross-validation procedures for model selection. Statistics Surveys, 4(none):40–79, 2010.
  105. Recent advances in features extraction and description algorithms: A comprehensive survey, 2017.
  106. A survey on image data augmentation for deep learning. Journal of Big Data, 6(1):60, 07 2019.
  107. Gradient-based hyperparameter optimization through reversible learning, 2015.
  108. An introduction to variable and feature selection. Journal of Machine Learning Research, 3:1157–1182, 2003.
  109. ZM Hira and DF Gillies. A review of feature selection and feature extraction methods applied on microarray data. Advances in Bioinformatics, 2015, 2015. Epub 2015 Jun 11.
  110. Interpretable time series classification using linear models and multi-resolution multi-domain symbolic representations. Data Mining and Knowledge Discovery, 33(4):1183–1222, 2019.

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.