Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

DOCTOR: A Multi-Disease Detection Continual Learning Framework Based on Wearable Medical Sensors (2305.05738v5)

Published 9 May 2023 in cs.LG, cs.HC, and eess.SP

Abstract: Modern advances in ML and wearable medical sensors (WMSs) in edge devices have enabled ML-driven disease detection for smart healthcare. Conventional ML-driven methods for disease detection rely on customizing individual models for each disease and its corresponding WMS data. However, such methods lack adaptability to distribution shifts and new task classification classes. In addition, they need to be rearchitected and retrained from scratch for each new disease. Moreover, installing multiple ML models in an edge device consumes excessive memory, drains the battery faster, and complicates the detection process. To address these challenges, we propose DOCTOR, a multi-disease detection continual learning (CL) framework based on WMSs. It employs a multi-headed deep neural network (DNN) and a replay-style CL algorithm. The CL algorithm enables the framework to continually learn new missions where different data distributions, classification classes, and disease detection tasks are introduced sequentially. It counteracts catastrophic forgetting with a data preservation method and a synthetic data generation (SDG) module. The data preservation method preserves the most informative subset of real training data from previous missions for exemplar replay. The SDG module models the probability distribution of the real training data and generates synthetic data for generative replay while retaining data privacy. The multi-headed DNN enables DOCTOR to detect multiple diseases simultaneously based on user WMS data. We demonstrate DOCTOR's efficacy in maintaining high disease classification accuracy with a single DNN model in various CL experiments. In complex scenarios, DOCTOR achieves 1.43 times better average test accuracy, 1.25 times better F1-score, and 0.41 higher backward transfer than the naive fine-tuning framework with a small model size of less than 350KB.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (43)
  1. Viswanatha Reddy Allugunti. 2022. A machine learning model for skin disease classification using convolution neural network. International Journal of Computing, Programming and Database Management 3, 1 (2022), 141–147.
  2. Dissecting FLOPs along input dimensions for greenAI cost estimations. In Machine Learning, Optimization, and Data Science. 86–100.
  3. Energy-based tuning of convolutional neural networks on multi-GPUs. Concurrency and Computation: Practice and Experience 31, 21 (2019).
  4. SMOTE: Synthetic minority over-sampling technique. Journal of Artificial Intelligence Research 16 (Jun. 2002), 321–357.
  5. Continual learning for recurrent neural networks: An empirical evaluation. Neural Networks 143 (2021), 607–627.
  6. A continual learning survey: Defying forgetting in classification tasks. IEEE Transactions on Pattern Analysis and Machine Intelligence 44, 7 (2022), 3366–3385.
  7. Accuracy is not the only metric that matters: Estimating the energy consumption of deep learning models. arXiv: 2304.00897 (2023).
  8. Intelligent model to predict early liver disease using machine learning technique. In Proceedings of the International Conference on Business Analytics for Technology and Security. 1–5.
  9. Generative adversarial networks. Commun. ACM 63, 11 (Oct. 2020), 139–144.
  10. CovidDeep: SARS-CoV-2/COVID-19 test based on wearable medical sensors and efficient neural networks. IEEE Transactions on Consumer Electronics 67, 4 (2021), 244–256.
  11. TUTOR: Training neural networks using decision rules as model priors. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 42, 2 (2023), 483–496.
  12. SCANN: Synthesis of compact and accurate neural networks. IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems 41, 9 (2022), 3012–3025.
  13. MHDeep: Mental health disorder detection system based on wearable sensors and artificial neural networks. ACM Transactions on Embedded Computing Systems 21, 6, Article 81 (Dec. 2022), 22 pages.
  14. REMIND your neural network to prevent catastrophic forgetting. In Proceedings of the European Conference on Computer Vision. 466–483.
  15. Denoising diffusion probabilistic models. In Advances in Neural Information Processing Systems, Vol. 33. 6840–6851.
  16. Diffusion models for medical image analysis: A comprehensive survey. arXiv: 2211.07804 (2022).
  17. Overcoming catastrophic forgetting in neural networks. Proceedings of the National Academy of Sciences 114, 13 (2017), 3521–3526.
  18. Optimal continual learning has perfect memory and is NP-hard. In Proceedings of the 37th International Conference on Machine Learning, Vol. 119. 5327–5337.
  19. Exploring system performance of continual learning for mobile and embedded sensing applications. In Proceedings of the 2021 IEEE/ACM Symposium on Edge Computing. 319–332.
  20. Zhizhong Li and Derek Hoiem. 2018. Learning without forgetting. IEEE Transactions on Pattern Analysis and Machine Intelligence 40, 12 (2018), 2935–2947.
  21. David Lopez-Paz and Marc' Aurelio Ranzato. 2017. Gradient episodic memory for continual learning. In Advances in Neural Information Processing Systems, Vol. 30.
  22. Online continual learning in image classification: An empirical survey. Neurocomputing 469 (2022), 28–51.
  23. Arun Mallya and Svetlana Lazebnik. 2018. PackNet: Adding multiple tasks to a single network by iterative pruning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 7765–7773.
  24. David McCaffary. 2021. Towards continual task learning in artificial neural networks: Current approaches and insights from neuroscience. arXiv: 2112.14146 (2021).
  25. Michael McCloskey and Neal J. Cohen. 1989. Catastrophic interference in connectionist networks: The sequential learning problem. In Psychology of Learning and Motivation. Vol. 24. 109–165.
  26. Breast cancer detection in mammogram: Combining modified CNN and texture feature based approach. Journal of Ambient Intelligence and Humanized Computing (2022), 1–10.
  27. Jorge A. Mendez and Eric Eaton. 2023. How to reuse and compose knowledge for a lifetime of tasks: A survey on continual learning and functional composition. arXiv: 2207.07730 (2023).
  28. Julia H. Miao and Kathleen H. Miao. 2018. Cardiotocographic diagnosis of fetal health based on multiclass morphologic pattern predictions using deep learning classification. International Journal of Advanced Computer Science and Applications 9, 5 (2018).
  29. Continual lifelong learning with neural networks: A review. Neural Networks 113 (2019), 54–71.
  30. Recent advances of continual learning in computer vision: An overview. arXiv: 2109.11369 (2021).
  31. iCaRL: Incremental classifier and representation learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition.
  32. Continual learning with deep generative replay. In Advances in Neural Information Processing Systems, Vol. 30.
  33. SpaceNet: Make free space for continual learning. Neurocomputing 439 (2021), 1–11.
  34. Prerit Terway and Niraj K. Jha. 2023. REPAIRS: Gaussian mixture model-based completion and optimization of partially specified systems. ACM Transactions on Embedded Computing Systems 22, 4, Article 69 (Jul. 2023), 36 pages.
  35. Gido M. van de Ven and Andreas S. Tolias. 2019. Three scenarios for continual learning. arXiv: 1904.07734 (2019).
  36. Online continual learning with contrastive vision transformer. In Proceedings of the European Conference on Computer Vision. 631–650.
  37. Learning to prompt for continual learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 139–149.
  38. Large scale incremental learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition.
  39. Hongxu Yin and Niraj K. Jha. 2017. A health decision support system for disease diagnosis based on wearable medical sensors and machine learning ensembles. IEEE Transactions on Multi-Scale Computing Systems 3, 4 (2017), 228–241.
  40. DiabDeep: Pervasive diabetes diagnosis based on wearable medical sensors and efficient neural networks. IEEE Transactions on Emerging Topics in Computing 9, 3 (2021), 1139–1150.
  41. Lifelong learning with dynamically expandable networks. arXiv: 1708.01547 (2017).
  42. Chang Yue and Niraj K. Jha. 2022. CTRL: Clustering training losses for label error detection. arXiv: 2208.08464 (2022).
  43. Continual learning through synaptic intelligence. In Proceedings of the 34th International Conference on Machine Learning, Vol. 70. 3987–3995.
Citations (6)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets