Self-Labeling in Multivariate Causality and Quantification for Adaptive Machine Learning (2404.05809v1)
Abstract: Adaptive ML aims to allow ML models to adapt to ever-changing environments with potential concept drift after model deployment. Traditionally, adaptive ML requires a new dataset to be manually labeled to tailor deployed models to altered data distributions. Recently, an interactive causality based self-labeling method was proposed to autonomously associate causally related data streams for domain adaptation, showing promising results compared to traditional feature similarity-based semi-supervised learning. Several unanswered research questions remain, including self-labeling's compatibility with multivariate causality and the quantitative analysis of the auxiliary models used in the self-labeling. The auxiliary models, the interaction time model (ITM) and the effect state detector (ESD), are vital to the success of self-labeling. This paper further develops the self-labeling framework and its theoretical foundations to address these research questions. A framework for the application of self-labeling to multivariate causal graphs is proposed using four basic causal relationships, and the impact of non-ideal ITM and ESD performance is analyzed. A simulated experiment is conducted based on a multivariate causal graph, validating the proposed theory.
- Y. Roh, G. Heo, and S. E. Whang, “A survey on data collection for machine learning: A big data - ai integration perspective,” IEEE Transactions on Knowledge and Data Engineering, vol. 33, no. 4, pp. 1328–1347, 2021.
- T. Fredriksson, D. I. Mattos, J. Bosch, and H. H. Olsson, “Data labeling: An empirical investigation into industrial challenges and mitigation strategies,” in Product-Focused Software Process Improvement, M. Morisio, M. Torchiano, and A. Jedlitschka, Eds. Cham: Springer International Publishing, 2020, pp. 202–216.
- J. Lu, A. Liu, F. Dong, F. Gu, J. Gama, and G. Zhang, “Learning under concept drift: A review,” IEEE Transactions on Knowledge and Data Engineering, vol. 31, no. 12, pp. 2346–2363, 2019.
- H. Yan, Y. Guo, and C. Yang, “Augmented self-labeling for source-free unsupervised domain adaptation,” in NeurIPS 2021 Workshop on Distribution Shifts: Connecting Methods and Applications, 2021.
- P. Zhou, C. Xiong, X. Yuan, and S. C. H. Hoi, “A theory-driven self-labeling refinement method for contrastive representation learning,” in Advances in Neural Information Processing Systems, vol. 34, 2021, pp. 6183–6197.
- M. Grzenda, H. M. Gomes, and A. Bifet, “Performance measures for evolving predictions under delayed labelling classification,” in 2020 International Joint Conference on Neural Networks (IJCNN), 2020, pp. 1–8.
- H. M. Gomes, M. Grzenda, R. Mello, J. Read, M. H. Le Nguyen, and A. Bifet, “A survey on semi-supervised learning for delayed partially labelled data streams,” ACM Comput. Surv., feb 2022.
- A. J. Ratner, C. M. De Sa, S. Wu, D. Selsam, and C. Ré, “Data programming: Creating large training sets, quickly,” in Advances in Neural Information Processing Systems, vol. 29, 2016.
- R. Stewart and S. Ermon, “Label-free supervision of neural networks with physics and domain knowledge,” in Thirty-First AAAI Conference on Artificial Intelligence, 2017.
- Y. Ren, A. H. Yen, and G.-P. Li, “A self-labeling method for adaptive machine learning by interactive causality,” IEEE Transactions on Artificial Intelligence, 2023.
- S. Sukhbaatar, J. Bruna, M. Paluri, L. Bourdev, and R. Fergus, “Training convolutional networks with noisy labels,” in 3rd International Conference on Learning Representations, Jan. 2015.
- E. García-Martín, C. F. Rodrigues, G. Riley, and H. Grahn, “Estimation of energy consumption in machine learning,” Journal of Parallel and Distributed Computing, vol. 134, pp. 75–88, 2019.
- B. Settles, M. Craven, and L. Friedland, “Active learning with real annotation costs,” in Proceedings of the NIPS workshop on cost-sensitive learning, vol. 1. Vancouver, CA:, 2008.
- P. Oza, V. A. Sindagi, V. V. Sharmini, and V. M. Patel, “Unsupervised domain adaptation of object detectors: A survey,” IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 1–24, 2023.
- J. Huang, D. Guan, A. Xiao, S. Lu, and L. Shao, “Category contrast for unsupervised domain adaptation in visual tasks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2022, pp. 1203–1214.
- Y.-C. Yu and H.-T. Lin, “Semi-supervised domain adaptation with source label adaptation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2023, pp. 24 100–24 109.
- P. Covington, J. Adams, and E. Sargin, “Deep neural networks for youtube recommendations,” in Proceedings of the 10th ACM conference on recommender systems, 2016, pp. 191–198.
- T. Xu, G. Goossen, H. K. Cevahir, S. Khodeir, Y. Jin, F. Li, S. Shan, S. Patel, D. Freeman, and P. Pearce, “Deep entity classification: Abusive account detection for online social networks,” in 30th USENIX Security Symposium (USENIX Security 21). USENIX Association, Aug. 2021, pp. 4097–4114. [Online]. Available: https://www.usenix.org/conference/usenixsecurity21/presentation/xu-teng
- M. Ragab, E. Eldele, Z. Chen, M. Wu, C.-K. Kwoh, and X. Li, “Self-supervised autoregressive domain adaptation for time series data,” IEEE Transactions on Neural Networks and Learning Systems, vol. 35, no. 1, pp. 1341–1351, 2024.
- H. He, O. Queen, T. Koker, C. Cuevas, T. Tsiligkaridis, and M. Zitnik, “Domain adaptation for time series under feature and label shifts,” in Proceedings of the 40th International Conference on Machine Learning, ser. Proceedings of Machine Learning Research, A. Krause, E. Brunskill, K. Cho, B. Engelhardt, S. Sabato, and J. Scarlett, Eds., vol. 202. PMLR, 23–29 Jul 2023, pp. 12 746–12 774.
- R. Cai, J. Chen, Z. Li, W. Chen, K. Zhang, J. Ye, Z. Li, X. Yang, and Z. Zhang, “Time series domain adaptation via sparse associative structure alignment,” Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, no. 8, pp. 6859–6867, May 2021.
- H. F. Gollob and C. S. Reichardt, “Taking account of time lags in causal models,” Child Development, vol. 58, no. 1, pp. 80–92, 1987.
- S. Sawesi, M. Rashrash, and O. Dammann, “The representation of causality and causation with ontologies: A systematic literature review,” Online Journal of Public Health Informatics, vol. 14, no. 1, Sep. 2022.
- C. Zhang, S. Bauer, P. Bennett, J. Gao, W. Gong, A. Hilmkil, J. Jennings, C. Ma, T. Minka, N. Pawlowski, and J. Vaughan, “Understanding causality with large language models: Feasibility and opportunities,” 2023.
- H. Huang, “Causal relationship over knowledge graphs,” in Proceedings of the 31st ACM International Conference on Information & Knowledge Management, ser. CIKM ’22. New York, NY, USA: Association for Computing Machinery, 2022, p. 5116–5119.
- C. Gan, J. Schwartz, S. Alter, D. Mrowca, M. Schrimpf, J. Traer, J. D. Freitas, J. Kubilius, A. Bhandwaldar, N. Haber, M. Sano, K. Kim, E. Wang, M. Lingelbach, A. Curtis, K. Feigelis, D. M. Bear, D. Gutfreund, D. Cox, A. Torralba, J. J. DiCarlo, J. B. Tenenbaum, J. H. McDermott, and D. L. K. Yamins, “Threedworld: A platform for interactive multi-modal physical simulation,” 2021.
- T. Chen and C. Guestrin, “Xgboost: A scalable tree boosting system,” in Proceedings of the 22nd acm sigkdd international conference on knowledge discovery and data mining, 2016, pp. 785–794.
- D.-H. Lee et al., “Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks,” in Workshop on challenges in representation learning, ICML, vol. 3, no. 2, 2013, p. 896.
- D. Berthelot, N. Carlini, I. Goodfellow, N. Papernot, A. Oliver, and C. A. Raffel, “Mixmatch: A holistic approach to semi-supervised learning,” in Advances in Neural Information Processing Systems, vol. 32, 2019.
- B. Zhang, Y. Wang, W. Hou, H. WU, J. Wang, M. Okumura, and T. Shinozaki, “Flexmatch: Boosting semi-supervised learning with curriculum pseudo labeling,” in Advances in Neural Information Processing Systems, vol. 34, 2021, pp. 18 408–18 419.
- Y. Wang, H. Chen, Q. Heng, W. Hou, Y. Fan, Z. Wu, J. Wang, M. Savvides, T. Shinozaki, B. Raj, B. Schiele, and X. Xie, “Freematch: Self-adaptive thresholding for semi-supervised learning,” in The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net, 2023. [Online]. Available: https://openreview.net/pdf?id=PDrUPTXJI_A
- H. Chen, R. Tao, Y. Fan, Y. Wang, J. Wang, B. Schiele, X. Xie, B. Raj, and M. Savvides, “Softmatch: Addressing the quantity-quality tradeoff in semi-supervised learning,” in The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023. OpenReview.net, 2023. [Online]. Available: https://openreview.net/pdf?id=ymt1zQXBDiF
- J. Choquette, W. Gandhi, O. Giroux, N. Stam, and R. Krashinsky, “Nvidia a100 tensor core gpu: Performance and innovation,” IEEE Micro, vol. 41, no. 2, pp. 29–35, 2021.
- D. Rolnick, A. Veit, S. J. Belongie, and N. Shavit, “Deep learning is robust to massive label noise,” CoRR, vol. abs/1705.10694, 2017. [Online]. Available: http://arxiv.org/abs/1705.10694
- B. Frenay and M. Verleysen, “Classification in the presence of label noise: A survey,” IEEE Transactions on Neural Networks and Learning Systems, vol. 25, no. 5, pp. 845–869, 2014.
- M. Li, M. Soltanolkotabi, and S. Oymak, “Gradient descent with early stopping is provably robust to label noise for overparameterized neural networks,” in Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, ser. Proceedings of Machine Learning Research, S. Chiappa and R. Calandra, Eds., vol. 108. PMLR, 26–28 Aug 2020, pp. 4313–4324.
- V. Khetan, R. Ramnani, M. Anand, S. Sengupta, and A. E. Fano, “Causal bert: Language models for causality detection between events expressed in text,” in Intelligent Computing, K. Arai, Ed. Cham: Springer International Publishing, 2022, pp. 965–980.
- Z. LYU, Z. Jin, R. Mihalcea, M. Sachan, and B. Schölkopf, “Can large language models distinguish cause from effect?” in UAI 2022 Workshop on Causal Representation Learning, 2022.
- E. Kıcıman, R. Ness, A. Sharma, and C. Tan, “Causal reasoning and large language models: Opening a new frontier for causality,” 2023.
- E. Yiu, E. Kosoy, and A. Gopnik, “Transmission versus truth, imitation versus innovation: What children can do that large language and language-and-vision models cannot (yet),” Perspectives on Psychological Science, vol. 0, no. 0, 2023.