Papers
Topics
Authors
Recent
2000 character limit reached

Model Complexity of Program Phases (2310.03865v1)

Published 5 Oct 2023 in cs.LG, cs.IT, and math.IT

Abstract: In resource limited computing systems, sequence prediction models must operate under tight constraints. Various models are available that cater to prediction under these conditions that in some way focus on reducing the cost of implementation. These resource constrained sequence prediction models, in practice, exhibit a fundamental tradeoff between the cost of implementation and the quality of its predictions. This fundamental tradeoff seems to be largely unexplored for models for different tasks. Here we formulate the necessary theory and an associated empirical procedure to explore this tradeoff space for a particular family of machine learning models such as deep neural networks. We anticipate that the knowledge of the behavior of this tradeoff may be beneficial in understanding the theoretical and practical limits of creation and deployment of models for resource constrained tasks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (28)
  1. The information complexity of learning tasks, their structure and their distance. Information and Inference: A Journal of the IMA, 10(1):51–72.
  2. Phase-aware CPU workload forecasting.
  3. Learning long-term dependencies with gradient descent is difficult. Transactions in Neural Networks, 5(2):157–166.
  4. Perceptron-based prefetch filtering. 2019 ACM/IEEE 46th Annual International Symposium on Computer Architecture (ISCA), pages 1–13.
  5. Evidence-based static branch prediction using machine learning. ACM Trans. Program. Lang. Syst., 19:188–222.
  6. Using machine learning techniques to analyze the performance of concurrent kernel execution on GPUs. Future Gener. Comput. Syst., 113:528–540.
  7. Real time detection of cache-based side-channel attacks using hardware performance counters. Appl. Soft Comput., 49:1162–1174.
  8. Run-time program-specific phase prediction for Python programs. In ManLang ’18.
  9. Clustering by compression. IEEE Transactions on Information Theory, 51:1523–1545.
  10. Block2vec: A deep learning strategy on mining block correlations in storage systems. 2016 45th International Conference on Parallel Processing Workshops (ICPPW), pages 230–239.
  11. Classification of computer hardware and performance prediction using statistical learning and neural networks.
  12. Hochreiter, S. (1998). The vanishing gradient problem during learning recurrent neural nets and problem solutions. Int. J. Uncertain. Fuzziness Knowl.-Based Syst., 6(2):107–116.
  13. Long short-term memory. Neural Computation, 9:1735–1780.
  14. Learning automatic schedulers with projective reparameterization.
  15. Miss Rate Estimation (MRE): a Novel Approach Toward L2 Cache Partitioning Algorithms for Multicore System.
  16. RHMD: Evasion-Resilient Hardware Malware Detectors. 2017 50th Annual IEEE/ACM International Symposium on Microarchitecture (MICRO), pages 315–327.
  17. EnsembleHMD: Accurate Hardware Malware Detectors with Specialized Ensemble Classifiers. IEEE Transactions on Dependable and Secure Computing, 17:620–633.
  18. A new approach to detecting execution phases using performance monitoring counters. In Architecture of Computing Systems - ARCS 2017.
  19. Machine learning based online performance prediction for runtime parallelization and task scheduling. 2009 IEEE International Symposium on Performance Analysis of Systems and Software, pages 89–100.
  20. Assessing the ability of LSTMs to learn syntax-sensitive dependencies. Transactions of the Association for Computational Linguistics, 4:521–535.
  21. Approximating rate-distortion graphs of individual data: Experiments in lossy compression and denoising. IEEE Transactions on Computers, 61:395–407.
  22. Learning internal representations by error propagation.
  23. Regression as classification. 2012 Proceedings of IEEE Southeastcon, pages 1–6.
  24. Program phase duration prediction and its application to fine-grain power management. 2013 IEEE Computer Society Annual Symposium on VLSI (ISVLSI), pages 127–132.
  25. Sequence to sequence learning with neural networks. In Neural Information Processing Systems.
  26. Kolmogorov’s structure functions and model selection. IEEE Transactions on Information Theory, 50:3265–3290.
  27. GPGPU performance and power estimation using machine learning. 2015 IEEE 21st International Symposium on High Performance Computer Architecture (HPCA), pages 564–576.
  28. Google’s neural machine translation system: Bridging the gap between human and machine translation. CoRR, abs/1609.08144.

Summary

We haven't generated a summary for this paper yet.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube