Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 147 tok/s
Gemini 2.5 Pro 52 tok/s Pro
GPT-5 Medium 27 tok/s Pro
GPT-5 High 30 tok/s Pro
GPT-4o 96 tok/s Pro
Kimi K2 188 tok/s Pro
GPT OSS 120B 398 tok/s Pro
Claude Sonnet 4.5 36 tok/s Pro
2000 character limit reached

"It is okay to be uncommon": Quantizing Sound Event Detection Networks on Hardware Accelerators with Uncommon Sub-Byte Support (2404.04386v1)

Published 5 Apr 2024 in cs.SD and eess.AS

Abstract: If our noise-canceling headphones can understand our audio environments, they can then inform us of important sound events, tune equalization based on the types of content we listen to, and dynamically adjust noise cancellation parameters based on audio scenes to further reduce distraction. However, running multiple audio understanding models on headphones with a limited energy budget and on-chip memory remains a challenging task. In this work, we identify a new class of neural network accelerators (e.g., NE16 on GAP9) that allows network weights to be quantized to different common (e.g., 8 bits) and uncommon bit-widths (e.g., 3 bits). We then applied a differentiable neural architecture search to search over the optimal bit-widths of a network on two different sound event detection tasks with potentially different requirements on quantization and prediction granularity (i.e., classification vs. embeddings for few-shot learning). We further evaluated our quantized models on actual hardware, showing that we reduce memory usage, inference latency, and energy consumption by an average of 62%, 46%, and 61% respectively compared to 8-bit models while maintaining floating point performance. Our work sheds light on the benefits of such accelerators on sound event detection tasks when combined with an appropriate search method.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (22)
  1. “CMSIS-NN: Efficient Neural Network Kernels for Arm Cortex-M CPUs,” arXiv preprint arXiv:1801.06601, 2018.
  2. “DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients,” CoRR, vol. abs/1606.06160, 2016.
  3. “Neural Network Distillation on IoT Platforms for Sound Event Detection,” in Interspeech 2019. Sep 2019, p. 3609–3613, ISCA.
  4. “Sound event detection with binary neural networks on tightly power-constrained IoT devices,” in 2020 ACM/IEEE International Symposium on Low Power Electronics and Design, ISLPED 2020; Virtual, Online; United States; 10 August 2020 through 12 August 2020. ACM, 2020.
  5. “UDC: Unified DNAS for compressible TinyML models,” arXiv preprint arXiv:2201.05842, 2022.
  6. “Mixed Precision DNNs: All you need is a good parametrization,” in International Conference on Learning Representations, 2019.
  7. “Syntiant NDP200,” https://www.syntiant.com/ndp200.
  8. “Reconfigurable Binary Engine,” https://github.com/pulp-platform/rbe/.
  9. “XNOR Neural Engine: a Hardware Accelerator IP for 21.6 fJ/op Binary Neural Network Inference,” IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems, vol. 37, no. 11, pp. 2940–2951, Nov 2018, arXiv:1807.03010 [cs].
  10. “Neural Engine 16-channels,” https://github.com/pulp-platform/ne16/, Jul 2023.
  11. GreenWaves Technologies, “GAP9 processor,” https://greenwaves-technologies.com/gap9_processor/.
  12. “Fracbits: Mixed precision quantization via fractional bit-widths,” in Proceedings of the AAAI Conference on Artificial Intelligence, 2021, vol. 35, pp. 10612–10620.
  13. “TinyLSTMs: Efficient Neural Speech Enhancement for Hearing Aids,” arXiv:2005.11138, May 2020.
  14. “Sound Event Detection Via Dilated Convolutional Recurrent Neural Networks,” 2020, IEEE International Conference on Acoustics, Speech and Signal Processing.
  15. “HiSSNet: Sound Event Detection and Speaker Identification via Hierarchical Prototypical Networks for Low-Resource Headphones,” in IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, 2023, pp. 1–5.
  16. “Mobilenetv2: Inverted residuals and linear bottlenecks,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4510–4520.
  17. “BBC Sound Effects,” https://sound-effects.bbcrewind.co.uk/.
  18. “The voice bank corpus: Design, collection and data analysis of a large regional accent speech database,” in 2013 International Conference Oriental COCOSDA held jointly with 2013 Conference on Asian Spoken Language Research and Evaluation (O-COCOSDA/CASLRE), 2013, pp. 1–4.
  19. “PACT: Parameterized clipping activation for quantized neural networks,” arXiv preprint arXiv:1805.06085, 2018.
  20. “GreenWaves nntool,” https://github.com/GreenWaves-Technologies/gap_sdk.
  21. “Saleae Logic Pro 8,” https://usd.saleae.com/products/saleae-logic-pro-8.
  22. “Saleae Logic 2,” https://www.saleae.com/downloads/.

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 1 tweet and received 0 likes.

Upgrade to Pro to view all of the tweets about this paper:

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube