Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 180 tok/s
Gemini 2.5 Pro 55 tok/s Pro
GPT-5 Medium 34 tok/s Pro
GPT-5 High 37 tok/s Pro
GPT-4o 95 tok/s Pro
Kimi K2 205 tok/s Pro
GPT OSS 120B 433 tok/s Pro
Claude Sonnet 4.5 38 tok/s Pro
2000 character limit reached

Limitations of Amplitude Encoding on Quantum Classification (2503.01545v1)

Published 3 Mar 2025 in quant-ph

Abstract: It remains unclear whether quantum machine learning (QML) has real advantages when dealing with practical and meaningful tasks. Encoding classical data into quantum states is one of the key steps in QML. Amplitude encoding has been widely used owing to its remarkable efficiency in encoding a number of $2{n}$ classical data into $n$ qubits simultaneously. However, the theoretical impact of amplitude encoding on QML has not been thoroughly investigated. In this work we prove that under some broad and typical data assumptions, the average of encoded quantum states via amplitude encoding tends to concentrate towards a specific state. This concentration phenomenon severely constrains the capability of quantum classifiers as it leads to a loss barrier phenomenon, namely, the loss function has a lower bound that cannot be improved by any optimization algorithm. In addition, via numerical simulations, we reveal a counterintuitive phenomenon of amplitude encoding: as the amount of training data increases, the training error may increase rather than decrease, leading to reduced decrease in prediction accuracy on new data. Our results highlight the limitations of amplitude encoding in QML and indicate that more efforts should be devoted to finding more efficient encoding strategies to unlock the full potential of QML.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.