Dice Question Streamline Icon: https://streamlinehq.com

Lower bounds and complexity dependence on locality and sparsity

Establish lower bounds for Hamiltonian learning from real-time evolution that clarify how the complexity of learning all parameters scales with the locality of the Hamiltonian and with the effective sparsity; in particular, ascertain whether dependence on effective sparsity is necessary and determine the optimal dependence.

Information Square Streamline Icon: https://streamlinehq.com

Background

A lower bound of (1/ε) log(1/δ) is known for estimating a single parameter, but the overall complexity for learning all parameters remains unclear.

Understanding whether and how the complexity must intrinsically depend on locality and effective sparsity is important for characterizing fundamental limits of Hamiltonian learning algorithms.

References

Can one prove lower bounds on Hamiltonian learning? A lower bound of \frac{1}{\eps}\log\frac{1}{\delta} is known for estimating one parameter. It is not known how the complexity of learning all parameters scales with underlying locality. Is a dependence on effective sparsity \sparse necessary? What is the optimal dependence?

Structure learning of Hamiltonians from real-time evolution (2405.00082 - Bakshi et al., 30 Apr 2024) in Discussion (Future directions), Item 1