Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 167 tok/s
Gemini 2.5 Pro 48 tok/s Pro
GPT-5 Medium 33 tok/s Pro
GPT-5 High 40 tok/s Pro
GPT-4o 92 tok/s Pro
Kimi K2 193 tok/s Pro
GPT OSS 120B 425 tok/s Pro
Claude Sonnet 4.5 37 tok/s Pro
2000 character limit reached

Differentiating Through Linear Solvers (2404.17039v2)

Published 25 Apr 2024 in cs.MS, cs.NA, and math.NA

Abstract: Computer programs containing calls to linear solvers are a known challenge for automatic differentiation. Previous publications advise against differentiating through the low-level solver implementation, and instead advocate for high-level approaches that express the derivative in terms of a modified linear system that can be solved with a separate solver call. Despite this ubiquitous advice, we are not aware of prior work comparing the accuracy of both approaches. With this article we thus empirically study a simple question: What happens if we ignore common wisdom, and differentiate through linear solvers?

Definition Search Book Streamline Icon: https://streamlinehq.com
References (9)
  1. An updated set of basic linear algebra subprograms (blas). ACM Transactions on Mathematical Software, 28(2):135–151, 2002.
  2. Christianson, B. Reverse accumulation and attractive fixed points. Optimization Methods and Software, 3(4):311–326, 1994.
  3. Christianson, B. Differentiating through conjugate gradient. Optimization Methods and Software, 33(4-6):988–994, 2018.
  4. The university of florida sparse matrix collection. ACM Trans. Math. Softw., 38(1), dec 2011. ISSN 0098-3500. doi: 10.1145/2049662.2049663. URL https://doi.org/10.1145/2049662.2049663.
  5. Differentiating the method of conjugate gradients. SIAM Journal on Matrix Analysis and Applications, 35(1):110–126, 2014.
  6. Evaluating derivatives: principles and techniques of algorithmic differentiation. SIAM, 2008.
  7. Derivative convergence for iterative equation solvers. Optimization methods and software, 2(3-4):321–355, 1993.
  8. The tapenade automatic differentiation tool: Principles, model, and specification. ACM Transactions on Mathematical Software (TOMS), 39(3):1–43, 2013.
  9. Saad, Y. SPARSKIT: a basic tool kit for sparse matrix computations, 1994.
Citations (2)

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com
Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 2 tweets and received 0 likes.

Upgrade to Pro to view all of the tweets about this paper: