Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Continual Learning with Guarantees via Weight Interval Constraints (2206.07996v1)

Published 16 Jun 2022 in cs.LG

Abstract: We introduce a new training paradigm that enforces interval constraints on neural network parameter space to control forgetting. Contemporary Continual Learning (CL) methods focus on training neural networks efficiently from a stream of data, while reducing the negative impact of catastrophic forgetting, yet they do not provide any firm guarantees that network performance will not deteriorate uncontrollably over time. In this work, we show how to put bounds on forgetting by reformulating continual learning of a model as a continual contraction of its parameter space. To that end, we propose Hyperrectangle Training, a new training methodology where each task is represented by a hyperrectangle in the parameter space, fully contained in the hyperrectangles of the previous tasks. This formulation reduces the NP-hard CL problem back to polynomial time while providing full resilience against forgetting. We validate our claim by developing InterContiNet (Interval Continual Learning) algorithm which leverages interval arithmetic to effectively model parameter regions as hyperrectangles. Through experimental results, we show that our approach performs well in a continual learning setup without storing data from previous tasks.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Maciej Wołczyk (27 papers)
  2. Karol J. Piczak (3 papers)
  3. Bartosz Wójcik (15 papers)
  4. Łukasz Pustelnik (2 papers)
  5. Paweł Morawiecki (8 papers)
  6. Jacek Tabor (106 papers)
  7. Tomasz Trzciński (116 papers)
  8. Przemysław Spurek (74 papers)
Citations (8)

Summary

We haven't generated a summary for this paper yet.