pith. machine review for the scientific record. sign in

arxiv: 1606.09282 · v3 · submitted 2016-06-29 · 💻 cs.CV · cs.LG· stat.ML

Recognition: unknown

Learning without Forgetting

Authors on Pith no claims yet
classification 💻 cs.CV cs.LGstat.ML
keywords datacapabilitieslearningtaskforgettingwithoutfine-tuningmethod
0
0 comments X
read the original abstract

When building a unified vision system or gradually adding new capabilities to a system, the usual assumption is that training data for all tasks is always available. However, as the number of tasks grows, storing and retraining on such data becomes infeasible. A new problem arises where we add new capabilities to a Convolutional Neural Network (CNN), but the training data for its existing capabilities are unavailable. We propose our Learning without Forgetting method, which uses only new task data to train the network while preserving the original capabilities. Our method performs favorably compared to commonly used feature extraction and fine-tuning adaption techniques and performs similarly to multitask learning that uses original task data we assume unavailable. A more surprising observation is that Learning without Forgetting may be able to replace fine-tuning with similar old and new task datasets for improved new task performance.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 3 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Memory-Efficient Looped Transformer: Decoupling Compute from Memory in Looped Language Models

    cs.CL 2026-05 unverdicted novelty 6.0

    MELT decouples reasoning depth from memory in looped LLMs by sharing a single gated KV cache per layer and using two-phase chunk-wise distillation from Ouro, delivering constant memory use while matching or beating st...

  2. Layered Mutability: Continuity and Governance in Persistent Self-Modifying Agents

    cs.AI 2026-04 unverdicted novelty 5.0

    Layered mutability framework claims governance difficulty in persistent self-modifying agents rises with rapid mutation, strong downstream coupling, weak reversibility, and low observability, producing compositional d...

  3. Layered Mutability: Continuity and Governance in Persistent Self-Modifying Agents

    cs.AI 2026-04 unverdicted novelty 5.0

    Persistent self-modifying AI agents exhibit compositional drift from mismatches across five mutability layers, with governance difficulty rising under rapid mutation, strong coupling, weak reversibility, and low obser...