Cross-domain structural pattern
New task learning overwrites previous task representations (catastrophic forgetting). Replay of old examples maintains previous capabilities. Sparse replay selects critical examples minimizing memory requirements. Gradient alignment between tasks reduces interference. Trade-off: memory efficiency versus retention quality versus computational overhead.
view paper→