We present a method for storing multiple models within a single set of
parameters. Models can coexist in superposition and still be retrieved
individually. In experiments with neural networks, we show that a surprisingly
large number of models can be effectively stored within a single parameter
instance. Furthermore, each of these models can undergo thousands of training
steps without significantly interfering with other models within the
superposition. This approach may be viewed as the online complement of
compression: rather than reducing the size of a network after training, we make
use of the unrealized capacity of a network during training.
Beschreibung
[1902.05522] Superposition of many models into one
%0 Generic
%1 cheung-2019
%A Cheung, Brian
%A Terekhov, Alex
%A Chen, Yubei
%A Agrawal, Pulkit
%A Olshausen, Bruno
%D 2019
%K artificial_intelligence machine_learning neural_networks
%T Superposition of many models into one
%U http://arxiv.org/abs/1902.05522
%X We present a method for storing multiple models within a single set of
parameters. Models can coexist in superposition and still be retrieved
individually. In experiments with neural networks, we show that a surprisingly
large number of models can be effectively stored within a single parameter
instance. Furthermore, each of these models can undergo thousands of training
steps without significantly interfering with other models within the
superposition. This approach may be viewed as the online complement of
compression: rather than reducing the size of a network after training, we make
use of the unrealized capacity of a network during training.
@misc{cheung-2019,
abstract = {We present a method for storing multiple models within a single set of
parameters. Models can coexist in superposition and still be retrieved
individually. In experiments with neural networks, we show that a surprisingly
large number of models can be effectively stored within a single parameter
instance. Furthermore, each of these models can undergo thousands of training
steps without significantly interfering with other models within the
superposition. This approach may be viewed as the online complement of
compression: rather than reducing the size of a network after training, we make
use of the unrealized capacity of a network during training.},
added-at = {2019-03-10T10:38:27.000+0100},
author = {Cheung, Brian and Terekhov, Alex and Chen, Yubei and Agrawal, Pulkit and Olshausen, Bruno},
biburl = {https://www.bibsonomy.org/bibtex/297e51df6c23dfca736cf2ba4c8daf451/jpvaldes},
description = {[1902.05522] Superposition of many models into one},
interhash = {4b8f3338345cf9fa42e4a4e964be4424},
intrahash = {97e51df6c23dfca736cf2ba4c8daf451},
keywords = {artificial_intelligence machine_learning neural_networks},
note = {cite arxiv:1902.05522},
timestamp = {2019-03-10T10:38:27.000+0100},
title = {Superposition of many models into one},
url = {http://arxiv.org/abs/1902.05522},
year = 2019
}