Multi-Task Generalization and Adaptation between Noisy Digit Datasets: An Empirical Study

Cite this publication

​Schneider S, Ecker AS, Macke JH, Bethge M. ​Multi-Task Generalization and Adaptation between Noisy Digit Datasets: An Empirical Study​. ​ Paper presented at:​ Neural Information Processing Systems (NeurIPS), Workshop on Continual Learning; ​2018​ - 2018​; ​​Montreal, Canada. 

Documents & Media

License

GRO License GRO License

Details

Authors
Schneider, Steffen; Ecker, Alexander S. ; Macke, Jakob H.; Bethge, Matthias
Abstract
Transfer learning for adaptation to new tasks is usually performed by either finetuning all model parameters or parameters in the final layers. We show that good target performance can also be achieved on typical domain adaptation tasks by adapting only the normalization statistics and affine transformations of layers throughout the network. We apply this adaptation scheme to supervised domain adaptation on common digit datasets and study robustness properties under perturbation by noise. Our results indicate that (1) adaptation to noise exceeds the difficulty of widely used digit benchmarks in domain adaptation,(2) the similarity of the optimal adaptation parameters for different domains is strongly predictive of generalization performance, and (3) generalization performance is highest with training on a rich environment or high noise levels.
Issue Date
2018
Conference
Neural Information Processing Systems (NeurIPS), Workshop on Continual Learning
Conference Place
Montreal, Canada
Event start
2018
Event end
2018
Language
English

Reference

Citations