Distributed Weight Consolidation: A Brain Segmentation Case Study

التفاصيل البيبلوغرافية
العنوان: Distributed Weight Consolidation: A Brain Segmentation Case Study
المؤلفون: McClure, Patrick, Zheng, Charles Y., Kaczmarzyk, Jakub R., Lee, John A., Ghosh, Satrajit S., Nielson, Dylan, Bandettini, Peter, Pereira, Francisco
سنة النشر: 2018
المجموعة: Computer Science
Statistics
مصطلحات موضوعية: Computer Science - Machine Learning, Computer Science - Computer Vision and Pattern Recognition, Statistics - Machine Learning
الوصف: Collecting the large datasets needed to train deep neural networks can be very difficult, particularly for the many applications for which sharing and pooling data is complicated by practical, ethical, or legal concerns. However, it may be the case that derivative datasets or predictive models developed within individual sites can be shared and combined with fewer restrictions. Training on distributed data and combining the resulting networks is often viewed as continual learning, but these methods require networks to be trained sequentially. In this paper, we introduce distributed weight consolidation (DWC), a continual learning method to consolidate the weights of separate neural networks, each trained on an independent dataset. We evaluated DWC with a brain segmentation case study, where we consolidated dilated convolutional neural networks trained on independent structural magnetic resonance imaging (sMRI) datasets from different sites. We found that DWC led to increased performance on test sets from the different sites, while maintaining generalization performance for a very large and completely independent multi-site dataset, compared to an ensemble baseline.
Comment: Published in NeurIPS 2018
نوع الوثيقة: Working Paper
URL الوصول: http://arxiv.org/abs/1805.10863
رقم الأكسشن: edsarx.1805.10863
قاعدة البيانات: arXiv