We extend feed-forward neural networks with a Dirichlet process prior over the weight distribution. This enforces a sharing on the network weights, which can reduce the overall number of parameters drastically. We alternately sample from the posterior of the weights and the posterior of assignments of network connections to the weights. This results in a weight sharing that is adopted to the given data. In order to make the procedure feasible, we present several techniques to reduce the computational burden. Experiments show that our approach mostly outperforms models with random weight sharing. Our model is capable of reducing the memory footprint substantially while maintaining a good performance compared to neural networks without weight sharing.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TPAMI.2018.2884905DOI Listing

Publication Analysis

Top Keywords

weight sharing
16
neural networks
12
networks weight
8
weight
5
sharing
5
bayesian neural
4
sharing dirichlet
4
dirichlet processes
4
processes extend
4
extend feed-forward
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!