Matrix factorization is an important mathematical problem encountered in the context of dictionary learning, recommendation systems, and machine learning. We introduce a decimation scheme that maps it to neural network models of associative memory and provide a detailed theoretical analysis of its performance, showing that decimation is able to factorize extensive-rank matrices and to denoise them efficiently. In the case of binary prior on the signal components, we introduce a decimation algorithm based on a ground-state search of the neural network, which shows performances that match the theoretical prediction.
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1103/PhysRevE.107.064308 | DOI Listing |
Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!