On the Downstream Performance of Compressed Word Embeddings.

Adv Neural Inf Process Syst

Department of Computer Science, Stanford University, Stanford, CA 94305.

Published: December 2019

Compressing word embeddings is important for deploying NLP models in memory-constrained settings. However, understanding what makes compressed embeddings perform well on downstream tasks is challenging-existing measures of compression quality often fail to distinguish between embeddings that perform well and those that do not. We thus propose the as a new measure. We relate the eigenspace overlap score to downstream performance by developing generalization bounds for the compressed embeddings in terms of this score, in the context of linear and logistic regression. We then show that we can lower bound the eigenspace overlap score for a simple uniform quantization compression method, helping to explain the strong empirical performance of this method. Finally, we show that by using the eigenspace overlap score as a selection criterion between embeddings drawn from a representative set we compressed, we can efficiently identify the better performing embedding with up to 2× lower selection error rates than the next best measure of compression quality, and avoid the cost of training a model for each task of interest.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC6935262PMC

Publication Analysis

Top Keywords

eigenspace overlap
12
overlap score
12
downstream performance
8
word embeddings
8
compressed embeddings
8
embeddings perform
8
perform well
8
compression quality
8
embeddings
6
compressed
4

Similar Publications

On the Downstream Performance of Compressed Word Embeddings.

Adv Neural Inf Process Syst

December 2019

Department of Computer Science, Stanford University, Stanford, CA 94305.

Compressing word embeddings is important for deploying NLP models in memory-constrained settings. However, understanding what makes compressed embeddings perform well on downstream tasks is challenging-existing measures of compression quality often fail to distinguish between embeddings that perform well and those that do not. We thus propose the as a new measure.

View Article and Find Full Text PDF

Laplacian mixture modeling for network analysis and unsupervised learning on graphs.

PLoS One

March 2019

Research Division, Nanobio.Md, San Francisco, CA, United States of America.

Laplacian mixture models identify overlapping regions of influence in unlabeled graph and network data in a scalable and computationally efficient way, yielding useful low-dimensional representations. By combining Laplacian eigenspace and finite mixture modeling methods, they provide probabilistic or fuzzy dimensionality reductions or domain decompositions for a variety of input data types, including mixture distributions, feature vectors, and graphs or networks. Provable optimal recovery using the algorithm is analytically shown for a nontrivial class of cluster graphs.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!