Deep neural networks (DNNs) usually contain massive parameters, but there is redundancy such that it is guessed that they could be trained in low-dimensional subspaces. In this paper, we propose a Dynamic Linear Dimensionality Reduction (DLDR) based on the low-dimensional properties of the training trajectory. The reduction method is efficient, supported by comprehensive experiments: optimizing DNNs in 40-dimensional spaces can achieve comparable performance as regular training over thousands or even millions of parameters. Since there are only a few variables to optimize, we develop an efficient quasi-Newton-based algorithm, obtain robustness to label noise, and improve the performance of well-trained models, which are three follow-up experiments that can show the advantages of finding such low-dimensional subspaces. The code is released (Pytorch: https://github.com/nblt/DLDR and Mindspore: https://gitee.com/mindspore/docs/tree/r1.6/docs/sample_code/dimension_reduce_training).
Download full-text PDF |
Source |
---|---|
http://dx.doi.org/10.1109/TPAMI.2022.3178101 | DOI Listing |
J Chem Theory Comput
December 2024
Department of Chemistry, Faculty of Science, Hokkaido University, Sapporo 060-0810, Japan.
In recent years, automated reaction path search methods have established the concept of a reaction route network. The Reaction Space Projector (ReSPer) visualizes the potential energy hypersurface into a lower-dimensional subspace using principal coordinates. The main time-consuming process in ReSPer is calculating the structural distance matrix, making it impractical for complex organic reaction route networks.
View Article and Find Full Text PDFiScience
December 2024
Brain Institute, Federal University of Rio Grande do Norte, Natal, RN 59078, Brazil.
bioRxiv
November 2024
Department of Systems and Computational Biology, Albert Einstein College of Medicine, Bronx, NY 10461, USA.
Phys Rev E
October 2024
Department of Physics, University of Houston, Houston, Texas 77204, USA.
Advances in microarray and sequencing technologies have made possible the interrogation of biological processes at increasing levels of complexity. The underlying biomolecular networks contain large numbers of nodes, yet interactions within the networks are not known precisely. In the absence of accurate models, one may inquire if it is possible to find relationships between the states of such networks under external changes, and in particular, if such relationships can be model-independent.
View Article and Find Full Text PDFCogn Neurodyn
October 2024
Institute for Cognitive Neurodynamics, Center for Intelligent Computing, School of Mathematics, East China University of Science and Technology, 130 Meilong Road, Shanghai, 200237 China.
Numerous electrophysiological experiments have reported that the prefrontal cortex (PFC) is involved in the process of working memory. PFC neurons continue firing to maintain stimulus information in the delay period without external stimuli in working memory tasks. Further findings indicate that while the activity of single neurons exhibits strong temporal and spatial dynamics (heterogeneity), the activity of population neurons can encode spatiotemporal information of stimuli stably and reliably.
View Article and Find Full Text PDFEnter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!