Deep neural networks (DNNs) usually contain massive parameters, but there is redundancy such that it is guessed that they could be trained in low-dimensional subspaces. In this paper, we propose a Dynamic Linear Dimensionality Reduction (DLDR) based on the low-dimensional properties of the training trajectory. The reduction method is efficient, supported by comprehensive experiments: optimizing DNNs in 40-dimensional spaces can achieve comparable performance as regular training over thousands or even millions of parameters. Since there are only a few variables to optimize, we develop an efficient quasi-Newton-based algorithm, obtain robustness to label noise, and improve the performance of well-trained models, which are three follow-up experiments that can show the advantages of finding such low-dimensional subspaces. The code is released (Pytorch: https://github.com/nblt/DLDR and Mindspore: https://gitee.com/mindspore/docs/tree/r1.6/docs/sample_code/dimension_reduce_training).

Download full-text PDF

Source
http://dx.doi.org/10.1109/TPAMI.2022.3178101DOI Listing

Publication Analysis

Top Keywords

low-dimensional subspaces
8
low dimensional
4
dimensional trajectory
4
trajectory hypothesis
4
hypothesis true
4
true dnns
4
dnns trained
4
trained tiny
4
tiny subspaces
4
subspaces deep
4

Similar Publications

In recent years, automated reaction path search methods have established the concept of a reaction route network. The Reaction Space Projector (ReSPer) visualizes the potential energy hypersurface into a lower-dimensional subspace using principal coordinates. The main time-consuming process in ReSPer is calculating the structural distance matrix, making it impractical for complex organic reaction route networks.

View Article and Find Full Text PDF
Article Synopsis
  • * This study focuses on the olfactory system, looking at how neurons in this area may communicate through a special region called a "communication subspace."
  • * The findings show that these neuronal groups in the olfactory pathway communicate based on breathing patterns and can send information back and forth, with the effectiveness of this communication affected by anesthesia.
View Article and Find Full Text PDF
Article Synopsis
  • Sensory processing relies on communication among different brain regions, but neural response variability can limit how effectively this information is transmitted.
  • The authors propose a mathematical framework that combines two concepts—linear Fisher information and communication subspaces—to analyze how this variability affects sensory information flow.
  • Their approach provides a way to partition information based on the relationship between population responses, aiming to help researchers better understand and design experiments related to how sensory information is shared in the brain.
View Article and Find Full Text PDF

Advances in microarray and sequencing technologies have made possible the interrogation of biological processes at increasing levels of complexity. The underlying biomolecular networks contain large numbers of nodes, yet interactions within the networks are not known precisely. In the absence of accurate models, one may inquire if it is possible to find relationships between the states of such networks under external changes, and in particular, if such relationships can be model-independent.

View Article and Find Full Text PDF

A working memory model based on recurrent neural networks using reinforcement learning.

Cogn Neurodyn

October 2024

Institute for Cognitive Neurodynamics, Center for Intelligent Computing, School of Mathematics, East China University of Science and Technology, 130 Meilong Road, Shanghai, 200237 China.

Numerous electrophysiological experiments have reported that the prefrontal cortex (PFC) is involved in the process of working memory. PFC neurons continue firing to maintain stimulus information in the delay period without external stimuli in working memory tasks. Further findings indicate that while the activity of single neurons exhibits strong temporal and spatial dynamics (heterogeneity), the activity of population neurons can encode spatiotemporal information of stimuli stably and reliably.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!