Designing rotationally invariant neural networks from PDEs and variational methods.

Res Math Sci

Mathematical Image Analysis Group, Faculty of Mathematics and Computer Science, Campus E1.7, Saarland University, 66041 Saarbrücken, Germany.

Published: August 2022

Partial differential equation models and their associated variational energy formulations are often rotationally invariant by design. This ensures that a rotation of the input results in a corresponding rotation of the output, which is desirable in applications such as image analysis. Convolutional neural networks (CNNs) do not share this property, and existing remedies are often complex. The goal of our paper is to investigate how diffusion and variational models achieve rotation invariance and transfer these ideas to neural networks. As a core novelty, we propose activation functions which couple network channels by combining information from several oriented filters. This guarantees rotation invariance within the basic building blocks of the networks while still allowing for directional filtering. The resulting neural architectures are inherently rotationally invariant. With only a few small filters, they can achieve the same invariance as existing techniques which require a fine-grained sampling of orientations. Our findings help to translate diffusion and variational models into mathematically well-founded network architectures and provide novel concepts for model-based CNN design.

Download full-text PDF

Source
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC9352643PMC
http://dx.doi.org/10.1007/s40687-022-00339-xDOI Listing

Publication Analysis

Top Keywords

rotationally invariant
12
neural networks
12
diffusion variational
8
variational models
8
rotation invariance
8
designing rotationally
4
neural
4
invariant neural
4
networks
4
networks pdes
4

Similar Publications

Group-equivariant neural networks have emerged as an efficient approach to model complex data, using generalized convolutions that respect the relevant symmetries of a system. These techniques have made advances in both the supervised learning tasks for classification and regression, and the unsupervised tasks to generate new data. However, little work has been done in leveraging the symmetry-aware expressive representations that could be extracted from these approaches.

View Article and Find Full Text PDF

A simple approach to rotationally invariant machine learning of a vector quantity.

J Chem Phys

November 2024

J. Heyrovský Institute of Physical Chemistry, Academy of Sciences of the Czech Republic, v.v.i., Dolejškova 3, 18223 Prague 8, Czech Republic.

Article Synopsis
  • Machine learning (ML) for predicting vector or tensor properties requires maintaining proper invariance with molecular rotation, unlike energy prediction which is simpler due to its scalar nature.
  • The proposed "rotate-predict-rotate" (RPR) technique involves three steps: rotating the molecule to align with its principal axes, predicting the vector property in that orientation, and then transforming the prediction back to the original orientation.
  • This RPR approach ensures covariance for vector properties and can extend to tensors, allowing for rapid training of accurate ML models across numerous molecular configurations.
View Article and Find Full Text PDF

Computationally Efficient Direction Finding for Conformal MIMO Radar.

Sensors (Basel)

September 2024

College of Computer and Information Technology, China Three Gorges University, Yichang 443002, China.

The use of conformal arrays offers a significant advancement in Multiple-Input-Multiple-Output (MIMO) radar, enabling the placement of antennas on irregular surfaces. For joint Direction-of-Departure (DOD) and Direction-of-Arrival (DOA) estimation in conformal-array MIMO radar, the current spectrum-searching methods are computationally too expensive, while the existing rotation-invariant method may suffer from phase ambiguity caused by the non-Nyquist spacing of the sensors. In this paper, an improved rotationally invariant technique is proposed.

View Article and Find Full Text PDF

Deciphering Molecular Embeddings with Centered Kernel Alignment.

J Chem Inf Model

October 2024

Department of Pharmaceutical Sciences, Division of Pharmaceutical Chemistry, Faculty of Life Sciences, University of Vienna, Josef-Holaubek-Platz 2, Vienna 1090, Austria.

Analyzing machine learning models, especially nonlinear ones, poses significant challenges. In this context, centered kernel alignment (CKA) has emerged as a promising model analysis tool that assesses the similarity between two embeddings. CKA's efficacy depends on selecting a kernel that adequately captures the underlying properties of the compared models.

View Article and Find Full Text PDF

Excited-state molecular dynamics simulations are crucial for understanding processes like photosynthesis, vision, and radiation damage. However, the computational complexity of quantum chemical calculations restricts their scope. Machine learning offers a solution by delivering high-accuracy properties at lower computational costs.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!