e3nn-tutorial-mrs-fall-2021

e3nn tutorial for Materials Research Society Fall Meeting 2021

Spacetime coordinates
Tutorials
    Intro
    Group Theory
    Invariants
    Molecular Dynamics
    Electron Densities
    Phonon Properties
Related Talks @MRS
Instructors

e3nn webpage
NequIP repository

MRS 2021 Fall Meeting Tutorial:
Symmetry-Aware Neural Networks
for the Materials Science with
e3nn

Understanding the role of symmetry in the physical sciences is critical for choosing an appropriate machine-learning method.

In this tutorial we will present how to incorporate Euclidean and permutation symmetries into neural networks and the benefits of these techniques for predicting a variety of properties of atomic systems (molecules, crystal, and beyond).

We will provide a cloud-based notebook environment for participants to run the code example used throughout the tutorial. Participants will leave the tutorial with theory and code resources in hand and a practical working knowledge of how symmetry considerations impact algorithm design in machine learning and beyond.

Spacetime coordinates

Monday, November 29, 2021
8:30 AM - 5:00 PM
Hynes, Level 2, Room 205

Hynes Convention Center Floor Plans

Schedule

8:30 - 9:30am Tutorial 1: Intro to Symmetry in ML
9:30 - 10:00am BREAK
10:00 - 11:00am Tutorial 2: Group theory, irreps, and tensor products
11:00 - noon Tutorial 3: Invariant functions for geometry of atomic structures
noon - 1:30pm LUNCH
1:30-2:30pm Tutorial 4: Molecular dynamics
2:30-3:00pm BREAK
3:00-4:00pm Tutorial 5: Electron densities
4:00-5:00pm Tutorial 6: Phonon properties

Tutorials and Materials

YouTube playlist for tutorial

Tutorial 1: Euclidean Symmetry in Machine Learning for Materials Science by Tess Smidt

Video | Slides

Tutorial 2: Group theory, irreducible representations, and tensor products and how to use them in e3nn to build Euclidean Neural Networks by Mario Geiger

Video | Slides

Tutorial 3: Analyzing geometry and structure of atomic configurations with equivariant and invariant functions by Martin Uhrin and Thomas Hardin

Video | Slides (Part 1 | Part 2) | Colab | Code

Tutorial 4: Molecular dynamics with NequIP by Simon Batzner and Alby Musaelian

Video | Slides | Colab | NequIP | example.yaml

Tutorial 5: Predicting Electron Densities with e3nn by Josh Rackers

Video | Slides | Colab

Tutorial 6: Predicting Phonon Properties of Crystal Structures by Nina Andrejevic and Zhantao Chen

Video | Slides | Colab | Code

Related Talks at MRS 2021 Fall Meeting

EQ04.01.03
Using Complete, Symmetry Invariant Representations of Atomic Environments to Predict New Ionic Liquids from Experimental Data

Martin Uhrin

Tuesday 11 AM, November 30, 2021
Hynes, Level 2, Room 205

The ultimate goal of any atomistic design process, be it for materials or molecules, is to start with one or more target properties and predict atomic geometries that are statistically likely to possess these properties. In this talk, I will discuss how to construct complete, symmetry invariant descriptions of atomic environments that can be inverted to recover the original environment. Using this representation I will show that it is possible to build a model that can predict the conductivity of ionic liquids, an important potential electrolyte for future battery technologies. By training on a dataset of experimental results the model learns how to attribute contributions to the overall conductivity to fragments of the anion and cation. This allows us to virtually screen thousands of anion/cation combinations to pinpoint those with high conductivities. Furthermore, it opens the door to designing entirely new molecules by combining fragments to achieve properties that go beyond those that are possible using current ionic liquids.

DS03.06.03
NequIP—E(3)-Equivariant Convolutions Enable Sample-Efficient, Scalable and Highly Accurate Machine Learning Interatomic Potentials

Simon Batzner, Albert Musaelian, Lixin Sun, Tess Smidt, Mario Geiger, Jonathan Mailoa, Mordechai Kornbluth, Nicola Molinari, Boris Kozinsky

Thursday 1:45 PM, November 30, 2021
Sheraton, 5th Floor, The Fens

We present Neural Equivariant Interatomic Potentials (NequIP), an E(3)-equivariant deep learning approach for learning interatomic potentials for molecular dynamics simulations. Instead of the commonly deployed invariant convolutions over scalar features, NequIP uses E(3)-equivariant convolutions over geometric tensors, better representing the symmetries of Euclidean space. The proposed model obtains state-of-the-art accuracy on a challenging set of diverse molecules and materials while at the same exhibiting remarkable sample efficiency. Interestingly, NequIP outperforms existing, invariant models with up to three orders of magnitude fewer training data and performs better than kernel methods, even on tiny data sets, thereby challenging the widely held belief that deep neural networks require massive training sets. We show results from molecular dynamics simulations using NequIP on a series of technologically relevant bulk materials as well as the folding of small proteins. The method is implemented in a scalable and highly efficient software implementation, integrated with the molecular dynamics code LAMMPS, and can be used to simulate large time- and length-scales at high accuracy and low computational cost.

DS03.06.02
DICE—A Linear-Scaling N-Body Interatomic Potential from E(3)-Equivariant Convolutions

Albert Musaelian, Simon Batzner, Lixin Sun, Steven Torrisi, Boris Kozinsky

Thursday 2:00 PM, November 30, 2021
Sheraton, 5th Floor, The Fens

Message Passing Graph Neural Networks (MPNNs) based on pairwise interactions have emerged as the leading paradigm for modeling atomistic systems by recursively propagating information along a molecular graph. While MPNNs have consistently been demonstrated to give low generalization errors, they inherently have a low level of interpretability, are not systematically improvable, and are difficult to scale to large numbers of atoms. Here, we introduce the Deep Interatomic Cluster Expansion (DICE), an equivariant neural network that leverages many-body information in a single interaction, without the need for message passing or convolutions. The method can be systematically improved by including higher-order interactions at linear cost, has physically meaningful hyperparameters, and is embarrassingly parallel. DICE builds on a novel, learnable E(3)-equivariant many-body representation that utilizes weighted tensor products of geometric features to describe N-point correlations of atoms. The proposed many-body representation overcomes the exponential scaling of a naive cluster expansion and instead scales linearly with the number of simultaneously correlated particles. We demonstrate that the use of higher-order correlations of atoms systematically improves the accuracy. We further find that DICE gives excellent performance across a wide variety of settings, outperforming both MPNNs as well as kernel-based methods on small data sets.

EN12.08.08
Neural Equivariant Interatomic Potentials for Large-Scale Electrolyte Simulations

Nicola Molinari, Albert Musaelian, Juan Felipe Gomez, Robert Newman, Simon Batzner, Boris Kozinsky

Thursday 3:15 PM, December 2, 2021
Hynes, Level 3, Ballroom A

Electrolytes control battery recharge time and efficiency, anode/cathode stability, and ultimately safety, thus their optimization is crucial for the design of modern energy storage devices. In this work, we focus on ionic liquid-based electrolytes. This promising class of electrolytes possesses superior chemical stability compared to standard organic solvents, however, poor and anomalous transport properties are hindering their applicability [1]. Unfortunately, classical energy models struggle to provide the level of accuracy needed to reliably predict and investigate ionic conductivity.

Here we study the applicability to such systems of state-of-the-art equivariant graph neural networks (NequIP [2]). Ionic liquid-based electrolytes provide a unique challenge due to their strong ionic interactions and liquid nature. Additionally, substantially diverse inter-atomic environments are expected as a function of lithium-salt doping [3], raising the question of model transferability. Our promising results show that we can circumnavigate the ad-hoc and system-dependent tweaks to the classical energy models (for instance charge rescaling or polarizability), while retaining computational speed and near-DFT accuracy for large-scale ionic transport molecular dynamics investigations.

Instructors

Tess Smidt (tsmidt@mit.edu)

Tess Smidt is an Assistant Professor of Electrical Engineering and Computer Science at MIT. Tess earned her SB in Physics from MIT in 2012 and her PhD in Physics from the University of California, Berkeley in 2018. Her research focuses on machine learning that incorporates physical and geometric constraints, with applications to materials design. Prior to joining the MIT EECS faculty, she was the 2018 Alvarez Postdoctoral Fellow in Computing Sciences at Lawrence Berkeley National Laboratory and a Software Engineering Intern on the Google Accelerated Sciences team where she developed Euclidean symmetry equivariant neural networks which naturally handle 3D geometry and geometric tensor data.

Mario Geiger (geiger.mario@gmail.com)

Mario Geiger is a PhD in the laboratory of Physics of Complex Systems at EPFL in Switzerland. He is a physicist and he studies the dynamics of neural networks. He also studies the theory of equivariant neural networks and is the main developer of e3nn, a library for neural networks aware of the Euclidean symmetries.

Martin Uhrin

Martin Uhrin is a senior postdoctoral research in the group of Nicola Marzari at EPFL in Switzerland. The focus of his research is the development and application of methods that accelerate materials discovery, particularly in the area of prediction and rationalisation of crystal structures. To this end, Martin has built up expertise in ab-initio structure prediction, empirical potential development, cluster prediction and deep learning. He has also contributed to the development of many community tools and was lead architect of the AiiDA workflow engine, part of the widely used AiiDA materials informatics platform. Martin earned his Ph.D. in computational physics from the University College London under the supervision of Chris Pickard. He holds an MPhys. in computational physics from the University of Edinburgh.

Thomas Hardin

Thomas Hardin is a Truman Fellow at Sandia National Laboratories. Thomas designs disordered materials and uses machine learning to discover low-dimensional manifolds underpinning the structure of glass. His other research work with disordered materials includes mesoscale and atomistic modeling, additive manufacturing, and microparticle impact testing. Thomas earned his doctorate degree in Materials Science and Engineering from the Massachusetts Institute of Technology, working with Professor Chris Schuh, and a dual bachelor’s degree in Mathematics and Mechanical Engineering from Brigham Young University, working with Professors Brent Adams and Eric Homer.

Simon Batzner

Simon Batzner is a PhD student at Harvard University, working under the supervision of Prof. Boris Kozinsky. His research interests focus on the development of Machine-Learning Interatomic Potentials for accelerated Molecular Dynamics simulations. Before joining Harvard, he obtained his Master’s at MIT, where he worked with Prof. Alexie Kolpak on problems in atomistic Machine Learning. He obtained his bachelor’s degree at the University of Stuttgart, Germany, during which he also spent one year working on the SOFIA mission at the NASA Armstrong Flight Research Center.

Alby Musaelian

Albert Musaelian is a PhD student in Boris Kozinsky’s Materials Intelligence Research (MIR) group at Harvard University. He works on the design and implementation of equivariant machine learning methods for atomistic simulations with a particular focus on machine-learning interatomic potentials.

Josh Rackers

Josh Rackers is a Truman Fellow at Sandia National Labs. Josh is interested in developing methods to enable quantum-accurate simulation of biological molecules. To this end, he is involved in research spanning from quantum chemistry through classical molecular dynamics simulation. Josh is particularly interested in using physics-aware machine learning (ML) models to predict charge densities of large biomolecular systems. Josh earned a Ph.D. in Biophysics from Washington University in St. Louis and a B.S. in Physics and Political Science from Ohio State University. Before venturing into graduate school, he taught high school physics and chemistry in Baltimore, Maryland.

Nina Andrejevic

Nina Andrejevic is a PhD student at the Massachusetts Institute of Technology working with Prof. Mingda Li in the Quantum Matter Group. Nina’s research interests include investigating electronic and phononic properties of emergent quantum materials using neutron and X-ray spectroscopies, as well as developing complementary quantum theories and machine learning methods to help interpret and extract key insights from experiments.

Zhantao Chen

Zhantao Chen is a PhD student at the Massachusetts Institute of Technology working with Prof. Mingda Li in the Quantum Matter Group and Prof. Jing Kong. Zhantao’s research interests include merging machine learning methods into computational physics to help extract knowledge from scattering measurements.