68
Tess Smidt 2018 Alvarez Postdoctoral Fellow in Computing Sciences Euclidean Neural Networks... rotation-, translation-, and permutation-equivariant convolutional neural networks for 3D point clouds ...for emulating ab initio calculations and generating atomic geometries. From Passive to Active: Generative and Reinforcement Learning with Physics Machine Learning for Physics at IPAM 2019.09.27

Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Tess Smidt2018 Alvarez Postdoctoral Fellow in Computing Sciences

Euclidean Neural Networks...rotation-, translation-, and permutation-equivariant convolutional neural networks for 3D point clouds

...for emulating ab initio calculations and generating atomic geometries.

From Passive to Active: Generative and Reinforcement Learning with Physics

Machine Learning for Physics at IPAM2019.09.27

Page 2: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

2

...where the electrons are...

Given an atomic structure,

Ene

rgy

(eV

)

Momentum

...and what the electrons are doing.

...use quantum theory and supercomputers to determine...

Si

What a computational materials physicist does:

Structure

Properties

Page 3: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

StructureQuantum Theory + Supercomputers

Properties

Hypothesize

Inverse Design

Zooooom!

We want to use deep learning to speed up these calculations, hypothesize new structures, perform inverse design, and organize these relations.

Map

Page 4: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

StructureQuantum Theory + Supercomputers

Properties

Hypothesize

Inverse Design

Zooooom!

We want to use deep learning to speed up these calculations, hypothesize new structures, perform inverse design, and organize these relations.

Map

What types of neural networks are best suited for these tasks?

Page 5: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

5

Neural networks are specially designed for different data types. Assumptions about the data type are built into how the network operates. W x

Page 6: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

6

Neural networks are specially designed for different data types. Assumptions about the data type are built into how the network operates.

Vectors ⇨ Dense NN

Components are independent.

W x

Page 7: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

7

Neural networks are specially designed for different data types. Assumptions about the data type are built into how the network operates.

Vectors ⇨ Dense NN 2D images ⇨ Convolutional NN

Components are independent. The same features can be found anywhere in an image. Locality.

W x

Page 8: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

8

Neural networks are specially designed for different data types. Assumptions about the data type are built into how the network operates.

Vectors ⇨ Dense NN 2D images ⇨ Convolutional NN Text ⇨ Recurrent NN

Components are independent. The same features can be found anywhere in an image. Locality.

Sequential data. Next input/output depends on input/output that has come before.

W x

Page 9: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

9

Neural networks are specially designed for different data types. Assumptions about the data type are built into how the network operates.

Vectors ⇨ Dense NN 2D images ⇨ Convolutional NN Text ⇨ Recurrent NN

Components are independent. The same features can be found anywhere in an image. Locality.

Sequential data. Next input/output depends on input/output that has come before.

What are our data types in materials physics? How do we build neural networks for these data types?

W x

Page 10: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

10

What assumptions do we want “built in” to our neural networks (for materials data)?

Atomic systems form geometric motifs that can appear at multiple locations and orientations.

The properties of physical systems transform predictably under rotation.

Two point masses with velocity and acceleration.

Same system, with rotated coordinates.

Page 11: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

11

What assumptions do we want “built in” to our neural networks (for materials data)?

Atomic systems form geometric motifs that can appear at multiple locations and orientations.

The properties of physical systems transform predictably under rotation.

Two point masses with velocity and acceleration.

Same system, with rotated coordinates.

Our data types are geometry and geometric tensors.These data types assume Euclidean symmetry (3D translations, 3D rotations, and inversion).

Page 12: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

12

Space has Euclidean symmetry, E(3). Objects break that symmetry. The broken symmetry is a subgroup of E(3).

Page 13: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

13

Space has Euclidean symmetry, E(3). Objects break that symmetry. The broken symmetry is a subgroup of E(3).

O(3)

3D rotations and inversions

Page 14: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

14

Space has Euclidean symmetry, E(3). Objects break that symmetry. The broken symmetry is a subgroup of E(3).

O(3) SO(2) + mirrors

(C∞v)

3D rotations and inversions

2D rotation and mirrors along cone axis

Page 15: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

15

Space has Euclidean symmetry, E(3). Objects break that symmetry. The broken symmetry is a subgroup of E(3).

O(3) SO(2) + mirrors

(C∞v)

Oh

3D rotations and inversions

2D rotation and mirrors along cone axis

Discrete rotations and mirrors

Page 16: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

16

Space has Euclidean symmetry, E(3). Objects break that symmetry. The broken symmetry is a subgroup of E(3).

O(3) Oh Pm-3m(221)

SO(2) + mirrors

(C∞v)

3D rotations and inversions

2D rotation and mirrors along cone axis

Discrete rotations and mirrors

Discrete rotations, mirrors, and translations

Page 17: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

17

Properties of a system must be compatible with symmetry.Which of these situations (inputs / outputs) are symmetrically allowed / forbidden?

m m

m m

m m

a.

b.

c.

Page 18: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

18

m m

m m

m m

a.

b.

c.

Properties of a system must be compatible with symmetry.Which of these situations (inputs / outputs) are symmetrically allowed / forbidden?

Page 19: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

19

m m

m m

m m

a.

b.

c.

m 2m

Properties of a system must be compatible with symmetry.Which of these situations (inputs / outputs) are symmetrically allowed / forbidden?

Page 20: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

20

m m

m m

m m

a.

b.

c.

m 2m

m mg

Properties of a system must be compatible with symmetry.Which of these situations (inputs / outputs) are symmetrically allowed / forbidden?

Page 21: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

21

We build neural networks with Euclidean symmetry, E(3) and SE(3).

● What neural networks with Euclidean symmetry can do.

● How Euclidean Neural Networks work.

● Applications of Euclidean Neural Networks.

Page 22: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Trained on 3D Tetris shapes in one orientation, these network can perfectly identify these shapes in any orientation.

TRA

INTE

ST

22

Chiral

Page 23: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

23

Given a molecule and a rotated copy, the predicted forces are the same up to rotation.(Predicted forces are equivariant to rotation.)

Page 24: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

24

To these networks, primitive unit cells, conventional unit cells, and supercells of the same crystal will produce the same output (assuming periodic boundary conditions).

Page 25: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

25

We build neural networks with Euclidean symmetry, E(3) and SE(3).

● What neural networks with Euclidean symmetry can do.

● How Euclidean Neural Networks work.○ Overview○ Input to network○ Network operations○ Visualizing kernels○ Interpreting input / output

● Applications of Euclidean Neural Networks.

Page 26: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

We use points. Images of atomic systems are sparse and imprecise.

vs.

We encode the symmetries of 3D Euclidean space (3D translation- and 3D rotation-equivariance).

Other atoms

Convolution center

We use continuous convolutions with atoms as convolution centers.

26

K. T. Schütt et al, NIPS 30 (2017). (arXiv: 1706.08566)

Page 27: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

We use points. Images of atomic systems are sparse and imprecise.

vs.

We encode the symmetries of 3D Euclidean space (3D translation- and 3D rotation-equivariance).

Other atoms

Convolution center

We use continuous convolutions with atoms as convolution centers.

27

K. T. Schütt et al, NIPS 30 (2017). (arXiv: 1706.08566)

Page 28: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

We use points. Images of atomic systems are sparse and imprecise.

vs.

We encode the symmetries of 3D Euclidean space (3D translation- and 3D rotation-equivariance).

Other atoms

Convolution center

We use continuous convolutions with atoms as convolution centers.

28

K. T. Schütt et al, NIPS 30 (2017). (arXiv: 1706.08566)

Page 29: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Translation equivariance

Rotation equivariance

29

Page 30: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Translation equivarianceConvolutional neural network ✓

Rotation equivariance?

30

Page 31: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Translation equivarianceConvolutional neural network ✓

Rotation equivarianceData augmentationRadial functions Want a network that both preserves geometry and exploits symmetry.

31

Page 32: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

32

Several groups converged on similar ideas around the same time.

Tensor field networks: Rotation- and translation-equivariant neural networks for 3D point clouds(arXiv:1802.08219)Tess Smidt*, Nathaniel Thomas*, Steven Kearnes, Lusann Yang, Li Li, Kai Kohlhoff, Patrick Riley

Points, nonlinearity on norm of tensors

Clebsch-Gordan Nets: a Fully Fourier Space Spherical Convolutional Neural Network(arXiv:1806.09231)Risi Kondor, Zhen Lin, Shubhendu Trivedi

Only use tensor product as nonlinearity, no radial function

3D Steerable CNNs: Learning Rotationally Equivariant Features in Volumetric Data(arXiv:1807.02547)Mario Geiger*, Maurice Weiler*, Max Welling, Wouter Boomsma, Taco Cohen

Efficient framework for voxels, gated nonlinearity

*denotes equal contribution

Page 33: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

33

Several groups converged on similar ideas around the same time.

Tensor field networks: Rotation- and translation-equivariant neural networks for 3D point clouds(arXiv:1802.08219)Tess Smidt*, Nathaniel Thomas*, Steven Kearnes, Lusann Yang, Li Li, Kai Kohlhoff, Patrick Riley

Points, nonlinearity on norm of tensors

Clebsch-Gordan Nets: a Fully Fourier Space Spherical Convolutional Neural Network(arXiv:1806.09231)Risi Kondor, Zhen Lin, Shubhendu Trivedi

Only use tensor product as nonlinearity, no radial function

3D Steerable CNNs: Learning Rotationally Equivariant Features in Volumetric Data(arXiv:1807.02547)Mario Geiger*, Maurice Weiler*, Max Welling, Wouter Boomsma, Taco Cohen

Efficient framework for voxels, gated nonlinearity

*denotes equal contribution

Tensor field networks + 3D steerable CNNs = Euclidean neural networks

Page 34: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

34

To be rotation-equivariant means that we can rotate our inputs OR rotate our outputs and we get the same answer.

Layerin outRot

Layerin outRot=

Page 35: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

35

The input to our network is geometry and features on that geometry.

Page 36: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

36

The input to our network is geometry and features on that geometry.We categorize our features by how they transform under rotation.

Features have “angular frequency” L where L is a positive integer.

Scalars

Vectors

3x3 Matrices

Frequency

Doesn’t change with rotation

Changes with same frequency as rotation

Page 37: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

37

Other atoms

Convolution center

with no symmetry:

with SO(3) symmetry:

Learned Parameters

The convolutional kernels are built from functions with “angular frequency” L⇨ Spherical harmonics.

Page 38: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

38

Page 39: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Let g be a 3d rotation matrix.

a-1 +a0 +a1

=

D is the Wigner-D matrix. It has shape and is a function of g.

Spherical harmonics of a given L transform together under rotation.

g

39

b-1 +b0 +b1

D

Page 40: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

40

Features and kernels are not simply scalars.We use tensor products with Clebsch-Gordan coefficients to combine.

D. Griffiths, Introduction to quantum mechanics

Same math involved in the addition of angular momentum.

Page 41: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Examples of tensor product: How to combine a scalar and a vector? Easy!

41

Angular Frequency

Page 42: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Examples of tensor product: How to combine two vectors? Many ways.

42

Dot product

Crossproduct

Outer product

Angular Frequency

Page 43: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

43

Features and kernels are not simply scalars.We use tensor products and Clebsch-Gordan coefficients to combine.

Scalar multiply

“i” and “j” are scalar channels

Scalar convolution

Page 44: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

44

Features and kernels are not simply scalars.We use tensor products and Clebsch-Gordan coefficients to combine.

Tensor convolutionScalar

multiply

Tensor product + Clebsch-Gordan:Takes ij -> k

“channel” includes specific spherical harmonic L“i” and “j” are scalar channels

Scalar convolution

Page 45: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

For L=1 ⇨ L=1, the filters will be a learned, radially-dependent linear combinations of the L = 0, 1, and 2 spherical harmonics.

45

L=2L=0 L=1

+ +

(3 x 3) x 3 ⇨ 3(3 x 3) = 1 + 3 + 5

Page 46: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

For L=1 ⇨ L=1, the filters will be a learned, radially-dependent linear combinations of the L = 0, 1, and 2 spherical harmonics.

46

L=2L=0 L=1

+ +

(3 x 3) x 3 ⇨ 3(3 x 3) = 1 + 3 + 5

Page 47: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

47

We can interpret our outputs as numerical features...

Scalars● Energy● Mass● Isotropic *

...

Vectors● Force● Velocity● Acceleration● Polarization

...

Matrices, Tensors, …● Moment of Inertia● Polarizability● Interaction of multipoles

...

m

Page 48: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

48

We can interpret our outputs as numerical features or geometry.

Output (0 ≤ L< 6 coefficients) of randomly initialized network applied to a tetrahedron with a center.

Page 49: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

49

We can interpret our outputs as numerical features or geometry.

Can generate point sets from signal peaks as output!Sets == permutation invariant

Output (0 ≤ L< 6 coefficients) of randomly initialized network applied to a tetrahedron with a center.

Page 50: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

50

We build neural networks with Euclidean symmetry, E(3) and SE(3).

● What neural networks with Euclidean symmetry can do.

● How Euclidean Neural Networks work.

● Applications of Euclidean Neural Networks.

Page 51: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Applications: Predicting ab initio forces for molecular dynamicsSimon Batzner (MIT/Harvard) and Boris Kozinsky (Harvard)Presented at APS March Meeting 2019

51

Dataset: MD17ab initio molecular dynamics trajectories of...

Direct prediction of forces rather than gradient of scalar energy.

Page 52: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Applications: Predicting molecular Hamiltonians with atom-centered basis sets

K. T. Schütt, M. Gastegger, A. Tkatchenko, K.-R. Müller, R. J. Maurer. arXiv:1906.10033 (2019)

Predict Hamiltonian matrix and get eigenvectors (wavefunctions) and eigenvalues (energies).

Page 53: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

K. T. Schütt, M. Gastegger, A. Tkatchenko, K.-R. Müller, R. J. Maurer. arXiv:1906.10033 (2019)

Predict Hamiltonian matrix and get eigenvectors (wavefunctions) and eigenvalues (energies).

Problem! Hamiltonian depends on coordinate system -- traditionally requires augmenting data.

Applications: Predicting molecular Hamiltonians with atom-centered basis sets

Page 54: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

O 1s 2s 2s 2p 2p 3d H 1s 2s 2p H 1s 2s 2p

Applications: Predicting molecular Hamiltonians with atom-centered basis sets

With Euclidean neural networks -- only need one example.Output is guaranteed to be equivariant!

Page 55: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

W. Wang, R. Gómez-Bombarelli. arXiv:1812.02706 (revised 2019)Applications: Coarse-grained geometries and recover all atoms picture

Page 56: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

W. Wang, R. Gómez-Bombarelli. arXiv:1812.02706 (revised 2019)

Tetrahedral chain.Centers of a tetrahedral chain.

Predict using spherical harmonic

signal

Test problem

Applications: Coarse-grained geometries and recover all atoms picture

Page 57: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Without symmetry breaking

Peaks == point locations

Symmetric configurations == degeneracy

Applications: Coarse-grained geometries and recover all atoms picture

Page 58: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Without symmetry breaking With symmetry breaking

Peaks == point locations

Symmetric configurations == degeneracy

Use second input to break symmetry

Applications: Coarse-grained geometries and recover all atoms picture

Page 59: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

59

Applications: Creating an autoencoder for discrete geometry

ContinuousLatent Representation(N dimensional vector)

Discrete geometry Discrete geometry

Reduce geometry to single point.

Create geometry from single point.

Page 60: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

60

ContinuousLatent Representation(N dimensional vector)

Discrete geometry Discrete geometry

Reduce geometry to single point.

Create geometry from single point.

Atomic structures are hierarchical and can be constructed from geometric motifs.

+ Encode geometry ✓+ Encode hierarchy ?+ Decode geometry ?+ Decode hierarchy ?

(Need to do this in a recursive manner)

Applications: Creating an autoencoder for discrete geometry

Page 61: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

61

Convolve

Bloom(Make copies and move)

Symmetric Cluster (Keep track of point origins)

Combine (CNN and Cluster Info)

Geometry

New

Geom

etry

*Edges are shown for visualization. May not be included.

How to encode: Recursively convert geometry to a vector

Page 62: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

62

How to decode: Recursively convert a vector to geometry

Bloom with new features

Cluster (Merge duplicate points)

Combine (Bloom features and

Cluster Info)

Geometry

New

Geom

etry

*Edges are shown for visualization. May not be included.

1st

2nd

Page 63: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

tensor field networks

Google Accelerated Science Team Stanford

Patrick Riley

Steve Kearnes

Nate Thomas

Lusann Yang

Kai Kohlhoff

LiLi

atomic architectssummer 2019

Hashim Piracha Mario Geiger Ben MillerTess Smidt

Page 64: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

● Euclidean neural networks operate on points/voxels and have symmetries of E(3).● Inputs to the network lower this symmetry to a subgroup of E(3).● Symmetry of outputs are constrained to the symmetry of the inputs.● The inputs and outputs of our network are geometry and geometric tensors. ● Convolutional filters are built from spherical harmonics with a learned radial function.

Applications: Molecular dynamics, predicting Hamiltonians, coarse-graining, autoencoders...

We expect these networks to be generally useful for physics, chemistry, and geometry.Reach out to me if you are interested and/or have any questions!

64

Tess [email protected]

se3cnn Code (PyTorch):https://github.com/mariogeiger/se3cnn

Tensor Field Networks (arXiv:1802.08219)3D Steerable CNNs (arXiv:1807.02547)

Page 65: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

Calling in backup (slides)!

65

Page 66: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

66

Features and kernels are not simply scalars.We use tensor products and Clebsch-Gordan coefficients to combine.

Page 67: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

67

Features and kernels are not simply scalars.We use tensor products and Clebsch-Gordan coefficients to combine.

singlet

triplet

Page 68: Euclidean Neural Networks From Passive to Active ...helper.ipam.ucla.edu/publications/mlpws1/mlpws1_16346.pdfWe want to use deep learning to speed up these calculations, hypothesize

68

Features and kernels are not simply scalars.We use tensor products and Clebsch-Gordan coefficients to combine.

L=0 L=1 L=2