## Abstract

In molecular dynamics (MD), neural network (NN) potentials trained bottom-up on quantum mechanical data have seen tremendous success recently. Top-down approaches that learn NN potentials directly from experimental data have received less attention, typically facing numerical and computational challenges when backpropagating through MD simulations. We present the Differentiable Trajectory Reweighting (DiffTRe) method, which bypasses differentiation through the MD simulation for time-independent observables. Leveraging thermodynamic perturbation theory, we avoid exploding gradients and achieve around 2 orders of magnitude speed-up in gradient computation for top-down learning. We show effectiveness of DiffTRe in learning NN potentials for an atomistic model of diamond and a coarse-grained model of water based on diverse experimental observables including thermodynamic, structural and mechanical properties. Importantly, DiffTRe also generalizes bottom-up structural coarse-graining methods such as iterative Boltzmann inversion to arbitrary potentials. The presented method constitutes an important milestone towards enriching NN potentials with experimental data, particularly when accurate bottom-up data is unavailable.

## Introduction

Molecular modeling has become a cornerstone of many disciplines, including computational chemistry, soft matter physics, and material science. However, simulation quality critically depends on the employed potential energy model that defines particle interactions. There are two distinct approaches for model parametrization^{1,2}: Bottom-up approaches aim at matching data from high-fidelity simulations, providing labeled data of atomistic configurations with corresponding target outputs. Labeled data allow straightforward differentiation for gradient-based optimization, at the expense of inherently limiting model accuracy to the quality imposed by the underlying data-generating simulation. On the other hand, top-down approaches optimize the potential energy model such that simulations match experimental data. From experiments, however, labeled data on the atomistic scale are not available. Experimental observables are linked only indirectly to the potential model via an expensive molecular mechanics simulation, complicating optimization.

A class of potentials with tremendous success in recent years are neural network (NN) potentials due to their flexibility and capacity of learning many-body interactions^{3,4}. The vast majority of NN potentials are trained via bottom-up methods^{5,6,7,8,9,10,11,12,13,14,15,16}. The objective is to match energies and/or forces from a data set, most commonly generated via density functional theory (DFT) for small molecules in vacuum^{17}. Within the data set distribution, state-of-the-art NN potentials have already reached the accuracy limit imposed by DFT, with the test error in predicting potential energy being around two orders of magnitude smaller than the corresponding expected DFT accuracy^{11,18}. In the limit of a sufficiently large data set without a distribution shift^{19,20} with respect to the application domain (potentially generated via active learning approaches^{21}), remaining deviations of predicted observables from experiments are attributable to uncertainty in DFT simulations^{11}—in line with literature reporting DFT being sensitive to employed functionals^{22}. More precise computational quantum mechanics models, e.g., the coupled cluster CCSD(T) method, improve DFT accuracy at the expense of significantly increased computational effort for data set generation^{23,24}. However, for larger systems such as macromolecules, quantum mechanics computations will remain intractable in the foreseeable future, preventing ab initio dataset generation altogether. Thus, the main obstacle in bottom-up learning of NN potentials is the currently limited availability of highly precise and sufficiently broad data sets.

Top-down approaches circumvent the need for reliable data-generating simulations. Leveraging experimental data in the potential optimization process has contributed greatly to the success of classical atomistic^{25,26} and coarse-grained^{27} (CG) force fields^{1}. Training difficulties have so far impeded a similar approach for NN potentials: Only recent advances in automatic differentiation (AD)^{28} software have enabled end-to-end differentiation of molecular dynamics (MD) observables with respect to potential energy parameters^{29,30}, by applying AD through the dynamics of a MD simulation^{29,30,31,32}. This direct reverse-mode AD approach saves all simulator operations on the forward pass to be used during gradient computation on the backward pass, resulting in excessive memory usage. Thus, direct reverse-mode AD for systems with more than hundred particles and a few hundred time steps is typically intractable^{29,30,31,32}. Numerical integration of the adjoint equations^{33,34} represents a memory-efficient alternative that requires to save only those atomic configurations that directly contribute to the loss. However, both approaches backpropagate the gradient through the entire simulation, which dominates computational effort and is prone to exploding gradients, as stated by Ingraham et al.^{31} and shown below.

Addressing the call for NN potentials trained on experimental data^{1}, we propose the Differentiable Trajectory Reweighting (DiffTRe) method. DiffTRe offers end-to-end gradient computation and circumvents the need to differentiate through the simulation by combining AD with previous work on MD reweighting schemes^{35,36,37,38}. For the common use case of time-independent observables, DiffTRe avoids exploding gradients and reduces the computational effort of gradient computations by around two orders of magnitude compared to backpropagation through the simulation. Memory requirements are comparable to the adjoint method. We showcase the broad applicability of DiffTRe on three numerical test cases: First, we provide insight into the training process on a toy example of ideal gas particles inside a double-well potential. Second, we train the state-of-the-art graph neural network potential DimeNet++^{11,12} for an atomistic model of the diamond from its experimental stiffness tensor. Finally, we learn a DimeNet++ model for CG water based on pressure, as well as radial and angular distribution functions. The last example shows how DiffTRe also generalizes bottom-up structural coarse-graining methods such as the iterative Boltzmann inversion^{39} or inverse Monte Carlo^{40} to many-body correlation functions and arbitrary potentials. DiffTRe allows to enhance NN potentials with experimental data, which is particularly relevant for systems where bottom-up data are unavailable or not sufficiently accurate.

## Results

### Differentiable Trajectory Reweighting

Top-down potential optimization aims to match the *K* outputs of a molecular mechanics simulation ** O** to experimental observables \(\tilde{{{{{{{{\boldsymbol{O}}}}}}}}}\). Therefore, the objective is to minimize a loss function

*L*(

**), e.g., a mean-squared error (MSE)**

*θ*where 〈〉 denotes the ensemble average, and 〈*O*_{k}(*U*_{θ})〉 depends on the potential energy *U*_{θ} parametrized by ** θ**. We will focus on the case where a MD simulation approximates 〈

*O*

_{k}(

*U*

_{θ})〉—with Monte Carlo

^{41}being a usable alternative. With standard assumptions on ergodicity and thermodynamic equilibrium, the ensemble average 〈

*O*

_{k}(

*U*

_{θ})〉 is approximated via a time average

where \({\{{{{{{{{{\bf{S}}}}}}}}}_{i}\}}_{i = 1}^{N}\) is the trajectory of the system, i.e., a sequence of *N* states consisting of particle positions and momenta. Due to the small time step size necessary to maintain numerical stability in MD simulations, states are highly correlated. Subsampling, i.e., only averaging over every 100th or 1000th state, reduces this correlation in Eq. (2).

As the generated trajectory depends on ** θ**, every update of

**during training would require a re-computation of the entire trajectory. However, by leveraging thermodynamic perturbation theory**

*θ*^{42}, it is possible to re-use decorrelated states obtained via a reference potential \(\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}\). Specifically, the time average is reweighted to account for the altered state probabilities

*p*

_{θ}(

**S**

_{i}) from the perturbed potential

*θ*^{35,36,42}:

Assuming a canonical ensemble, state probabilities follow the Boltzmann distribution \({p}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}({{{{{{{{\bf{S}}}}}}}}}_{i}) \sim {e}^{-\beta H({{{{{{{{\bf{S}}}}}}}}}_{i})}\), where *H*(**S**_{i}) is the Hamiltonian of the state (sum of potential and kinetic energy), *β* = 1/(*k*_{B}*T*), *k*_{B} Boltzmann constant, *T* temperature. Inserting *p*_{θ}(**S**_{i}) into Eq. (3) allows computing weights as a function of ** θ** (the kinetic energy cancels)

For the special case of \({{{{{{{\boldsymbol{\theta }}}}}}}}=\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}\), *w*_{i} = 1/*N*, recovering Eq. (2). Note that similar expressions to Eq. (4) could be derived for other ensembles, e.g., the isothermal–isobaric ensemble, via respective state probabilities *p*_{θ}(**S**_{i}). In practice, the reweighting ansatz is only applicable given small potential energy differences. For large differences between ** θ** and \(\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}\), by contrast, few states dominate the average. In this case, the effective sample size

^{37}

is reduced and the statistical error in 〈*O*_{k}(*U*_{θ})〉 increases (Eq. (3)).

Reweighting can be exploited for two purposes that are linked to speedups in the forward and backward pass, respectively: first, reweighting reduces computational effort as decorrelated states from previous trajectories can often be re-used^{37}. Second, and most importantly, reweighting establishes a direct functional relation between 〈*O*_{k}(*U*_{θ})〉 and ** θ**. This relation via

**w**provides an alternative end-to-end differentiable path for computing the gradient of the loss ∇

_{θ}

*L*: differentiating through the reweighting scheme replaces the backward pass through the simulation. Leveraging this alternative differentiation path, while managing the effective sample size

*N*

_{eff}, are the central ideas behind the DiffTRe method.

The workflow of the DiffTRe algorithm consists of the following steps: first, an initial reference trajectory is generated from the canonical ensemble, e.g., via a stochastic or deterministic thermostat, from an initial state **S**_{init} and reference potential \(\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}\) (Fig. 1a). Initial equilibration states are disregarded and the following states are subsampled yielding decorrelated states \({\{{{{{{{{{\bf{S}}}}}}}}}_{i}\}}_{i = 1}^{N}\). Together with their reference potential energies \({\{{U}_{\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}}({{{{{{{{\bf{S}}}}}}}}}_{i})\}}_{i = 1}^{N}\), these states are saved for re-use during reweighting. In the next step, the reweighting scheme is employed to compute ∇_{θ}*L* with respect to current parameters ** θ**, where initially \({{{{{{{\boldsymbol{\theta }}}}}}}}=\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}\). An optimizer subsequently uses ∇

_{θ}

*L*to improve

**. This procedure of reweighting, gradient computation and updating is repeated as long as the statistical error from reweighting is acceptably small, i.e.,**

*θ**N*

_{eff}is larger than a predefined \({\bar{N}}_{{{{{{{{\rm{eff}}}}}}}}}\). As soon as \({N}_{{{{{{{{\rm{eff}}}}}}}}} \, < \, {\bar{N}}_{{{{{{{{\rm{eff}}}}}}}}}\), a new reference trajectory needs to be sampled using the current

**as the new \(\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}\). At least one**

*θ***update per reference trajectory is ensured because initially**

*θ**N*

_{eff}=

*N*. Using the last generated state

**S**

_{N}as

**S**

_{init}for the next trajectory counteracts overfitting to a specific initial configuration. In addition, \({p}_{\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}}({{{{{{{{\bf{S}}}}}}}}}_{{{{{{{{\rm{init}}}}}}}}})\) is reasonably high when assuming small update steps, reducing necessary equilibration time for trajectory generation. Saving only \({\{{{{{{{{{\bf{S}}}}}}}}}_{i}\}}_{i = 1}^{N}\) and \({\{{U}_{\hat{{{{{{{{\boldsymbol{\theta }}}}}}}}}}({{{{{{{{\bf{S}}}}}}}}}_{i})\}}_{i = 1}^{N}\) from the simulation entails low-memory requirements similar to the adjoint method. DiffTRe assumes that deviations in predicted observables are attributable to an inaccurate potential

*U*

_{θ}rather than a statistical sampling error. Accordingly,

*N*and the subsampling ratio

*n*need to be chosen to yield a sufficiently small statistical error. Optimal values for

*N*and

*n*depend on the specific system, target observables, and the thermodynamic-state point.

Computation of ∇_{θ}*L* via reverse-mode AD through the reweighting scheme comprises a forward pass starting with computation of the potential *U*_{θ}(**S**_{i}) and weight *w*_{i} for each **S**_{i} (Eq. (4); Fig. 1a). Afterward, reweighted observables 〈*O*_{k}(*U*_{θ})〉 (Eq. (3)) and the resulting loss *L*(** θ**) (Eq. (1)) are calculated. The corresponding backward pass starts at

*L*(

**) and stops at parameters**

*θ***in the potential energy computation**

*θ**U*

_{θ}(

**S**

_{i}). The differentiation path defined by the reweighting ansatz is therefore independent of the trajectory generation.

Evaluation of *U*_{θ}(**S**_{i}) (Fig. 1b) involves computing the pairwise distance matrix **D** from atom positions of **S**_{i}, that are fed into a learnable potential \({U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}^{{{{{{{{\rm{model}}}}}}}}}\) and a prior potential *U*^{prior}. Both potential components are combined by adding the predicted potential energies

In subsequent examples of diamond and CG water, \({U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}^{{{{{{{{\rm{model}}}}}}}}}\) is a graph neural network operating iteratively on the atomic graph defined by **D**. *U*^{prior} is a constant potential approximating a priori-known properties of the system, such as the Pauli exclusion principle (e.g., Eq. (12)). Augmenting NN potentials with a prior is common in the bottom-up coarse-graining literature^{8,10} to provide qualitatively correct behavior in regions of the potential energy surface (PES) not contained in the dataset, but reachable by the CG model. By contrast, DiffTRe does not rely on pre-computed data sets. Rather, the prior serves to control the data (trajectory) generation in the beginning of the optimization. In addition, *U*^{prior} reformulates the problem from learning \({U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}^{{{{{{{{\rm{model}}}}}}}}}\) directly to learning the difference between *U*^{prior} and the optimal potential given the data^{10}. A well-chosen *U*^{prior} therefore represents a physics-informed initialization accelerating training convergence. Suitable *U*^{prior} can often be found in the literature: Classical force fields such as AMBER^{25} and MARTINI^{27} define reasonable interactions for bio-molecules and variants of the Embedded Atom Model^{43} (EAM) provide potentials for metals and alloys. Note that *U*^{prior} is not a prior in the Bayesian sense providing a pervasive bias on learnable parameters in the small data regime. If *U*^{prior} is in contradiction with the data, \({U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}^{{{{{{{{\rm{model}}}}}}}}}\) will correct for *U*^{prior} as a result of the optimization. In the next section, we further illustrate for a toy problem the interplay between prior, gradients and the learning process in DiffTRe, and provide a comparison to direct reverse-mode AD through the simulation.

### Double-well toy example

We consider ideal gas particles at a temperature *k*_{B}*T* = 1 trapped inside a one-dimensional double-well potential (Fig. 2a) parametrized by

The goal is to learn ** θ** such that \({U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}(x)={U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}^{{{{{{{{\rm{model}}}}}}}}}(x)+{U}^{{{{{{{{\rm{prior}}}}}}}}}(x)\) matches

*U*(

*x*). We select a cubic spline as \({U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}^{{{{{{{{\rm{model}}}}}}}}}\), which acts as a flexible approximator for twice continuously differentiable functions. The cubic spline is parametrized via the potential energy values of 50 control points \({\{{x}_{j},{U}_{j}\}}_{j = 1}^{50}\) evenly distributed over

*x*∈ [0, 1]. Analogous to NN potentials in subsequent problems, we randomly initialize \({U}_{j} \sim {{{{{{{\mathcal{N}}}}}}}}(0,0.0{1}^{2}{k}_{B}T)\). Initializing

*U*

_{j}= 0 leads to largely identical results in this toy problem. The harmonic single-well potential

*U*

^{prior}(

*x*) =

*λ*(

*x*− 0.5)

^{2}, with scale

*λ*= 75, encodes the prior knowledge that particles cannot escape the double-well. We choose the normalized density profile

*ρ*(

*x*)/

*ρ*

_{0}of ideal gas particles as the target observable. The resulting loss function is

where *ρ*(*x*) is discretized via *K* bins. 〈*ρ*(*x*_{k})〉 are approximated based on *N* = 10,000 states after skipping 1000 states for equilibration, where a state is retained every 100 time steps. We minimize Eq. (8) via an Adam^{44} optimizer with learning rate decay. For additional DiffTRe and simulation parameters, see Supplementary Method 1.1.

Initially, *ρ*/*ρ*_{0} resulting from *U*^{prior}(*x*) deviates strongly from the target double-well density (Fig. 2b). The loss curve illustrates successful optimization over 200 update steps (Fig. 2c). The wall-clock time per parameter update Δ*t* clearly shows two distinct levels: at the start of the optimization, update steps are rather large, significantly reducing *N*_{eff}. Hence a new reference trajectory generation is triggered with each update (average Δ*t* ≈ 39.2 s). Over the course of the simulation, updates of \({U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}^{{{{{{{{\rm{model}}}}}}}}}(x)\) become smaller and reference trajectories are occasionally re-used (average Δ*t* ≈ 2.76 s). After optimization, the target density is matched well. The learned potential energy function *U*_{θ}(*x*) recovers the data-generating potential *U*(*x*) (Supplementary Fig. 1a); thus, other thermodynamic and kinetic observables will match reference values closely. However, this conclusion does not apply in realistic applications, where learned potentials are in general not unique^{2} due to the limited number of target observables that can be considered in practice.

The effect of *U*^{prior} on the training process is twofold: First, by encoding prior knowledge, it simplifies convergence, as \({U}_{{{{{{{{\boldsymbol{\theta }}}}}}}}}^{{{{{{{{\rm{model}}}}}}}}}(x)\) only needs to adapt the single-well prior instead of learning large energy barriers from scratch. Second, *U*^{prior} also impacts the information content of the gradient by controlling the generation of trajectories in the beginning of the optimization (Eq. (15)). The local support of the cubic spline allows analyzing this relation empirically (Supplementary Fig. 2): The gradient is nonzero only in regions of the PES that are included in the reference trajectory. Hence, other regions of the PES are not optimized despite delivering a nonzero contribution to the loss. A well-chosen prior potential should therefore yield trajectories that are as close as possible to trajectories sampled from the true potential. However, satisfactory learning results can be obtained for a sensible range of prior scales (Supplementary Fig. 3).

We study the robustness of our results by varying the random seed that controls the initialization of the spline as well as the initial particle positions and velocities. Results from the variation study in Supplementary Fig. 4 demonstrate that the predicted *ρ*(*x*)/*ρ*_{0} is robust to the random initialization. The corresponding *U*_{θ}(*x*) exhibits some variance at the left well boundary, mirroring difficult training in this region due to vanishing gradients for vanishing predicted densities (Supplementary Fig. 2) and minor influence of the exact wall position on the resulting density profile (Supplementary Fig. 4a).

For comparison, we have implemented gradient computation via direct reverse-mode AD through the simulation. This approach clearly suffers from the exploding gradients problem (Fig. 2d): The gradient magnitude increases exponentially as a function of the simulation length. Without additional modifications (e.g., as implemented by Ingraham et al.^{31}), these gradients are impractical for longer trajectories. By contrast, gradients computed via DiffTRe show constant magnitudes irrespective of the simulation length.

To measure the speed-up over direct reverse-mode AD empirically, we simulate the realistic case of an expensive potential by substituting the numerically inexpensive spline with a fully connected neural network with two hidden layers and 100 neurons each. We measure speedups of *s*_{g} = 486 for gradient computations and *s* = 3.7 as overall speed-up per update when a new reference trajectory is sampled. However, these values are rather sensitive to the exact computational and simulation setup. Memory overflow in the direct AD method constrained trajectory lengths to ten retained states and a single state for equilibration (a total of 1100 time steps). Measuring speed-up for one of the real-world problems below would be desirable, but is prevented by the memory requirements of direct AD.

The measured speed-up values are in line with theoretical considerations: While direct AD backpropagates through the whole trajectory generation, DiffTRe only differentiates through the potential energy computation of decorrelated states \({\{{{{{{{{{\bf{S}}}}}}}}}_{i}\}}_{i = 1}^{N}\) (Fig. 1). From this algorithmic difference, we expect speed-up values that depend on the subsampling ratio *n*, the number of skipped states during equilibration *N*_{equilib} and the cost multiple of backward passes with respect to forward passes *G* (details in Supplementary Method 2)

For this toy example setup, the rule-of-thumb estimates in Eq. (9) yield *s*_{g} = 330 and *s* = 4, agreeing with the measured values. In the next sections, we showcase the effectiveness of DiffTRe in real-world, top-down learning of NN potentials.

### Atomistic model of diamond

To demonstrate the applicability of DiffTRe to solids on the atomistic scale, we learn a DimeNet++^{12} potential for diamond from its experimental elastic stiffness tensor **C**. Due to symmetries in the diamond cubic crystal, **C** only consists of three distinct stiffness moduli \({\tilde{C}}_{11}=1079\) GPa, \({\tilde{C}}_{12}=124\) GPa and \({\tilde{C}}_{44}=578\) GPa^{45} (in Voigt notation). In addition, we assume the crystal to be in a stress-free state ** σ** =

**0**for vanishing infinitesimal strain

**=**

*ϵ***0**. These experimental data define the loss

where loss weights *γ*_{σ} and *γ*_{C} counteract the effect of different orders of magnitude of observables. To demonstrate learning, we select the original Stillinger–Weber potential^{46} parametrized for silicon as *U*^{prior}. We have adjusted the length and energy scales to *σ*_{SW} = 0.14 nm and *ϵ*_{SW} = 200 kJ/mol, reflecting the smaller size of carbon atoms. We found learning to be somewhat sensitive to *U*^{prior} in this example because weak prior choices can lead to unstable MD simulations. Simulations are run with a cubic box of size *L* ≈ 1.784 nm containing 1000 carbon atoms (Fig. 3a) to match the experimental density (*ρ* = 3512 kg/m^{3})^{45} exactly. The temperature in the experiment (*T* = 298.15 K^{45}) determines the simulation temperature. Each trajectory generation starts with 10 ps of equilibration followed by 60 ps of production, where a decorrelated state is saved every 25 fs. We found these trajectories to yield observables with acceptably small statistical noise. The stress tensor ** σ** is computed via Eq. (13) and the stiffness tensor

**C**via the stress fluctuation method (Eq. (14)). Further details are summarized in Supplementary Method 4.

Figure 3 visualizes convergence of the stress (b) and stiffness components (c). Given that the model is only trained on rather short trajectories, we test the trained model on a trajectory of 10 ns length to ensure that the model neither overfitted to initial conditions nor drifts away from the targets. The resulting stress and stiffness values *σ*_{1} = 0.29 GPa, *σ*_{4} = 0.005 GPa, *C*_{11} = 1070 GPa, *C*_{12} = 114 GPa, and *C*_{44} = 560 GPa are in good agreement with respective targets. These results could be improved by increasing the trajectory length, which reduces statistical sampling errors. The corresponding inverse stress–strain relation is given by the compliance tensor **S** = **C**^{−1}, which can be constructed from by Young’s modulus *E* = 1047 GPa, shear modulus *G* = 560 GPa, and Poisson’s ratio *ν* = 0.097. The training loss curve and wall-clock time per update Δ*t* are displayed in Supplementary Fig. 5a.

Computing the stress–strain curve (Supplementary Fig. 5b) from the trained model in the linear regime (*ϵ*_{i} < 0.005) verifies that computing **C** via Eq. (14) yields the same result as explicitly straining the box and measuring stresses. In addition, this demonstrates that the DimeNet++ potential generalizes from the training box (** ϵ** = 0) to boxes under small strain. We also strained the box beyond the linear regime, creating a distribution shift

^{19,20}, to test generalization to unobserved state points. The predicted stress–strain curve in Fig. 3d shows good agreement with DFT data

^{47}for medium-sized natural strains \({e}_{1}={{{{{{\mathrm{log}}}}}}}\,(1+{\epsilon }_{1}) \, < \, 0.02\). For large strains, the deviation quickly increases, including an early fracture. These incorrect predictions of the learned potential are due to limited extrapolation capacities of NN potentials: states under large strain are never encountered during training, leading to large uncertainty in predicted forces. Incorporating additional observables linked to states of large strain into the optimization, such as the point of maximum stress, should improve predictions.

To test the trained DimeNet++ potential on held-out observables, we compute the phonon density of states (PDOS). The predicted PDOS deviates from the experiment^{48}, analogous to a Stillinger–Weber potential optimized for diamond^{49} (Supplementary Fig. 5c). The evolution of the predicted PDOS over the course of the optimization is shown in Supplementary Fig. 5d. Deviations of held-out observables are expected given that top-down approaches allow learning potentials that are consistent with target experimental observables but lack theoretical convergence guarantees of bottom-up schemes (in the limit of a sufficiently large data set and a sufficiently expressive model)^{2}. In principle, we expect sufficiently expressive top-down models to converge to the true potential in the limit of an infinite number of matched target observables. In practice, however, many different potentials can reproduce a sparse set of considered target observables, rendering the learned potential non-unique^{2}. In this particular example, we show that many different potentials can reproduce the target stress and stiffness, but predict different PDOSs: While predicted stress and stiffness values are robust to random initialization of NN weights and initial particle velocities within the statistical sampling error, the corresponding predicted PDOSs vary to a great extent (Supplementary Fig. 6). Incorporating additional observables more closely connected to phonon properties into the loss function could improve the predicted PDOS.

### Coarse-grained water model

Finally, we learn a DimeNet++ potential for CG water. Water is a common benchmark problem due to its relevance in bio-physics simulations and its pronounced 3-body interactions, which are challenging for classical potentials^{50}. We select a CG-mapping, where each CG particle is centered at the oxygen atom of the corresponding atomistic water molecule (Fig. 4a). This allows using experimental oxygen–oxygen radial (RDF) and angular distribution functions (ADF) as target observables. Given that the reference experiment^{51} was carried out at ambient conditions (*T* = 296.15 K), we can additionally target a pressure \(\tilde{p}=1\) bar. Hence, we minimize

As the prior potential, we select the repulsive term of the Lennard–Jones potential

Drawing inspiration from atomistic water models, we have chosen the length scale of the SPC^{52} water model as *σ*_{R} = 0.3165 nm as well as a reduced energy scale of *ϵ*_{R} = 1 kJ/mol to counteract the missing Lennard–Jones attraction term in Eq. (12). We build a cubic box of length 3 nm with 901 CG particles, implying a density of *ρ* = 998.28 g/l, to match the experimental water density of *ρ* = 997.87 g/l at 1 bar. Trajectory generation consists of 10 ps of equilibration and 60 ps of subsequent production, where a decorrelated state is saved every 0.1 ps. For additional details, see Supplementary Method 2.3.

Figure 4b–d displays properties predicted by the final trained model during a 10 ns production run: DiffTRe is able to train a DimeNet++ potential that simultaneously matches experimental oxygen RDF, ADF, and pressure to the line thickness. The evolution of predicted RDFs and ADFs as well as the loss and wall-clock times per update are displayed in Supplementary Fig. 7a–c. The learning process is robust to weak choices of *U*^{prior}: DiffTRe is able to converge to the same prediction quality as with the reference prior even if *σ*_{R} is misestimated by ±0.1 nm (approximately ±30%) compared to the classical SPC water model (Supplementary Fig. 8a, b). This represents a large variation given that within common atomistic water models, *σ*_{R} varies by <0.5%^{53}.

To test the learned potential on held-out observables, we compute the tetrahedral order parameter *q*^{54} and the self-diffusion coefficient *D*. *q* ≈ 0.569 matches the experimental value of \(\tilde{q}=0.576\) closely. This is expected as *q* considers the structure of four nearest neighbor particles, which is closely related to the ADF. The learned CG water model predicts a larger self-diffusion coefficient than were experimentally measured (*D* = 10.91 μm^{2}/ms vs. \(\tilde{D}=2.2\,\upmu {{{{{{{{\rm{m}}}}}}}}}^{2}/{{{{{{{\rm{ms}}}}}}}}\))^{55}. With the same simulation setup, a single-site tabulated potential parametrized via iterative Boltzmann inversion^{39} with pressure correction^{39,56} predicts *D* = 14.15 μm^{2}/ms. These results are in line with the literature: Due to smoother PESs, CG models exhibit accelerated dynamical processes compared to atomistic models^{2}. For CG water models specifically, diffusion coefficients decrease with increasing number of interaction sites^{57}. In this context, the decreasing diffusion coefficients over the course of the optimization (Supplementary Fig. 7d) could indicate that *U*_{θ} acts effectively as a single-site model in the beginning, while learning 3-body interactions during the optimization casts *U*_{θ} more similar to multisite CG models. Obtained results are robust to random initialization of NN weights and initial particle velocities, both for predicted target (Supplementary Fig. 8c, d) and held-out observables (*D* = 10.93 ± 0.20 μm^{2}/ms).

The accuracy of predicted 2 and 3-body interactions (Fig. 4c, d) showcases the potency of graph neural network potentials in top-down molecular modeling: capturing 3-body interactions is essential for modeling water given that pair potentials trained via force matching fail to reproduce both RDF and ADF of the underlying high-fidelity model^{50}. Other top-down CG water models with simple functional form tend to deviate from the experimental RDF^{58,59}. Deviations from experimental structural properties, albeit smaller in size, also arise in DFT simulations^{22,60}, limiting the accuracy of bottom-up trained NN potentials^{8}.

## Discussion

In this work, we demonstrate numerically efficient learning of NN potentials from experimental data. The main advantages of our proposed DiffTRe method are its flexibility and simplicity: DiffTRe is applicable to solid and fluid materials, coarse-grained and atomistic models, thermodynamic, structural and mechanical properties, as well as potentials of arbitrary functional form. To apply DiffTRe, practitioners only need to set up a MD simulation with corresponding observables and a loss function, while gradients are computed conveniently in an end-to-end fashion via AD. The demonstrated speedups and limited memory requirements promote application to larger systems.

Without further adaptations, DiffTRe can also be applied as a bottom-up model parametrization scheme. In this case, a high-fidelity simulation, rather than an experiment, provides target observables. For CG models, DiffTRe generalizes structural coarse-graining schemes such as iterative Boltzmann inversion^{39} or Inverse Monte Carlo^{40}. DiffTRe overcomes the main limitations of these approaches: First, structural coarse-graining is no longer restricted to one-dimensional potentials, and matching many-body correlation functions (e.g., ADFs) is therefore feasible. Second, the user can integrate additional observables into the optimization without relying on hand-crafted iterative update rules, for instance for pressure-matching^{39,56}. This is particularly useful if an observable needs to be matched precisely (e.g., pressure in certain multiscale simulations^{61}). Matching many-body correlation functions will likely allow structural bottom-up coarse-graining to take on significance within the new paradigm of many-body CG potentials^{8,9,10}.

For the practical application of DiffTRe, a few limitations need to be considered. The reweighting scheme renders DiffTRe invariant to the sequence of states in the trajectory. Hence, dynamical properties cannot be employed as target observables. In addition, the NN potential test cases considered in this work required a reasonably chosen prior potential. Lastly, two distinct sources of overfitting when learning from experimental data for a single system need to be accounted for^{1}: To avoid overfitting to a specific initial state, DiffTRe uses a different initial state for each reference trajectory. Moreover, increasing the system size and trajectory length ensures representative reference trajectories. Irrespective of overfitting, generalization to different systems, observables, and thermodynamic-state points remains to be addressed, for instance via training on multi-systemic experimental data sets. To this end, an in-depth assessment of out-of-sample properties of top-down learned NN potentials is required.

From a machine learning (ML) perspective, DiffTRe belongs to the class of end-to-end differentiable physics approaches^{62,63,64}. These approaches are similar to reinforcement learning in that the target outcome of a process (here a MD simulation) represents the data. A key difference is the availability of gradients through the process, allowing for efficient training. Differentiable physics approaches, increasingly popular in control applications^{34,65,66,67}, enable direct training of the ML model via the physics simulator, advancing the ongoing synthesis of ML and physics-based methods.

Finally, the combination of bottom-up and top-down approaches for learning NN potentials, i.e., considering information from both the quantum and macroscopic scale, represents an exciting avenue for future research. For top-down approaches, pre-training NN potentials on bottom-up data sets can serve as a sensible extrapolation for the PES in areas unconstrained by the experimental data. In DiffTRe, a pre-trained model could also circumvent the need for a prior potential. Bottom-up trained NN potentials, on the other hand, can be enriched with experimental data, which enables targeted refinement of the potential. This is particularly helpful for systems in which DFT accuracy is insufficient or the generation of a quantum mechanical data set is computationally intractable.

## Methods

### Differentiable histogram binning

To obtain an informative gradient \(\frac{\partial L}{\partial {{{{{{{\boldsymbol{\theta }}}}}}}}}\), predicted observables need to be continuously differentiable. However, many common observables in MD, including density and structural correlation functions, are computed by discrete histogram binning. To obtain a differentiable observable, the (discrete) Dirac function used in binning can be approximated by a narrow Gaussian probability density function (PDF)^{34}. Similarly, we smooth the non-differentiable cutoff in the definition of ADFs via a Gaussian cumulative distribution function (CDF) centered at the cutoff (details on differentiable density, RDF, and ADF in Supplementary Method 3).

### Stress–strain relations

Computing the virial stress tensor *σ*^{V} for many-body potentials, e.g., NN potentials, under periodic boundary conditions requires special attention. This is due to the fact that most commonly used formulas are only valid for non-periodic boundary conditions or pairwise potentials^{68}. Therefore, we resort to the formulation proposed by Chen et al.^{69}, which is well suited for vectorized computations in NN potentials.

where *N*_{p} is the number of particles, ⊗ represents the dyadic or outer product, *m*_{k} and **v**_{k} are mass and thermal excitation velocity of particle *k*, **R** and **F** are (*N*_{p} × 3) matrices containing all particle positions and corresponding forces, **h** is the (3 × 3) lattice tensor spanning the simulation box, and \({{\Omega }}=\det ({{{{{{{\bf{h}}}}}}}})\) is the box volume.

Due to the equivalence of the ensemble-averaged virial stress tensor 〈*σ*^{V}〉 and the Cauchy stress tensor *σ*^{70}, we can compute the elastic stiffness tensor from MD simulations and compare it to continuum mechanical experimental data (details in Supplementary Method 5). In the canonical ensemble, the isothermal elastic stiffness tensor **C** can be calculated at constant strain ** ϵ** via the stress fluctuation method

^{71}:

with the Born contribution to the stress tensor \({\sigma }_{ij}^{B}=\frac{1}{{{\Omega }}}\frac{\partial U}{\partial {\epsilon }_{ij}}\), the Born contribution to the stiffness tensor \({C}_{ijkl}^{B}=\frac{1}{{{\Omega }}}\frac{{\partial }^{2}U}{\partial {\epsilon }_{ij}\partial {\epsilon }_{kl}}\) and Kronecker delta *δ*_{ij}. Eq. (14) integrates well into DiffTRe by reweighting individual ensemble average terms (Eq. (3)) and combining the reweighted averages afterwards. Implementing the stress fluctuation method in differentiable MD simulations is straightforward: AD circumvents manual derivation of strain-derivatives, which is non-trivial for many-body potentials^{72}.

### Statistical mechanics foundations

Thermodynamic fluctuation formulas allow to compute the gradient \(\frac{\partial L}{\partial {{{{{{{\boldsymbol{\theta }}}}}}}}}\) from ensemble averages^{73,74,75}. Specifically, considering a MSE loss for a single observable *O*(*U*_{θ}) in the canonical ensemble^{73},

It can be seen that the AD routine in DiffTRe estimates \(\frac{\partial L}{\partial {{{{{{{\boldsymbol{\theta }}}}}}}}}\) by approximating ensemble averages in Eq. (15) via reweighting averages (Derivation in Supplementary Method 5). End-to-end differentiation through the reweighting scheme simplifies optimization by combining obtained gradients from multiple observables. This is particularly convenient for observables that are not merely averages of instantaneous quantities, e.g., the stiffness tensor **C** (Eq. (14)).

### DimeNet++

We employ a custom implementation of DimeNet++^{11,12} that fully integrates into Jax MD^{29}. Our implementation takes advantage of neighbor lists for efficient computation of the sparse atomic graph. We select the same NN hyperparameters as in the original publication^{12} except for the embedding sizes, which we reduced by factor 4. This modification allowed for a significant speed-up while retaining sufficient capacity for the problems considered in this work. For diamond, we have reduced the cutoff to 0.2 nm yielding an atomic graph, where each carbon atom is connected to its four covalently bonded neighbors. A comprehensive list of employed DimeNet++ hyperparameters is provided in Supplementary Method 6.

## Data availability

Simulation setups and trained DimeNet++ models have been deposited in https://github.com/tummfm/difftre. The data generated in this study are provided in the paper or in the Supplementary information file.

## Code availability

The code for DiffTRe and its application to the three test cases is available at https://github.com/tummfm/difftre^{76}.

## References

- 1.
Fröhlking, T., Bernetti, M., Calonaci, N. & Bussi, G. Toward empirical force fields that match experimental observables.

*J. Chem. Phys.***152**, 230902 (2020). - 2.
Noid, W. G. Perspective: coarse-grained models for biomolecular systems.

*J. Chem. Phys.***139**, 090901 (2013). - 3.
Schütt, K. T., Arbabzadah, F., Chmiela, S., Müller, K. R. & Tkatchenko, A. Quantum-chemical insights from deep tensor neural networks.

*Nat. Commun.***8**, 13890 (2017). - 4.
Noé, F., Tkatchenko, A., Müller, K. R. & Clementi, C. Machine learning for molecular simulation.

*Annu. Rev. Phys. Chem.***71**, 361–390 (2020). - 5.
Behler, J. & Parrinello, M. Generalized neural-network representation of high-dimensional potential-energy surfaces.

*Phys. Rev. Lett.***98**, 146401 (2007). - 6.
Schütt, K. T. et al. SchNet: a continuous-filter convolutional neural network for modeling quantum interactions. in

*Advances in Neural Information Processing Systems*Vol. 30, 992–1002 (Curran Associates, Inc., 2017). - 7.
Gilmer, J., Schoenholz, S. S., Riley, P. F., Vinyals, O. & Dahl, G. E. Neural message passing for quantum chemistry. in

*Proceedings of the 34th International Conference on Machine Learning*. 1263–1272 (PMLR, 2017). - 8.
Zhang, L., Han, J., Wang, H., Car, R. & Weinan, W. E. DeePCG: constructing coarse-grained models via deep neural networks.

*J. Chem. Phys.***149**, 034101 (2018). - 9.
Wang, J. et al. Machine learning of coarse-grained molecular dynamics force fields.

*ACS Cent. Sci.***5**, 755–767 (2019). - 10.
Husic, B. E. et al. Coarse graining molecular dynamics with graph neural networks.

*J. Chem. Phys.***153**, 194101 (2020). - 11.
Klicpera, J., Groß, J. & Günnemann, S. Directional message passing for molecular graphs. In

*8th International Conference on Learning Representations, ICLR*(2020). - 12.
Klicpera, J., Giri, S., Margraf, J. T. & Günnemann, S. Fast and uncertainty-aware directional message passing for non-equilibrium molecules. in

*Machine Learning for Molecules Workshop at NeurIPS*(2020). - 13.
Qiao, Z., Welborn, M., Anandkumar, A., Manby, F. R. & Miller, T. F. OrbNet: deep learning for quantum chemistry using symmetry-adapted atomic-orbital features.

*J. Chem. Phys.***153**, 124111 (2020). - 14.
Vlachas, P. R., Zavadlav, J., Praprotnik, M. & Koumoutsakos, P. Accelerated simulations of molecular systems through learning of their effective dynamics. Preprint at https://arxiv.org/abs/2011.14115 (2021).

- 15.
Jain, A. C. P., Marchand, D., Glensk, A., Ceriotti, M. & Curtin, W. A. Machine learning for metallurgy III: a neural network potential for Al-Mg-Si.

*Phys. Rev. Mater.***5**, 053805 (2021). - 16.
Ko, T. W., Finkler, J. A., Goedecker, S. & Behler, J. A fourth-generation high-dimensional neural network potential with accurate electrostatics including non-local charge transfer.

*Nat. Commun.***12**, 398 (2021). - 17.
Ramakrishnan, R., Dral, P. O., Rupp, M. & Von Lilienfeld, O. A. Quantum chemistry structures and properties of 134 kilo molecules.

*Sci. Data***1**, 1–7 (2014). - 18.
Faber, F. A. et al. Prediction errors of molecular machine learning models lower than hybrid DFT error.

*J. Chem. Theory Comput.***13**, 5255–5264 (2017). - 19.
Cubuk, E. D. & Schoenholz, S. S. Adversarial forces of physical models. in

*3rd NeurIPS workshop on Machine Learning and the Physical Sciences*(2020). - 20.
Schwalbe-Koda, D., Tan, A. R. & Gómez-Bombarelli, R. Differentiable sampling of molecular geometries with uncertainty-based adversarial attacks.

*Nat. Commun.***12**, 5104 (2021). - 21.
Zhang, L., Lin, D.-Y., Wang, H., Car, R. & Weinan, E. Active learning of uniformly accurate interatomic potentials for materials simulation.

*Phys. Rev. Mater.***3**, 023804 (2019). - 22.
Gillan, M. J., Alfè, D. & Michaelides, A. Perspective: how good is DFT for water?

*J. Chem. Phys.***144**, 130901 (2016). - 23.
Smith, J. S. et al. Approaching coupled cluster accuracy with a general-purpose neural network potential through transfer learning.

*Nat. Commun.***10**, 2930 (2019). - 24.
Sauceda, H. E., Vassilev-Galindo, V., Chmiela, S., Müller, K. R. & Tkatchenko, A. Dynamical strengthening of covalent and non-covalent molecular interactions by nuclear quantum effects at finite temperature.

*Nat. Commun.***12**, 442 (2021). - 25.
Cornell, W. D. et al. A second generation force field for the simulation of proteins, nucleic acids, and organic molecules.

*J. Am. Chem. Soc.***117**, 5179–5197 (1995). - 26.
Oostenbrink, C., Villa, A., Mark, A. E. & Van Gunsteren, W. F. A biomolecular force field based on the free enthalpy of hydration and solvation: the GROMOS force-field parameter sets 53A5 and 53A6.

*J. Comput. Chem.***25**, 1656–1676 (2004). - 27.
Marrink, S. J., Risselada, H. J., Yefimov, S., Tieleman, D. P. & De Vries, A. H. The MARTINI force field: coarse grained model for biomolecular simulations.

*J. Phys. Chem. B***111**, 7812–7824 (2007). - 28.
Baydin, A. G., Pearlmutter, B. A., Radul, A. A. & Siskind, J. M. Automatic differentiation in machine learning: a survey.

*J. Mach. Learn. Res.***18**, 1–43 (2018). - 29.
Schoenholz, S. S. & Cubuk, E. D. JAX MD: A Framework for Differentiable Physics. in

*Advances in Neural Information Processing Systems*Vol. 33, 11428–11441 (Curran Associates, Inc., 2020). - 30.
Doerr, S. et al. Torchmd: a deep learning framework for molecular simulations.

*J. Chem. Theory Comput.***17**, 2355–2363 (2021). - 31.
Ingraham, J., Riesselman, A., Sander, C. & Marks, D. Learning protein structure with a differentiable simulator. in

*7th International Conference on Learning Representations, ICLR*(2019). - 32.
Goodrich, C. P., King, E. M., Schoenholz, S. S., Cubuk, E. D. & Brenner, M. P. Designing self-assembling kinetics with differentiable statistical physics models.

*Proc. Natl Acad. Sci. USA***118**, e2024083118 (2021). - 33.
Chen, R. T. Q., Rubanova, Y., Bettencourt, J. & Duvenaud, D. Neural ordinary differential equations. in

*Advances in Neural Information Processing Systems*Vol. 31 (Curran Associates, Inc., 2018). - 34.
Wang, W., Axelrod, S. & Gómez-Bombarelli, R. Differentiable molecular simulations for control and learning. in

*ICLR 2020 Workshop on Integration of Deep Neural Models and Differential Equations.*(2020). - 35.
Norgaard, A. B., Ferkinghoff-Borg, J. & Lindorff-Larsen, K. Experimental parameterization of an energy function for the simulation of unfolded proteins.

*Biophys. J.***94**, 182–192 (2008). - 36.
Li, D. W. & Brüschweiler, R. Iterative optimization of molecular mechanics force fields from NMR data of full-length proteins.

*J. Chem. Theory Comput.***7**, 1773–1782 (2011). - 37.
Carmichael, S. P. & Shell, M. S. A new multiscale algorithm and its application to coarse-grained peptide models for self-assembly.

*J. Phys. Chem. B***116**, 8383–8393 (2012). - 38.
Wang, L. P., Chen, J. & Van Voorhis, T. Systematic parametrization of polarizable force fields from quantum chemistry data.

*J. Chem. Theory Comput.***9**, 452–460 (2013). - 39.
Reith, D., Pütz, M. & Müller-Plathe, F. Deriving effective mesoscale potentials from atomistic simulations.

*J. Comput. Chem.***24**, 1624–1636 (2003). - 40.
Lyubartsev, A. P. & Laaksonen, A. Calculation of effective interaction potentials from radial distribution functions: a reverse Monte Carlo approach.

*Phys. Rev. E***52**, 3730–3737 (1995). - 41.
Binder, K., Heermann, D., Roelofs, L., Mallinckrodt, A. J. & McKay, S. Monte Carlo simulation in statistical physics.

*Comput. Phys.***7**, 156 (1993). - 42.
Zwanzig, R. W. High-temperature equation of state by a perturbation method. I. Nonpolar gases.

*J. Chem. Phys.***22**, 1420–1426 (1954). - 43.
Daw, M. S. & Baskes, M. I. Embedded-atom method: derivation and application to impurities, surfaces, and other defects in metals.

*Phys. Rev. B***29**, 6443 (1984). - 44.
Kingma, D. P. & Ba, J. L. Adam: a method for stochastic optimization. in

*3rd International Conference on Learning Representations, ICLR*(2015). - 45.
McSkimin, H. J., Andreatch, P. & Glynn, P. The elastic stiffness moduli of diamond.

*J. Appl. Phys.***43**, 985–987 (1972). - 46.
Stillinger, F. H. & Weber, T. A. Computer simulation of local order in condensed phases of silicon.

*Phys. Rev. B***31**, 5262–5271 (1985). - 47.
Jensen, B. D., Wise, K. E. & Odegard, G. M. Simulation of the elastic and ultimate tensile properties of diamond, graphene, carbon nanotubes, and amorphous carbon using a revised reaxFF parametrization.

*J. Phys. Chem. A***119**, 9710–9721 (2015). - 48.
Dolling, G. & Cowley, R. A. The thermodynamic and optical properties of germanium, silicon, diamond and gallium arsenide.

*Proc. Phys. Soc.***88**, 463 (1966). - 49.
Barnard, A. S., Russo, S. P. & Leach, G. I. Nearest neighbour considerations in stillinger-weber type potentials for diamond.

*Mol. Simul.***28**, 761–771 (2002). - 50.
Scherer, C. & Andrienko, D. Understanding three-body contributions to coarse-grained force fields.

*Phys. Chem. Chem. Phys.***20**, 22387–22394 (2018). - 51.
Soper, A. K. & Benmore, C. J. Quantum differences between heavy and light water.

*Phys. Rev. Lett.***101**, 065502 (2008). - 52.
Berendsen, H. J. C., Postma, J. P. M., van Gunsteren, W. F. & Hermans, J. Interaction models for water in relation to protein hydration. in

*Intermolecular forces*. (ed. Pullman, B.) 331–342 (Springer, 1981). - 53.
Wu, Y., Tepper, H. L. & Voth, G. A. Flexible simple point-charge water model with improved liquid-state properties.

*J. Chem. Phys.***124**, 024503 (2006). - 54.
Errington, J. R. & Debenedetti, P. G. Relationship between structural order and the anomalies of liquid water.

*Nature***409**, 318–321 (2001). - 55.
Mills, R. Self-diffusion in normal and heavy water in the range 1-45

^{∘}.*J. Phys. Chem.***77**, 685–688 (1973). - 56.
Wang, H., Junghans, C. & Kremer, K. Comparative atomistic and coarse-grained study of water: what do we lose by coarse-graining?

*Eur. Phys. J. E***28**, 221–229 (2009). - 57.
Matysiak, S., Clementi, C., Praprotnik, M., Kremer, K. & Delle Site, L. Modeling diffusive dynamics in adaptive resolution simulation of liquid water.

*J. Chem. Phys.***128**, 024503 (2008). - 58.
Molinero, V. & Moore, E. B. Water modeled as an intermediate element between carbon and silicon.

*J. Phys. Chem. B***113**, 4008–4016 (2009). - 59.
Chan, H. et al. Machine learning coarse grained models for water.

*Nat. Commun.***10**, 379 (2019). - 60.
Distasio, R. A., Santra, B., Li, Z., Wu, X. & Car, R. The individual and collective effects of exact exchange and dispersion interactions on the ab initio structure of liquid water.

*J. Chem. Phys.***141**, 084502 (2014). - 61.
Thaler, S., Praprotnik, M. & Zavadlav, J. Back-mapping augmented adaptive resolution simulation.

*J. Chem. Phys.***153**, 164118 (2020). - 62.
Belbute-Peres, F. D. A., Smith, K. A., Allen, K. R., Tenenbaum, J. B. & Kolter, J. Z. End-to-end differentiable physics for learning and control. in

*Advances in Neural Information Processing Systems*Vol. 31 (Curran Associates, Inc., 2018). - 63.
Innes, M. et al. A differentiable programming system to bridge machine learning and scientific computing. Preprint at https://arxiv.org/abs/1907.07587 (2019).

- 64.
Hu, Y. et al. DiffTaichi: differentiable programming for physical simulation. in

*8th International Conference on Learning Representations, ICLR*(2020). - 65.
Degrave, J., Hermans, M., Dambre, J. & Wyffels, F. A differentiable physics engine for deep learning in robotics.

*Front. Neurorobot.***13**, 6 (2019). - 66.
Holl, P., Koltun, V. & Thuerey, N. Learning to control PDEs with differentiable physics. in

*8th International Conference on Learning Representations, ICLR*(2020). - 67.
Schäfer, F., Kloc, M., Bruder, C. & Lörch, N. A differentiable programming method for quantum control.

*Mach. Learn. Sci. Technol.***1**, 35009 (2020). - 68.
Thompson, A. P., Plimpton, S. J. & Mattson, W. General formulation of pressure and stress tensor for arbitrary many-body interaction potentials under periodic boundary conditions.

*J. Chem. Phys.***131**, 154107 (2009). - 69.
Chen, X. et al. TensorAlloy: an automatic atomistic neural network program for alloys.

*Comput. Phys. Commun.***250**, 107057 (2020). - 70.
Subramaniyan, A. K. & Sun, C. T. Continuum interpretation of virial stress in molecular simulations.

*Int. J. Solids Struct.***45**, 4340–4346 (2008). - 71.
Van Workum, K., Yoshimoto, K., De Pablo, J. J. & Douglas, J. F. Isothermal stress and elasticity tensors for ions and point dipoles using Ewald summations.

*Phys. Rev. E - Stat. Nonlinear, Soft Matter Phys.***71**, 061102 (2005). - 72.
Van Workum, K., Gao, G., Schall, J. D. & Harrison, J. A. Expressions for the stress and elasticity tensors for angle-dependent potentials.

*J. Chem. Phys.***125**, 144506 (2006). - 73.
Di Pierro, M. & Elber, R. Automated optimization of potential parameters.

*J. Chem. Theory Comput.***9**, 3311–3320 (2013). - 74.
Wang, L. P. et al. Systematic improvement of a classical molecular model of water.

*J. Phys. Chem. B***117**, 9956–9972 (2013). - 75.
Wang, L. P., Martinez, T. J. & Pande, V. S. Building force fields: an automatic, systematic, and reproducible approach.

*J. Phys. Chem. Lett.***5**, 1885–1891 (2014). - 76.
Thaler, S. & Zavadlav, J. Learning neural network potentials from experimental data via Differentiable Trajectory Reweighting. https://github.com/tummfm/difftre, https://doi.org/10.5281/zenodo.5643099 (2021).

## Funding

Open Access funding enabled and organized by Projekt DEAL.

## Author information

### Affiliations

### Contributions

S.T. conceptualized, implemented, and applied the DiffTRe method and conducted MD simulations as well as postprocessing. S.T. and J.Z. planned the study, analyzed and interpreted the results, and wrote the paper.

### Corresponding authors

## Ethics declarations

### Competing interests

The authors declare no competing interests.

## Additional information

**Peer review information** *Nature Communications* thanks Ekin Cubuk and the other, anonymous, reviewer(s) for their contribution to the peer review of this work.

**Publisher’s note** Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

## Supplementary information

## Rights and permissions

**Open Access** This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

## About this article

### Cite this article

Thaler, S., Zavadlav, J. Learning neural network potentials from experimental data via Differentiable Trajectory Reweighting.
*Nat Commun* **12, **6884 (2021). https://doi.org/10.1038/s41467-021-27241-4

Received:

Accepted:

Published:

DOI: https://doi.org/10.1038/s41467-021-27241-4

## Comments

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.