When identifying, within EU-MATHS-IN, a subject that would rally a majority of companies, from banks to aeronautics, from the health sector to the need of renewal in the energy business, Digital Twins emerged as a rather natural choice .
The concept was first created by NASA. Throughout its entire life cycle, a product or process can be accompanied by a virtual representation, called its Digital Twin. Digital Twins allow novel digital assistance for design optimisation, process control, lifecycle management, predictive maintenance, or risk analysis. Digital Twins have become so important to business today that they were identified as one of Gartner’s Top 10 Strategic Technology Trends [18, 27, 47]. They are becoming a business imperative, covering the entire lifecycle of an asset or process and forming the foundation for connected products and services. New business opportunities will emerge, benefitting from the cooperation between large companies, SMEs, startups and academia. To turn this vision into reality, novel mathematical and computer science technologies are required to describe, structure, integrate and interpret across many engineering disciplines.
Supporting this development requires a combination of efforts. High fidelity modelling is key to account for multi-physics and multi-scale systems, and to identify new design levers at the smallest scales. It is often derived from first principle approaches and relies on the power provided by High Performance Computing to deliver the expected prediction in a reasonable time frame. On the other hand, reduced order modelling must provide real-time estimates to enable system optimisation, or in combination with statistical learning to achieve efficient modelling compliant with available data in real time during operation. An example of that kind of need is given by autonomous transport. In particular, corresponding solutions must be realised on the edge to provide sufficiently fast and robust interactions with the real process. The fields of application are growing rapidly in the health sector (e.g., www.digitwin.org) and in climate monitoring .
Key to digital twinning is a joint use of data and first principle approaches. In the frequent cases where large amounts of data prove not to be available, this complementarity makes it possible to realise Smart Data concepts, fostering the efficiency and the robustness of predictions and enabling the quantification of associated uncertainties and risks.
The economic impact of related applications is estimated to cover a worldwide market of 90 billion euros per year by 2025. These opportunities are clearly reinforced by the high concentration of simulation firms in Europe, the highest in the world. Europe also benefits from a long-standing proficiency in mathematics. Several reports [11, 14, 15] have additionally shown the economic impact of mathematical sciences in the US, the Netherlands and France. Furthermore, US studies  in the past have also highlighted the European strength in the field.
This being said, the combined use of statistical learning and first principle modelling is not new. Model calibration is an old topic. The Kalman filter lay at the core of space conquest, and has been essential in the efficient control of remote systems. There remain, however, several challenges for achieving a clever way of combining simulation methods for complex systems, data collection, correction procedures and uncertainty quantification. Besides this, Digital Twins are expected to become self-learning objects that can actually provide guidance and high level services in operation, digesting sensor data with the help of reduced models. A significant part of these ambitions relies on a set of mathematical methods as well as integrated environments.
2 A favourable environment
2.1 European Commission Programs
Horizon Europe2 is a Research and Innovation funding program that is in place until 2027, with a total budget of 95.5 billion euros. Its three pillars are (I) Excellent Science, (II) Global Challenges and European Industrial Competitiveness, (III) Innovative Europe. The European Research Council (President: J.-P. Bourguignon) supports pillar I, while the DG Connect Directorate (Dir: K. Rouhana) specifically supports the present topic, as well as High Performance or Quantum Computing, as part of pillar II. Calls are the mechanisms through which individual projects get funded.
Of key importance is the EuroHPC Joint Undertaking. This is a joint initiative between the EU, European countries and private partners to develop a World Class Supercomputing Ecosystem in Europe. The initiative has two major partners onboard: ETP4HPC and the Big Data Value Association, which reflects the ambition to establish synergies between highly accurate simulations and data analytics. It is also oriented towards large companies and SMEs, in order to boost innovation potential and competitiveness, while widening the use of HPC in Europe. EU-MATHS-IN is involved in the TransContinuum Initiative, as a binding effort between ETP4HPC and BDVA, and contributes to a Strategic Research Agenda coordinated by Zoltán Horváth.
2.2 Some open platforms and funded projects
Open software enables new actors including SMEs, potentially with the help of a proper knowledgeable accompaniment, to access highly technological solutions, develop new ideas and start new businesses depending on associated exploitation rights. It clearly shifts development effort from low-value work to value creation. Open platforms can also act as binding environments to interface commercial products with clear added value. Sharing software components between academia and industry may be a way to reinforce the European momentum on the development of new mathematical algorithms in order to, e.g., take advantage of new European HPC architectures or data-simulation hybridation. The development of better interoperability is expected to accelerate innovation and boost European leadership.
Open software has been developed ranging from simulation tools to data analysis with applications like mechanics or biology. Furthermore, it has clearly evolved to provide integrated environments, especially relying on Python interfacing which is a key element to foster collaboration, to integrate various expertises and to make content widely accessible. Let us mention here two projects supported by the European Community that clearly support this ambition.
Societal challenges are increasing in complexity, and contributing to their resolution requires a holistic approach. It is necessary to provide decision-makers with tools that allow long-term risk analysis, improvements or even optimisation and control. One of the key technologies in this process is the use of mathematical Modelling, Simulation and Optimisation (MSO) methods, which have proven to be effective tools for solving problems such as the realistic prediction of wind fields, solar radiation, air pollution and forest fires, prediction of climate change, improving the filtration process for drinking water treatment and optimisation methods for intensity-modulated radiation therapy. These methods are highly complex and are typically processed via the most modern tools of ICT, including high performance computing and access to big data bases; they usually require the support of skilled experts, who are often not available, in particular in small and medium-sized businesses. The main goal of this project is to construct an e-infrastructure that provides, in a user-driven, integrative way, tailored access to the necessary services, resources and even tools for fast prototyping, also providing the service producers with the mathematical framework. The e-infrastructure consists of an integrated MSO application catalogue containing models, software, validation and benchmark and the MSO cloud, a user-friendly cloud infrastructure for selected MSO applications and developing frameworks from the catalogue. This will reduce the ‘time-to-market’ for consultants working on the above-mentioned societal challenges.
Open Dream Kit
OpenDreamKit is a project that brings together a range of projects and associated software to create and strengthen virtual research environments. The most widely-used research environment is the Jupyter Notebook from which computational research and data processing can be directed. The OpenDreamKit project provides interfaces to well-established research codes and tools so that they can be used seamlessly and combined from within a Jupyter Notebook. OpenDreamKit also supports open source research codes directly by investing into structural improvements and new features to not only connect all of these tools, but also enrich them and make them more sustainable.
3 A mathematical toolbox
We emphasise here four complementary perspectives on the joint exploitation of simulation and data: (i) techniques coming from optimal command, ranging from calibration to filtering, that allow identification of hidden parameters, model correction and handling noisy forcing terms, (ii) solution space reduction that enables fast solving and efficient correction, (iii) multi-fidelity co-kriging in order to merge and prioritise the feedback issuing from simulations and measurements on given observables, (iv) physics-inspired neural networks that directly learn the model solution, based on an a priori (set of) underlying model(s). Each of these approaches corresponds to a different balance in the roles of physical models and collected data. In this text, we do not claim any exhaustivity, and mathematical descriptions will remain formal. The idea is to provide a fairly broad overview of the field while remaining accessible to the vast majority of mathematician readers.
3.1 Optimal control: From calibration to filtering
Let be a vector space of parameters. For every , the solution of the “best-knowledge” model is defined to be the solution of the partial differential system
in a weak formal form, where is a Hilbert space and its dual. The parameter space encompasses modelling choices, the domain shape and the boundary conditions, as well as forcing terms.
Additionally, the physical quantity
is known, potentially with some noise, at the sampling points as . The purpose of our quest is to take benefit from this information to estimate outside the sampling points, or to improve the solution itself given by the model.
Calibration and model correction
As an example, let us augment into in order to account for a modelling error term in as we consider the augmented system
where . For each sampling point , the idea is to find the most adequate parameter of the model close to and the model correction in order to best account for the measurement . The calibration and model correction in the vicinity of can be formulated as finding
where in , and stands for the Hilbertian norms in , and . The scalar coefficients are taken sufficiently large. As classically given by the techniques of optimal control , the solution is characterised by the following system:
in , and respectively, with . Equation (3) is called the adjoint problem and equation (4) defines the gradient of the cost function with respect to , that must vanish. The gradient expression enables an iterative calibration in order to avoid the resolution of the coupled system (2)–(3)–(4). One has and we use the notation in .
Once the calibration and error corrections are performed at sampling points, let us introduce the functions obtained by kriging under the form
that comply with values , , at sampling points . For every , an updated model can be formulated as finding the solution such that
with the estimator . For purposes of efficiency, the spaces and can be replaced by some reduced basis approximation in (5).
Parameters and measurements can be considered as random variables. The relevance of this point of view is supported by their potential discrete natures and by the uncertainties and noise attached to them. Assume (resp. ) is the probability density followed by (resp. ). Bayesian inference provides the conditional density
where is called the prior, i.e., the a priori distribution expected on ; is the output likelihood given , i.e., the uncertainty on the output measurement or simulation. It results in an assessment of the distribution, known as posterior, that can be used as a new prior and so on, until uncertainties are judged satisfactory [22, 26].
From a practical standpoint, a Markov Chain Monte Carlo approach can be used to simulate samples according to the distribution followed by , and a surrogate model – for instance relying on reduced bases – can be used to diminish the computational cost required to determine the output .
Selection of parameters
In order to make the approaches as efficient as possible, the parametric space must be reduced. A rather natural approach relies on variance decomposition, i.e. on the identification of subspaces within for which the covariance matrix possesses the largest components; This can be done by Sobol decomposition or Principal Component Analysis for linear models . Observe that in the vicinity of , the privileged most influential subspaces are given by the singular value decomposition of in , the differential being computed by the adjoint method.
Sometimes, the parameters correspond to local characteristics of the model, such as material laws, and can be selected without involving the full resolution of the model. This is the case in Ortiz et al. [24, 28], which rely on local measurements or simulations with a closest-point projection approach.
Filtering: From space conquest to cardiovascular modelling
The complementarity between first principle modelling and data analytics was pioneered in a unique manner by space conquest. This clever combination makes it possible to benefit from the predictive power of simple dynamic models and the ability to cope with noise and uncertainties within the environment; as a result, it offers the possibility of automated decisions when long transmission times allow for neither full real-time feedback on the system state nor efficient human steering. We describe the main associated ideas within a linear framework; cf. [5, 6, 13, 30].
Filtering and optimal control
The above motivation was the main boost for the development of Kalman filtering, which is closely related to optimal control. Let describe the state of the system at time such that
where , , stands for an unknown forcing contribution and for the uncertainty on the initial condition. Measurements over , with , up to an error , are available in order to help estimate the real trajectory, through the determination of and such that
is infimised. Optimality is achieved for and , where the adjoint state is a solution of the backward system in time
In order to avoid the difficulty of a two-end problem in and , the optimal solution can be proven to decompose as , where is the operator solution from Riccati’s equation
The component obeys the filtered dynamics
where is Kalman’s gain; it reaches the optimal trajectory at time , as . Beyond this Linear Quadratic setting, filtering can incorporate robust control, and adapt in many ways to the case of nonlinear systems (Extended Kalman Filter, Unscented Kalman Filter).
Dynamic programming is of particular importance in order to proceed to nonlinear extensions. For every and , let us define the cost-to-come function
where the minimum is taken over controls and trajectories with end-points and . The cost function is a solution of the Hamilton–Jacobi–Bellman equation
in the sense of viscosity solutions, where
Observe that the optimality equations above read
Taking as an initial condition, let us assume the HJB equation admits a solution . Then, for all the optimal command is given by
This provides the estimated dynamics
In the present linear setting, one has 
Assume equation (6) is interpreted as astochastic differential equation, where and are zero-mean independent Gaussian processes with covariance matrices and respectively. The best mean square estimator follows the same equation as in the previous paragraphs with and . The covariance matrix obeys Riccati’s equation.
Stability and control
Kalman’s gain achieves control optimality in the sense detailed above. Nevertheless, it can prove costly to determine, and difficult to access for distributed systems. As a matter of fact, some feedback terms acting as Lyapunov functions can suffice for practical purposes. Let us assume in the above that the gain is chosen in the form with a symmetric definite positive matrix and a coefficient to determine. It follows that the error satisfies
and can be made rapidly decreasing provided that and remain moderate and is taken sufficiently large. This approach has been implemented with multiple refinements by Moireau et al. , comparing in-depth the displacement vs. velocity controls.
3.2 Solution space reduction
The notion of reduced bases makes it possible to resolve the equations of the model in a low-dimensional subspace of solutions, rather than a large full finite element space for instance. It is reputed to go back to Rayleigh’s intuition, and is key to benefit from surrogate models that comply with the first physical principles, with high computational efficiency. It is particularly useful for high-dimensional models, as in Quantum Mechanics for instance. The approach has been made particularly popular by the work of Maday, Patera et al. . Some interesting challenges rise when handling highly nonlinear problems ; it is particularly striking when contact mechanics is involved , when structure preservation is concerned [21, 20], or when local accuracy is of particular importance . The efficiency of the method is well established when the Kolmogorov -width of the solution manifold rapidly decreases as ; cf. [2, 8].
Parameterised-Background Data-Weak (PBDW) approach
be the solution manifold for the “best-knowledge” model. It can be gradually approximated with diminishing errors by solutions of the model in reduced spaces .
The true solution is unknown, but can be partially captured by experimental observations given as continuous linear forms of the solution: . They provide the scalar quantities . Let us write for the associated liftings satisfying for all , where denotes the inner product in the Hilbert space . It is imperative that the sensors be numerous enough () to control the components of the reduced solution in the selected space ; more specifically, one must have where is orthogonal to in .
and the norm is infimised. It boils down to finding and such that
This has been extended to a dynamic setting by Benaceur , in collaboration with Patera. The method allows for a real-time correction of the solution, based upon available measurements. In case of noisy measurements, a regularisation is required and the following functional:
is infimised in order to compromise between the minimisation of the term and the constraint (7), thus avoiding overfitting.
Data-driven reduced modelling
Each time the above reconstruction generates a prediction for a given state of the system, the vector can be replaced by in the reduced basis for the system. This can be done by the dynamic reduced basis low rank adaption proposed by Peherstorfer and Wilcox .
Another point of view consists in fitting the expression of the operators involved in the best-knowledge model, when projected onto an a priori reduced basis. This idea was proposed by Peherstorfer and Wilcox , who named it operator inference.
Finally, composite spaces can be constructed by the assembling of subdomains in which key driving parameters are retained and local reduced bases are adopted; coupling between subdomains can be performed via Lagrange multipliers, and the hidden parameters (for instance describing certain levels of damage within a structure) can be determined by statistical classification methods. Such approaches have been developed by Patera et al., and are called Port-Reduced Reduced-Basis Component (PR-RBC) methods; cf. [7, 16, 48].
Tensor approximation of solutions
Since functions in separable form are dense in spaces of sufficiently regular functions, one can decompose the solution of the parameterised model problem in the form
by means of Proper Generalised Decompositions . Of course, can easily incorporate a wide variety of correction terms, so as to account for the current state of the system. This general philosophy has been widely popularised by Chinesta and Nouy; see [9, 40] and the references therein.
3.3 Multi-fidelity prediction and co-kriging
In many practical cases, several sources of information (simulation and measurement campaigns) provide predictions of various accuracies for the observables. Let be the parameter space under consideration. For every , the -th measurement or simulation campaign is performed at sampling points with the model and provides data points . The accuracy increases with and the most accurate forecast therefore corresponds to .
We set , and recall the construction of the universal kriging as a non-biased approximation of with minimum variance in the form
Universal kriging additionally assumes that is a Gaussian process with unknown average of the form
and space correlation , where is a variance scaling factor and a parameter of the space-correlation function . The parameters , and can be determined through maximum likelihood. The non-biased minimum variance predictor is achieved for , where the matrices are given by their components
and the vectors by
where the following Gaussian processes are defined by their means and covariance matrices as
is a Gaussian process with distribution
where and .
The combined use of reduced models, full accuracy simulations and measurements clearly allow for very efficient and accurate surrogate models. The co-kriging above, further developed in , has the huge advantage of recursiveness and as a result, a very accessible computational cost.
3.4 Physics-Informed Neural Network
Neural networks have had great success with classification problems, together with Support Vector Machines for instance . They generate functions that are dense among continuous functions (cf. Cybenko ), and conciliate smoothness with the ability to represent thresholds quite accurately. Fitting can be performed by a back descent gradient inspired by optimal control techniques. Furthermore, robust and powerful Python libraries, like Tensor Flow, are freely available.
Physics-Informed Neural Networks were introduced by Raissi, Perdikaris and Karniadakis . They combine the statistical learning of the solution, say on the space-time domain sampled on points as , under the penalised constraint that is expected to resolve a partial differential equation of the form in . This reads as the infimisation
which is close to (1) when . A Bayesian approach can be used to identify the parameters from the model, as developed in . Lucor et al. have developed the approach for the thermo-mechanical simulation of an incompressible viscous flow ; cf. Figures 5 and 6.
The combined use of first principle models and data analytics is an avenue for predictive sciences. It is a privileged way to synergise the modelling knowledge present in simulation software with the relevance of available data, while guaranteeing a high level of predictiveness in operation. Beyond the necessity of a growing mathematical toolbox to handle problems of optimal control with extreme efficiency, several challenges are implied: (i) the necessity of developing porosity at the interface between competences (numerical analysis, optimal control and automatism, high performance computing, statistics, computer sciences), (ii) the need for integrated development environments, with a role to play in the question of open software, and (iii) data protection, as data becomes an outstanding source of value.
Acknowledgements I thank Krzysztof Burnecki for the kind invitation to write this article. I am indebted to Philippe Moireau, Amina Benaceur, Didier Lucor and Anne Sergent for providing the illustrations displayed in Section 3, and for our exchanges on the corresponding advances. The promotion of Digital Twins also owes thanks to my colleagues from the EU-MATHS-IN Industrial Core Team, with special thanks to Dirk Hartman, Zoltán Horváth, Wil Schilders, Cor van Kruijsdijk and Hubert Tardieu. In this regard, Section 1 presents several collective statements.
ATOS, Bosch, Dassault Aviation, EcoMT, EY, Michelin, NORS Group, Shell, Siemens
- Y. S. Abu-Mostafa, M. Magdon-Ismail and H-T. Lin, Learning from Data. AMLbook.com (2012)
- M. Barrault, Y. Maday, N. C. Nguyen and A. T. Patera, An ‘empirical interpolation’ method: application to efficient reduced-basis discretization of partial differential equations. C. R. Math. Acad. Sci. Paris339, 667–672 (2004)
- P. Bauer, B. Stevens and W. Hazeleger, A digital twin of earth for the green transition. Nature Climate Change11, 80–83 (2021)
- A. Benaceur, Réduction de modèles en thermo-mécanique. PhD thesis, Paris Est University (2018)
- A. Bensoussan, Filtrage optimal des systèmes linéaires. Dunod (1971)
- A. Bensoussan, Stochastic Control of Partially Observable Systems. Cambridge University Press, Cambridge (1992)
- M. A. Bhouri and A. Patera, A two-level parameterized model-order reduction approach for time-domain elastodynamics. Comput. Methods Appl. Mech. Engrg.385, 114004 (2021)
- P. Binev, A. Cohen, W. Dahmen, R. DeVore, G. Petrova and P. Wojtaszczyk, Convergence rates for greedy algorithms in reduced basis methods. SIAM J. Math. Anal.43, 1457–1472 (2011)
- F. Chinesta, E. Cueto, E. Abisset-Chavanne, J. L. Duval and F. El Khaldi, Virtual, digital and hybrid twins: a new paradigm in data-based engineering and engineered data. Arch. Comput. Methods Eng.27, 105–134 (2020)
- F. Chinesta and P. Ladevèze (eds.), Separated Representations and PGD-Based Model Reduction. CISM International Centre for Mechanical Sciences. Courses and Lectures 554, Springer, Vienna (2014)
- CMI, Etude de l’impact socio-économique des mathématiques en france [CMI, A study of the soci-economical impact of mathematics in France] (2015)
- G. Cybenko, Approximation by superpositions of a sigmoidal function. Math. Control Signals Systems2, 303–314 (1989)
- B. d’Andréa Novel and M. Cohen de Lara, Commande linéaire des systèmes dynamiques. Modélisation. Analyse. Simulation. Commande. [Modeling. Analysis. Simulation. Control], Masson, Paris (1994)
- Deloitte, Measuring the economic benefits of mathematical sciences research in the UK. Deloitte (2012)
- Deloitte, Mathematical sciences and their value for the dutch economy. Deloitte (2014)
- J. L. Eftang and A. T. Patera, Port reduction in parametrized component static condensation: approximation and a posteriori error estimation. Internat. J. Numer. Methods Engrg.96, 269–302 (2013)
- EU-MATHS-IN Industrial Core Team, Modelling, simulation and optimization in a data rich environment: a window of opportunity to boost innovation in Europe. Foundation EU-MATHS-IN, European Service Network of Mathematics for Industry (2018)
- D. Hartmann and H. Van der Auweraer, Digital twins. Siemens. arXiv:2001.09747 (2019)
- T. Hastie, R. Tibshirani and J. Friedman, The Elements of Statistical Learning. 2nd ed., Springer Series in Statistics, Springer, New York (2009)
- P. Hauret, Multiplicative reduced basis for elasticity. submitted (2021)
- J. Hesthaven and C. Pagliantini, Structure-preserving reduced basis methods for Hamiltonian systems with a nonlinear poisson structure (2018)
- D. Higdon, J. Gattiker, B. Williams and M. Rightley, Computer model calibration using high-dimensional output. J. Amer. Statist. Assoc.103, 570–583 (2008)
- B. Iooss and P. Lemaître, A review on global sensitivity analysis methods. In Uncertainty Management in Simulation-Optimization of Complex Systems, Oper. Res./Comput. Sci. Interfaces Ser. 59, Springer (2015)
- K. Karapiperis, L. Stainier, M. Ortiz and J. E. Andrade, Data-driven multiscale modeling in mechanics. J. Mech. Phys. Solids147, 104239, 16 (2021)
- M. C. Kennedy and A. O’Hagan, Predicting the output from a complex computer code when fast approximations are available. Biometrika87, 1–13 (2000)
- M. C. Kennedy and A. O’Hagan, Bayesian calibration of computer models. J. R. Stat. Soc. Ser. B Stat. Methodol.63, 425–464 (2001)
- M. Kerremans, D. Cearley, A. Velosa and M. Walker, Top 10 strategic technology trends for 2019: digital twins.www.gartner.com/en/documents/3904569/top-10-strategic-technology-trends-for-2019-digital-twin, Gartner (2019)
- T. Kirchdoerfer and M. Ortiz, Data-driven computational mechanics. Comput. Methods Appl. Mech. Engrg.304, 81–101 (2016)
- L. Le Gratiet and J. Garnier, Recursive co-kriging model for design of computer experiments with multiple levels of fidelity. Int. J. Uncertain. Quantif.4, 365–386 (2014)
- J.-L. Lions, Contrôle optimal de systèmes gouvernés par des équations aux dérivées partielles. Dunod, Paris; Gauthier-Villars, Paris (1968)
- D. Lucor, A. Agrawal and A. Sergent, Physics-aware deep neural networks for surrogate modeling of turbulent natural convection, preprint, arXiv:2103.03565 (2021)
- Y. Maday, Reduced basis method for the rapid and reliable solution of partial differential equations. In International Congress of Mathematicians. Vol. III, Eur. Math. Soc., Zürich, 1255–1270 (2006)
- Y. Maday, N. C. Nguyen, A. T. Patera and G. S. H. Pau, A general multipurpose interpolation procedure: the magic points. Commun. Pure Appl. Anal.8, 383–404 (2009)
- Y. Maday, A. T. Patera, J. D. Penn and M. Yano, A parameterized-background data-weak approach to variational data assimilation: formulation, analysis, and application to acoustics. Internat. J. Numer. Methods Engrg.102, 933–965 (2015)
- Y. Maday, A. T. Patera, J. D. Penn and M. Yano, PBDW state estimation: noisy observations; configuration-adaptive background spaces; physical interpretations. In CANUM 2014 – 42e Congrès National d’Analyse Numérique, ESAIM Proc. Surveys 50, EDP Sci., Les Ulis, 144–168 (2015)
- Y. Maday and T. Taddei, Adaptive PBDW approach to state estimation: noisy observations; user-defined update spaces. SIAM J. Sci. Comput.41, B669–B693 (2019)
- P. Moireau, A discrete-time optimal filtering approach for non-linear systems as a stable discretization of the Mortensen observer. ESAIM Control Optim. Calc. Var.24, 1815–1847 (2018)
- P. Moireau, C. Bertoglio, N. Xiao, C. A. Figueroa, C. A. Taylor, D. Chapelle and J.-F. Gerbeau, Sequential identification of boundary support parameters in a fluid-structure vascular model using patient image data. Biomechanics and Modeling in Mechanobiology12, 475–496 (2012)
- P. Moireau, D. Chapelle and P. Le Tallec, Filtering for distributed mechanical systems using position measurements: perspectives in medical imaging. Inverse Problems25, 035010, 25 (2009)
- A. Nouy, Proper generalized decompositions and separated representations for the numerical solution of high dimensional stochastic problems. Arch. Comput. Methods Eng.17, 403–434 (2010)
- NSF and WTEC, www.nsf.gov/pubs/reports/sbes_final_report.pdf; www.wtec.org/sbes/SBES-GlobalFinalReport.pdf.
- B. Peherstorfer and K. Willcox, Dynamic data-driven reduced-order models. Comput. Methods Appl. Mech. Engrg.291, 21–41 (2015)
- B. Peherstorfer and K. Willcox, Data-driven operator inference for nonintrusive projection-based model reduction. Comput. Methods Appl. Mech. Engrg.306, 196–215 (2016)
- P. Perdikaris, D. Venturi, J. O. Royset and G. E. Karniadakis, Multi-fidelity modelling via recursive co-kriging and Gaussian–Markov random fields. Proc. R. Soc. A.471 (2015)
- M. Raissi, P. Perdikaris and G. E. Karniadakis, Physics-informed neural networks: a deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys.378, 686–707 (2019)
- S. Riffaud, M. Bergmann, C. Farhat, S. Grimberg and A. Iollo, The DGDD method for reduced-order modeling of conservation laws. J. Comput. Phys.437, 110336, 19 (2021)
- Digital twins – explainer video. youtu.be/ObGhB9CCHP8, Siemens AG (2019)
- T. Taddei, J. D. Penn, M. Yano and A. T. Patera, Simulation-based classification: a model-order-reduction approach for structural health monitoring. Arch. Comput. Methods Eng.25, 23–45 (2018)
- L. Yang, X. Meng and G. E. Karniadakis, B-PINNs: Bayesian physics-informed neural networks for forward and inverse PDE problems with noisy data. J. Comput. Phys.425, 109913, 23 (2021)
Cite this article
Patrice Hauret, At the crossroads of simulation and data analytics. Eur. Math. Soc. Mag. 121 (2021), pp. 9–18DOI 10.4171/MAG/20