WO2024215873A1 - System and methods for simulating the physical behavior of objects - Google Patents
System and methods for simulating the physical behavior of objects Download PDFInfo
- Publication number
- WO2024215873A1 WO2024215873A1 PCT/US2024/024050 US2024024050W WO2024215873A1 WO 2024215873 A1 WO2024215873 A1 WO 2024215873A1 US 2024024050 W US2024024050 W US 2024024050W WO 2024215873 A1 WO2024215873 A1 WO 2024215873A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- subdomains
- subdomain
- deep learning
- model
- training data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/20—Design optimisation, verification or simulation
- G06F30/27—Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/20—Design optimisation, verification or simulation
- G06F30/23—Design optimisation, verification or simulation using finite element methods [FEM] or finite difference methods [FDM]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computing arrangements based on specific mathematical models
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/55—Controlling game characters or game objects based on the game progress
- A63F13/57—Simulating properties, behaviour or motion of objects in the game world, e.g. computing tyre load in a car race game
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2111/00—Details relating to CAD techniques
- G06F2111/10—Numerical modelling
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2111/00—Details relating to CAD techniques
- G06F2111/18—Details relating to CAD techniques using virtual or augmented reality
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2119/00—Details relating to the type or aim of the analysis or the optimisation
- G06F2119/08—Thermal analysis or thermal optimisation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2119/00—Details relating to the type or aim of the analysis or the optimisation
- G06F2119/14—Force analysis or force optimisation, e.g. static or dynamic forces
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06G—ANALOGUE COMPUTERS
- G06G7/00—Devices in which the computing operation is performed by varying electric or magnetic quantities
- G06G7/48—Analogue computers for specific processes, systems or devices, e.g. simulators
Definitions
- FEM finite element method
- Embodiments of the present disclosure provide an artificial intelligence (Al)-driven system and methods for modeling the response/behavior of physical objects within a digital environment.
- digital models of physical objects are processed using a domain decomposition method (DDM) model configured to solve a boundary value problem by splitting the digital model into a plurality of subdomains (e.g., smaller boundary value problems on subdomains).
- DDM domain decomposition method
- Each subdomain can then be individually solved (e.g., as a partial differential equation) using a deep learning model to predict horizontal and vertical displacements or the whole displacement field, which in turn can be used to update boundary conditions for the neighboring subdomains. From the respective boundary conditions, a displacement field and/or a stress field can be predicted.
- the outputs of the proposed system and methods can be employed to generate data (e.g., output data, user interface data, digital structures) for various applications, including simulations, computer-aided design (CAD) systems, and virtual reality (VR) systems, for example, to model and solve engineering design problems.
- data e.g., output data, user interface data, digital structures
- CAD computer-aided design
- VR virtual reality
- a system for simulating a physical behavior of objects can include: a processor; and memory having instructions stored thereon that, when executed by the processor, cause the system to: obtain a digital model of a physical object; decompose the digital model into a plurality of subdomains, wherein the plurality of subdomains at least partially overlap (often 50% overlap is used to facilitate the construction of subdomains), independently solve each subdomain of the plurality of subdomains using a deep learning model, wherein the deep learning model is trained to predict horizontal and vertical displacements (or the whole displacement field) for each subdomain of the plurality of subdomains; iteratively update displacement boundary conditions for each of the plurality of subdomains based on the respective predicted horizontal and vertical displacements (or the whole displacement field); and once convergence is achieved, predict at least one of a displacement field or a stress field using the updated boundary conditions for each subdomain of the plurality of subdomains.
- the instructions further cause the system to: train the deep learning model to predict a response of each of the plurality of subdomains using a training data set, wherein the training data set includes at least one of subdomain geometry, material properties, horizontal and vertical displacements or the whole displacement field, and boundary conditions for a plurality of different types of subdomains.
- the training data set is generated using finite element domain decomposition method (FE-DDM) simulations or by extracting multiple subdomains from finite element simulation of the field in a larger domain to solve each of the plurality of different types of subdomains.
- FE-DDM finite element domain decomposition method
- the training data set is generated using finite element simulations to solve each of the plurality of different types of subdomains.
- the deep learning model is one of a plurality of deep learning models, wherein the instructions further cause the system to: train the plurality of deep learning models, wherein each of the plurality of deep learning models is trained to predict a response of a unique type of subdomain.
- the plurality of deep learning models are trained using respective training data sets, wherein each of the training data sets includes at least one of midline horizontal and vertical displacements or the whole displacement field, subdomain geometry, material properties, and boundary conditions for a respective unique type of subdomain.
- the plurality of deep learning models are trained using respective training data sets, wherein each of the training data sets includes at least one of subdomain geometry, material properties, displacements along planes parallel to X-Y, X-Z, and Y-Z planes or the whole displacement field, subdomain geometry, material properties, and boundary conditions for a respective unique type of subdomain for a 3 -dimensional (3D) physical object.
- the instructions further cause the system to: identify one of the plurality of deep learning models that is applicable to each of the plurality of subdomains prior to independently solving each subdomain, wherein each of the identified deep learning models is used to solve a respective one of the plurality of subdomains.
- the stress field is predicted using a second deep learning model.
- the digital model is decomposed into the plurality of subdomains using an overlapping Schwarz method.
- the plurality of subdomains overlap by 50%.
- the plurality of subdomains overlap by between 10% and 80%.
- the deep learning model is a convolutional neural network or Fourier Neural Operator.
- the digital model of the physical object is a 2-dimensional (2D) representation of the physical object.
- the digital model of the physical object is a 3 -dimensional (3D) model representation of the physical object.
- the instructions further cause the system to: generate output data for the physical object based at least on the predicted displacement field over the entire domain and/or stress field.
- the output data is employed in a computer-aided design system, simulation, or virtual reality system.
- a method of simulating the physical behavior of an object can include: obtaining a digital model of the object; decomposing the digital model into a plurality of subdomains, wherein the plurality of subdomains at least partially overlap; independently solving each subdomain of the plurality of subdomains using a deep learning model, wherein the deep learning model is trained to predict horizontal and vertical displacements (or the whole displacement field) for each subdomain of the plurality of subdomains; iteratively update boundary conditions for each of the plurality of subdomains based on the respective predicted horizontal and vertical displacements (or the whole displacement field); and once convergence is achieved, predict at least one of a displacement field or a stress field using the updated boundary conditions for each subdomain of the plurality of subdomains.
- a method of simulating the physical behavior of an object can include: obtaining a digital model of the object; decomposing the digital model into a plurality of subdomains, wherein the plurality of subdomains at least partially overlap; independently solving each subdomain of the plurality of subdomains using a deep learning model, wherein the deep learning model is trained to predict a displacement field or solution field (e.g., temperature field, electromagnetic field) for each subdomain of the plurality of subdomains; iteratively update boundary conditions for each of the plurality of subdomains based on the respective predicted displacement fields or solution fields; and once convergence is achieved, predict at least one of a displacement field over the entire domain, a solution field over the entire domain, or a stress field using the updated boundary conditions for each subdomain of the plurality of subdomains.
- a displacement field or solution field e.g., temperature field, electromagnetic field
- a non-transitory computer-readable medium has instructions stored thereon that when executed by a processor, cause a computing device to: obtain a digital model of a physical object; decompose the digital model into a plurality of subdomains, wherein the plurality of subdomains at least partially overlap; independently solve each subdomain of the plurality of subdomains using a deep learning model, wherein the deep learning model is trained to predict horizontal and vertical displacements (or the whole displacement field) for each subdomain of the plurality of subdomains; iteratively update boundary conditions for each of the plurality of subdomains based on the respective predicted horizontal and vertical displacements (or the whole displacement field); and once convergence is achieved, predict at least one of a displacement field or a stress field using the updated boundary conditions for each subdomain of the plurality of subdomains.
- FIG. l is a block diagram of an example computing device, according to some implementations.
- FIG. 2 is a flow chart of a process for simulating the physical behavior of an object using a digital model, according to some implementations.
- FIGS. 3A-3B are schematic diagrams illustrating domain partitioning and updating Boundary Conditions (BCs) of a subdomain in non-overlapping and overlapping domain decomposition (DDM) based on the field approximated in neighboring subdomains, according to some implementations.
- BCs Boundary Conditions
- FIGS. 4A-4C are schematic diagrams illustrating the approximation of the field in a porous domain using both the non-overlapping and overlapping DDM techniques, according to some implementations.
- FIGS. 5A-5B are graphs illustrating the effect of the number of subdomains and the overlap percentage between neighboring subdomains in the overlapping DDM.
- FIG. 6 is a schematic diagram illustrating using subdomains with 50% overlap to discretize a domain in the Deep Learning-Driven Domain Decomposition (DLD 3 ) method, according to some implementations.
- DLD 3 Deep Learning-Driven Domain Decomposition
- FIG. 7 are schematic diagrams illustrating partitioning a domain with arbitrary geometry and BCs for the DLD 3 method, according to some implementations.
- FIG. 8 are schematic diagrams illustrating subdividing a domain into overlapping subdomains, according to some implementations.
- FIG. 9 A shows an example of a virtually reconstructed geometrical model.
- FIG. 9B depicts a small portion of the conforming mesh generated using Conforming to Interface Structured Adaptive Mesh Refinement (CISAMR).
- CISAMR Conforming to Interface Structured Adaptive Mesh Refinement
- FIGS. 10A-10B are schematic diagrams showing the Finite Element (FE) approximation of displacement and strain fields in the y-direction for the domain and the corresponding mesh generated using the CIS AMR algorithm shown in FIGS. 9A-9B.
- FE Finite Element
- FIG. 11 is a schematic diagram illustrating extracting random subdomain geometries and the corresponding field/BC as entries into the training dataset, according to some implementations.
- FIGS. 12A-12D are schematic diagrams illustrating four different Fourier Neural Operator (FNO) models trained based on the subdomain geometry and applied BC, according to some implementations.
- FNO Fourier Neural Operator
- FIG. 13 is a flowchart diagram illustrating a method for extracting random subdomain geometries and the corresponding field/BC as entries into the training dataset, according to some implementations.
- FIG. 14 is a graph depicting training and validation losses plotted against the number of epochs during the training of the FNO model.
- FIG. 15 is a schematic diagram that shows FNO prediction of the magnitude of the displacement field in several subdomains and the corresponding distribution of the error vs finite element (FE) simulation of the field using conforming meshes.
- FE finite element
- FIGS. 16A-16B are schematic diagrams corresponding with a first problem in a conducted study.
- FIGS. 17A-17C are results for the first problem from the conducted study.
- FIGS. 18A-18B are schematic diagrams corresponding with a second problem in a conducted study.
- FIGS. 19A-19C illustrate results for the second problem from the conducted study.
- FIGS. 20A-20B are schematic diagrams corresponding with a third problem in a conducted study.
- FIGS. 21 A-21B show results for the third problem from the conducted study.
- a system and methods for simulating the physical behavior of objects are shown, according to various implementations.
- “physical behavior” generally refers to a mechanical response, thermal response, chemical response, or the like for an object.
- the system and methods disclosed herein relate to an artificial intelligence (Al)-driven modeling technique for modeling the response/behavior of physical objects within a digital environment.
- Al artificial intelligence
- digital models of physical objects are processed using a domain decomposition method (DDM) model configured to solve a boundary value problem by splitting the digital model into a plurality of subdomains (e.g., smaller boundary value problems on subdomains).
- DDM domain decomposition method
- Each subdomain can then be individually solved (e.g., as a partial differential equation) using a deep learning model to predict horizontal and vertical displacements (or the whole displacement field), which in turn can be used to predict boundary conditions for the respective subdomain. From the respective boundary conditions, a displacement field and/or a stress field can be predicted.
- DLD 3 This technique - sometimes referred to herein as DLD 3 - can address several disadvantages of FEM as described above, as well as limitations of various other Al techniques for predicting the physics-based response of problems with arbitrary geometry and boundary conditions.
- DLD 3 can result in: 1) reduced labor cost associated with the modeling process - for example, no need for mesh generation; 2) reduced computational cost; 3) enabling the simulation of massive problems not feasible using existing computational resources and FE-based algorithms; and 4) addresses the challenges associated with the generalizability of Al models to problems with arbitrary shapes and boundary conditions.
- DLD 3 can be used to simulate a variety of response problems for an object, including but not limited to mechanical, thermal, and chemical responses.
- DLD 3 can also be applied to transient heat transfer problems, plasticity problems, and so on.
- DLD 3 can be applied to both 2-dimensional (2D) and three-dimensional (3D) digital models.
- the Deep Learning-Driven Domain Decomposition (DLD 3 ) algorithm described herein is a generalizable Artificial Intelligence (Al)-driven technique for simulating two- dimensional linear elasticity problems with arbitrary geometry and boundary conditions (BCs).
- DLD 3 uses a set of pre-trained Al models capable of predicting the linear elastic displacement field in small subdomains of a given domain with various geometries/BCs.
- the overlapping Schwarz domain decomposition method (DDM) is then utilized to iteratively update the subdomain BCs to approximate the problem response by enforcing a continuous displacement field in the domain.
- the Fourier Neural Operator (FNO) model was chosen as the Al engine used in the DLD 3 algorithm due to its data efficiency and high accuracy.
- This disclosure contemplates that other model architectures can be used.
- This disclosure presents a framework relying on geometry reconstruction and automated meshing algorithms to acquire millions of data used for training these FNO models based on high-fidelity finite element (FE) simulation results.
- FE finite element
- the Finite Element Method (FEM) and commercial finite element software are widely used to simulate the thermal and mechanical behaviors of materials/structures across various industries, such as automotive, aerospace, and defense.
- FEM Finite Element Method
- commercial finite element software are widely used to simulate the thermal and mechanical behaviors of materials/structures across various industries, such as automotive, aerospace, and defense.
- the operational and computational costs associated with performing FE simulations could be significant.
- These challenges often result in compromises such as the oversimplification of the problem geometry/microstructure, using coarse meshes, or minimizing the number of simulations, which undermine the fidelity of results.
- the high operational cost of an FE analysis primarily stems from the intricate and time-consuming process of setting up the model, involving tasks such as drawing accurate computer-aided design (CAD) models and creating conforming meshes.
- CAD computer-aided design
- DNNs deep neural networks
- PC A principal component analysis
- DNN deep belief networks
- DNN deep autoencoders
- CNN convolutional neural networks
- GANs generative adversarial networks
- U-Net CNN-based encoder-decoder models
- GANs have also been used for predicting the stress/strain fields when the domain geometry is given as the input [28, 29]
- RNNs Recurrent neural networks
- LSTM long short-term memory
- GRU gated recurrent units
- AI/ML models mentioned above were originally developed in other areas of research such as computer vision and natural language processing and then applied to various problems in solid/fluid mechanics.
- AI/ML models specifically developed for predicting the response of partial differential equations among which physics- informed neural networks (PINNs) [12, 10, 15] is one of the most successful techniques for modeling a wide array of problems.
- PINNs offer the ability to simulate complex phenomena using a small set of training data by embedding prior knowledge of physical laws (boundary conditions, stress-strain relationships, etc.) in the training process to enhance the data set and facilitate learning.
- the original FNO model and its variants have been implemented for predicting the response of a wide range of linear and nonlinear mechanics problems.
- Embodiments of the present disclosure provide a generalizable Al-driven modeling technique, coined Deep Learning-Driven Domain Decomposition (DLD 3 ), for simulating two-dimensional linear elasticity problems with arbitrary geometry and boundary conditions (BCs).
- DLD 3 Deep Learning-Driven Domain Decomposition
- a set of pre-trained AI/ML models capable of accurately predicting the displacement field in small subdomains of a larger problem with various geometries/BCs are deployed in the Schwarz overlapping domain decomposition method (DDM) to approximate the linear elastic response.
- DDM Schwarz overlapping domain decomposition method
- the training dataset is then constructed by extracting millions of subdomains (images) and their corresponding displacement fields and BCs, as well as material property fields (elastic modulus and Poisson’s ratio) from the simulation results.
- material property fields material property fields
- the DLD 3 method can predict the response of a wide array of problems with distinct geometries and BCs.
- a brief overview of the overlapping and non-overlapping DDM techniques is provided herein, where we also discuss the reasoning for using the former in the DLD 3 technique. We delve into the DLD 3 algorithm and provide a detailed discussion on constructing the training dataset and the subsequent training/performance of the FNO models used in this method.
- CAD computer- aided design
- VR virtual reality
- DLD 3 directly uses imaging data such as micro-CT and ultrasonic images to perform a simulation, which is highly suitable for applications such as modeling porosity defects in additively manufactured parts and structures subjected to corrosion attack.
- the DLD 3 algorithm has the potential to perform massive simulations without running into issues such as convergence difficulty or lack of memory observed in FE.
- Embodiments of the present disclosure solve various challenges including those associated with accurate approximation of the field near curved edges and material interfaces.
- the proposed system can be embodied as standalone software or potentially as an add-on for FEM software.
- DLD 3 can specifically be tuned for simulating the digital manufacturing of flexible parts such as wire harnesses in real time.
- existing digital manufacturing software packages use either coarse meshes or rely on alternative techniques such as reduced-order models, which essentially compromise the accuracy.
- the ability to simulate a physical phenomenon in real-time opens the door to using DLD 3 as the underlying simulation engine in augmented/virtual reality software, where the simulations are often not physics-based.
- DLD 3 can significantly enhance the predictive maintenance (e.g., of the aging Air Force fleet) based on non-destructive inspection data such as imaging via automated physics-based simulations.
- outputs generated via the DLD 3 methods described herein can be used to generate 2D or 3D models or objects that a user can manipulate directly in a CAD design, simulation, or VR environment (for example, to apply forces to the generated object in the context of solving an engineering design problem).
- Embodiments of the present disclosure can be variously employed in structural analysis (e.g., buildings, aircraft, mechanical components) including vibrational and acoustic analysis, fluid dynamics and computational fluid dynamics, and electronic device design (e.g., motors, circuits, transformers) to predict the performance of objects and as part of product development and testing.
- DLD 3 can be used to simulate material response using raw imaging data as the only input which highly reduces the operational costs and computational costs of such systems.
- Computing device 100 may be generally configured to implement or execute the various processes and methods described herein.
- Computing device 100 may be any suitable computing device, such as a desktop computer, a laptop computer, a server, a workstation, a smartphone, or the like.
- Computing device 100 generally includes a processing circuit 102 that includes a processor 104 and a memory 110.
- Processor 104 can be a general- purpose processor, an ASIC, one or more FPGAs, a group of processing components, or other suitable electronic processing structures.
- processor 104 is configured to execute program code stored on memory 110 to cause computing device 100 to perform one or more operations, as described below in greater detail.
- Memory 110 can include one or more devices (e.g., memory units, memory devices, storage devices, etc.) for storing data and/or computer code for completing and/or facilitating the various processes described in the present disclosure.
- memory 110 includes tangible (e.g., non-transitory), computer-readable media that store code or instructions executable by processor 104.
- Tangible, computer-readable media refers to any physical media that can provide data that causes computing device 100 to operate in a particular fashion.
- Example tangible, computer-readable media may include, but is not limited to, volatile media, non-volatile media, removable media, and non-removable media implemented in any method or technology for storage of information such as computer- readable instructions, data structures, program modules, or other data.
- memory 110 can include RAM, ROM, hard drive storage, temporary storage, non-volatile memory, flash memory, optical memory, or any other suitable memory for storing software objects and/or computer instructions.
- Memory 110 can include database components, object code components, script components, or any other type of information structure for supporting the various activities and information structures described in the present disclosure.
- Memory 110 can be communicably connected to processor 104, such as via processing circuit 102, and can include computer code for executing (e.g., by processor 104) one or more processes described herein.
- processor 104 and/or memory 110 can be implemented using a variety of different types and quantities of processors and memory.
- processor 104 may represent a single processing device or multiple processing devices.
- memory 110 may represent a single memory device or multiple memory devices.
- computing device 100 may be implemented within a single computing device (e.g., one server, one housing, etc.). In other embodiments, computing device 100 may be distributed across multiple servers or computers (e.g., that can exist in distributed locations). For example, computing device 100 may include multiple distributed computing devices (e.g., multiple processors and/or memory devices) in communication with each other that collaborate to perform operations.
- an application may be partitioned in such a way as to permit concurrent and/or parallel processing of the instructions of the application.
- the data processed by the application may be partitioned in such a way as to permit concurrent and/or parallel processing of different portions of a data set by two or more computers.
- virtualization software may be employed by computing device 100 to provide the functionality of a number of servers that is not directly bound to the number of computers in computing device 100.
- Memory 110 is shown to include a domain decomposition (DDM) engine 112 configured to decompose digital models (e.g., 2D or 3D representations) of physical objects into a plurality of subdomains.
- DDM domain decomposition
- Memory 110 is also shown to include a deep learning (DL) engine 114 configured to solve each subdomain generated by DDM engine 112. More generally, as described herein, DDM engine 112 and DL engine 114 may operate cooperatively to decompose a digital model into subdomains, solve each subdomain, and iteratively update boundary conditions for each subdomain.
- DDM engine 112 and/or DL engine 114 may be configured to predict a linear-elastic response for each subdomain, for example, displacement and/or stress fields from the solved subdomains and/or updated boundary conditions.
- DDM engine 112 may be configured to implement an overlapping DDM technique to generate a plurality of overlapping subdomains.
- the overlapping DDM technique is the overlapping Schwarz method.
- DDM engine 112 generates a plurality of subdomains that overlap by 50%.
- DL engine 114 may solve each subdomain (e.g., as partial differential equation problems) to predict horizontal and vertical displacements (or the whole displacement field). In the case of a 50% overlap of each subdomain, DL engine 114 may predict midline horizontal and vertical displacements. The predicted displacements can then be used to update boundary conditions for each of the subdomains and/or to predict a displacement and/or stress field.
- DL engine 114 solves each subdomain using a single, pre-trained deep learning model.
- the deep learning model may be a convolution neural network or a Fourier Neural Operator; however, it will be appreciated that other suitable types of deep learning models can be used.
- the deep learning model is stored in a database 116 of memory 110.
- DL engine 114 is configured to train the deep learning model to solve a plurality of different types of subdomains.
- the deep learning model may be trained by a remote computing device and later transmitted to or retrieved by DL engine 114 for use. In either case, the deep learning model may be trained to predict the response of a variety of different types of subdomains using a training data set.
- DL engine 114 solves for each subdomain using a dedicated or separate deep learning model.
- a plurality of deep learning models can be trained, each to solve a different type of subdomain.
- DL engine 114 or a remote device can train the plurality of deep learning models using respective training data sets. For example, multiple different neural networks (NNs), CNNs, FNOs, or other suitable models may be trained using respective training data sets.
- each training data set is associated with a particular type of subdomain problem.
- training data may be generated using similar methods.
- each training data set includes the subdomain geometry, material properties, displacement/traction boundary conditions (input data), and horizontal and vertical displacements or the whole displacement field (output data or label data) for a respective type of subdomain.
- each of the training data sets includes the subdomain geometry, material properties, displacements along planes parallel to X-Y, X-Z, and Y-Z planes (or the whole displacement field) and displacement/traction boundary conditions for a respective unique type of subdomain.
- the training data sets are generated using FE-DDM simulations.
- the training data sets are generated using finite element (FE) simulations to solve each of the plurality of different types of subdomains.
- FE finite element
- a large number of FE simulations may be conducted on multiple subdomains with various geometries and boundary conditions.
- the solution fields described herein are not limited to displacement fields for linear elastic problems and can include or refer to other data types and solution fields for various problems including non-linear problems.
- this disclosure contemplates that the system and methods described herein can be used to determine temperature fields for heat transfer problems and electromagnetic fields for electrical problems.
- training each deep learning model includes: providing, to a deep learning model that is being trained, boundary conditions at multiple boundary points (e.g., 80 points, 21 along each boundary in 2D) for a subdomain.
- the boundary conditions relate to a specific type of subdomain that the deep learning model is being trained to evaluate.
- the deep learning model then predicts displacements along the horizontal and vertical midlines (or the whole displacement field) of the subdomain. A polynomial curve can then be fitted to the resulting midline displacement values to extrapolate points for updating the adjacent subdomain boundary conditions during DDM iterations.
- DL engine 114 is further configured to identify one of the plurality of deep learning models that is applicable to each of the plurality of subdomains prior to independently solving each subdomain. For example, multiple trained deep learning models may be stored in database 116 such that DL engine 114 can identify a type of each subdomain, retrieve the appropriate model(s) from database 116, and solve each subdomain with the appropriate deep learning model. In some implementations, DL engine 114 can further execute a model to predict a stress field and/or displacement field based on the predicted horizontal and vertical displacements (or the whole displacement field) for each subdomain.
- the term “artificial intelligence” is defined herein to include any technique that enables one or more computing devices or comping systems (i.e., a machine) to mimic human intelligence.
- Al includes but is not limited to, knowledge bases, machine learning, representation learning, and deep learning.
- machine learning is defined herein to be a subset of Al that enables a machine to acquire knowledge by extracting patterns from raw data.
- Machine learning techniques include, but are not limited to, logistic regression, support vector machines (SVMs), decision trees, Naive Bayes classifiers, and artificial neural networks.
- representation learning is defined herein to be a subset of machine learning that enables a machine to automatically discover representations needed for feature detection, prediction, or classification from raw data.
- Representation learning techniques include, but are not limited to, autoencoders.
- the term “deep learning” is defined herein to be a subset of machine learning that that enables a machine to automatically discover representations needed for feature detection, prediction, classification, etc. using layers of processing. Deep learning techniques include, but are not limited to, artificial neural networks or multilayer perceptron (MLP).
- MLP multilayer perceptron
- Machine learning models include supervised, semi -supervised, and unsupervised learning models.
- a supervised learning model the model learns a function that maps an input (also known as feature or features) to an output (also known as target or targets) during training with a labeled data set (or dataset).
- an unsupervised learning model the model learns patterns (e.g., structure, distribution, etc.) within an unlabeled data set.
- a semi- supervised model the model learns a function that maps an input (also known as feature or features) to an output (also known as target or target) during training with both labeled and unlabeled data.
- An artificial neural network is a computing system including a plurality of interconnected neurons (e.g., also referred to as “nodes”).
- the nodes can be implemented using a computing device (e.g., a processing unit and memory as described herein).
- the nodes can be arranged in a plurality of layers such as an input layer, output layer, and optionally one or more hidden layers.
- An ANN having hidden layers can be referred to as a deep neural network or multilayer perceptron (MLP).
- MLP multilayer perceptron
- Each node is connected to one or more other nodes in the ANN.
- each layer is made of a plurality of nodes, where each node is connected to all nodes in the previous layer.
- nodes in a given layer are not interconnected with one another, i.e., the nodes in a given layer function independently of one another.
- nodes in the input layer receive data from outside of the ANN
- nodes in the hidden layer(s) modify the data between the input and output layers
- nodes in the output layer provide the results.
- Each node is configured to receive an input, implement an activation function (e.g., binary step, linear, sigmoid, tanh, or rectified linear unit (ReLU) function), and provide an output in accordance with the activation function. Additionally, each node is associated with a respective weight.
- ANNs are trained with a dataset to maximize or minimize an objective function.
- the objective function is a cost function, which is a measure of the ANN’S performance (e.g., an error such as LI or L2 loss) during training, and the training algorithm tunes the node weights and/or bias to minimize the cost function.
- a cost function which is a measure of the ANN’S performance (e.g., an error such as LI or L2 loss) during training
- the training algorithm tunes the node weights and/or bias to minimize the cost function.
- any algorithm that finds the maximum or minimum of the objective function can be used for training the ANN. Training algorithms for ANNs include, but are not limited to, backpropagation.
- a CNN is a type of deep neural network that has been applied, for example, to image analysis applications. Unlike traditional neural networks, each layer in a CNN has a plurality of nodes arranged in three dimensions (width, height, and depth). CNNs can include different types of layers, e.g., convolutional, pooling, and fully-connected (also referred to herein as “dense”) layers.
- a convolutional layer includes a set of filters and performs the bulk of the computations.
- a pooling layer is optionally inserted between convolutional layers to reduce the computational power and/or control overfitting (e.g., by downsampling).
- a fully connected layer includes neurons, where each neuron is connected to all of the neurons in the previous layer. The layers are stacked similar to traditional neural networks.
- computing device 100 further includes a user interface 120 that enables a user to interact with computing device 100.
- user interface 120 may include a screen, such as an LED or LCD screen, that can display data, text, and other graphical elements to a user.
- user interface 120 may include a screen on which digital models of objects are displayed.
- user interface 120 includes one or more user input devices, such as a mouse, a keyboard, a joystick, a number pad, a drawing pad or digital pen, a touch screen, and the like. These user input device(s) are generally configured to be manipulated by a user to input data to, or to otherwise interact with computing device 100.
- computing device 100 may include a mouse and/or keyboard that a user can use to input, retrieve, and/or manipulate digital models.
- Computing device 100 is also shown to include a communications interface 130 that facilitates communications between computing device 100 and any external components or devices.
- communications interface 130 can facilitate communications between computing device 100 and a back-end server or other remote computing device.
- communications interface 130 also facilitates communications to a plurality of external user devices.
- communications interface 130 can facilitate communications between communications interface 130 can be or can include a wired or wireless communications interface (e.g., jacks, antennas, transmitters, receivers, transceivers, wire terminals, etc.) for conducting data communications, or a combination of wired and wireless communication interfaces.
- wired or wireless communications interface e.g., jacks, antennas, transmitters, receivers, transceivers, wire terminals, etc.
- communications via communications interface 130 are direct (e.g., local wired or wireless communications) or via a network (e.g., a WAN, the Internet, a cellular network, etc.).
- a network e.g., a WAN, the Internet, a cellular network, etc.
- communications interface 130 may include one or more Ethernet ports for communicably coupling computing device 100 to a network (e.g., the Internet).
- communications interface 130 can include a Wi-Fi transceiver for communicating via a wireless communications network.
- communications interface 130 may include cellular or mobile phone communications transceivers.
- process 200 for simulating a physical behavior (e.g., mechanical response, thermal response, chemical response, etc.) of an object from a digital model is shown, according to some implementations.
- process 200 can be used to predict the linear elastic response of an object; however, it should be appreciated that process 200 can more generally be applied to any physical response problem, for example, mechanical, thermal, chemical, etc.
- process 200 is implemented by computing device 100, as described above. However, it should be understood that process 200 can, more generally, be implemented or executed by any suitable computing device. It will be appreciated that certain steps of process 200 may be optional and, in some implementations, process 200 may be implemented using less than all the steps. It will also be appreciated that the order of steps shown in FIG. 2 is not intended to be limiting.
- a digital model of a physical object is obtained.
- the digital model can be either a 2D or 3D model of the physical object.
- the digital model is retrieved from a database, downloaded from a media device (e.g., a portable memory device), or otherwise obtained by a user of computing device 100.
- the digital model can be created and/or stored locally on computing device 100.
- the digital model is decomposed into a plurality of overlapping subdomains.
- a DDM model may be applied to the digital model to generate the subdomains.
- the DDM model implements an overlapping Schwarz method to decompose the digital model.
- the plurality of subdomains overlap by 50%, which can enhance the DDM solver convergence and facilitate training the DL model; however, it should be appreciated that in various other implementations, the subdomains may overlap by any amount (e.g., 10%, 20%, 30%, 40%, 50%, 60%, 70%, 80% etc.).
- non-overlapping DDM when solving multiphase problems (similar to fluid-solid interaction or modeling a two-phase composite material).
- each subdomain is solved using a suitable deep learning model to predict vertical and horizontal displacements or the whole displacement field (e.g., for a 2D subdomain).
- a suitable deep learning model to predict vertical and horizontal displacements or the whole displacement field (e.g., for a 2D subdomain).
- a single deep learning model e.g., a CNN
- a plurality of different deep learning models are trained, each to predict displacement values for a particular type of subdomain.
- the deep learning model(s) may be configured to directly predict midline horizontal and vertical displacements or the whole displacement field and then extract midline horizontal and vertical displacements.
- boundary conditions for each of the subdomains are iteratively updated based on the predicted displacements.
- the boundary conditions along each subdomain boundary may be iteratively updated to enforce the continuity of forces/displacements.
- the boundary conditions are iteratively updated until convergence is achieved (e.g., via the Schwarz DDM).
- a displacement field and/or a stress field are predicted.
- a displacement and/or stress field can be predicted based on the updated boundary conditions from step 208, for example, after convergence.
- the displacement field is predicted based on the updated boundary conditions and geometry of each subdomain.
- these data points e.g., the updated boundary conditions and geometry of each subdomain
- the deep learning model is different from the deep learning model(s) used to predict the horizontal and vertical displacements (midline horizontal and vertical displacements).
- the deep learning model for predicting the displacement field may be a different type of deep learning model and/or may be trained on a different data set to make different predictions (e.g., predictions of a displacement field instead of predictions of horizontal and vertical displacements).
- the stress field is predicted from the displacement field.
- a geometry of each subdomain is provided as input to a deep learning model for predicting the stress field.
- hierarchical deep learning predictions are initiated to recursively evaluate the displacement along the midline of each subdomain, and then on the midlines of its quarter subdomains, etc., until the full displacement field is determined.
- the deep learning model is trained to predict an entire displacement field.
- the displacement field is simply predicted/visualized for all subdomains.
- a stress field can be recovered from the resulting displacement values.
- a second deep learning model is trained to predict the stress field from the boundary displacements and/or the displacement field.
- output data is generated for the physical object, for example, based at least in part on the predicted displacement field over the entire domain and/or stress field.
- the generated output data can be employed in a computer-aided design system, simulation, or virtual reality system.
- the output data is used to generate user interface data (e.g., models, image data, simulation data, renderings, or the like) depicting the predicted displacement field and/or stress field.
- the output data can include new training data for machine learning model(s).
- the DLD 3 technique relies on the utilization of DDM to break down a large domain into smaller subdomains, the response of which can accurately be predicted using a pretrained ML model, to approximate the displacement field.
- DDM dynamic light-to-distance domain
- the response of which can accurately be predicted using a pretrained ML model, to approximate the displacement field.
- FIGS. 3A-3B illustrate domain partitioning and updating BCs of a subdomain in non-overlapping (FIG. 3 A) and overlapping DDM (FIG. 3B) based on the field approximated in neighboring subdomains.
- the core idea of both the non-overlapping and overlapping DDM is to approximate the field in each subdomain independently (e.g., using FEM) and use this solution to update BCs of neighboring subdomains. This process is iteratively continued until the continuity of the field (and tractions in the non-overlapping DDM) is satisfied over the entire domain.
- the first step of the modeling process for a DDM simulation is to subdivide the domain into smaller subdomains, as shown in FIGS. 3A-3B.
- the field approximated in each subdomain (T) subject to initial BCs is utilized to update its neighboring subdomain BCs using the fixed- point iteration (FPI) algorithm.
- FPI fixed- point iteration
- a structured partitioning scheme c/ FIG. 3 A
- Dirichlet (displacement) BC is enforced along the left and bottom edges of each subdomain, while the right and top edges have Neumann (traction) BC.
- the iterative FPI process begins by approximating the field in subdomain (T) and using this solution to update the Dirichlet BCs along the left edge of @ and the bottom edge of (5).
- u n and t n are the nodal vectors of displacement and traction BCs along subdomain edges at iteration n.
- the overlapping Schwarz method has a more straightforward algorithm that only involves updated Dirichlet BCs of subdomain edges based on the field approximated in its neighboring overlapping subdomains. Therefore, in an overlapping DDM simulation, we no longer require recovering stresses within a subdomain or the tractions along its edges, which not only facilitates the implementation but also improves the accuracy due to the higher error involved in the recovery of gradients (stresses/tractions) vs the field. Further, this algorithm does not require using an underrelaxation approach and therefore determining the appropriate value of P for updating BCs.
- the letters “L”, “B”, “R”, and “T” in the superscripts refer to “Left”, “Bottom”, “Right”, and “Top” regions of the subdomain boundary, as labeled in FIG. 3B.
- the BCs along different regions of the subdomain @ edges are updated based on the order in which the subdomains are being visited to approximate the field in them (from (T) to @) to use the most up-to-date field in this process. For example, although the bottom left (BL) portion of the subdomain @ edge overlaps with both (T) and @ , since the field in the latter is simulated last, only the approximated field in @ is utilized to update this portion of the
- FIGS. 4A-4C illustrate the approximation of the field in a porous domain using both the non-overlapping and overlapping DDM techniques, together with the overlapping FE- DDM approximation of the stress field using 3 x 3 subdomains with 50% overlap and the corresponding error vs an FE simulation conducted on the entire field (direct numerical simulation).
- FIG. 4A shows domain geometry and applied boundary condition
- FIG. 4B shows overlapping FE-DDM approximation of the normal stress field in the y- direction
- FIG. 4C shows the corresponding error vs a direct numerical simulation using 9 subdomains with 50% overlap.
- a high overlap of 50% not only significantly expedites the overlapping DDM convergence but also reduces the total computational cost.
- using a higher overlap percentage significantly reduces the number of iterations needed for the overlapping DDM convergence. This significant reduction is more than enough to offset the increased time associated with approximating the field in each subdomain, resulting in a notable drop in the overall simulation time.
- FIGS. 5A-5B are graphs illustrating the effect of the number of subdomains and the overlap percentage between neighboring subdomains in the overlapping DDM on the number of iterations (FIG. 5 A) and the overall computational costs (FIG. 5B).
- the total cost of a DLD 3 is proportional to the number of DDM iterations, which as shown in FIGS. 5A-5B, could considerably be reduced using subdomains with 50% overlap. While we could change this percentage depending on the domain dimensions, keeping that at 50% for all subdomains also facilitates the DLD 3 implementation, as the trained Al model only needs to predict the displacements along horizontal and vertical midlines of subdomains at each iteration.
- FIG. 6 illustrates using subdomains with 50% overlap 602 to discretize a domain in the DLD 3 method to implement the overlapping Schwarz method to approximate the field.
- an Al model 604 is trained to predict the displacement field along the horizontal and vertical midlines of each subdomain based on the Dirichlet BCs applied on the edges of this subdomain, i.e., u ⁇ , Un, Un> ⁇ n- As shown in FIG. 6, this characteristic highly facilitates subdividing a domain into overlapping subdomains, we first discretize the domain using a structured mesh and then merge all neighboring grids sharing a node to construct a subdomain.
- One challenge toward the successful implementation of the DLD 3 method is to construct the training dataset, and then select an appropriate Al model and train the Al model to accurately predict the field in each subdomain.
- this training dataset consists of the BCs applied along square-shaped subdomain edges and the corresponding displacement field within the domain (or simply the displacement along its horizontal and vertical midlines, i.e., u ⁇ 1 , n). While constructing such a dataset is a time-consuming process is a time-consuming process, it is a rather straightforward task that requires applying various BCs to a squareshaped domain discretized using a structured mesh and using FEM to approximate the field.
- FIG. 7 illustrates partitioning a domain with arbitrary geometry and BCs for the DLD 3 implementation.
- Various colors in the image on the right correspond to different types of subdomains depending on their geometry (with vs. without curved edges) and BCs (purely Dirichlet, Dirichlet-Neumann, etc.), where different DL models are used to predict their midline displacements during the DLD 3 simulation.
- subdomains are marked with different colors indicating their classification into different types depending on the geometric feature (with or without a curved edge) and BC type (Purely Dirichlet or Dirichlet-Neumann BCs, applied along straight or curved edges).
- BC type Purly Dirichlet or Dirichlet-Neumann BCs, applied along straight or curved edges.
- the orange subdomains shown in FIG. 7 are the same type previously shown in FIG. 6, while the cases shown in other colors represent other types of subdomains requiring the training of new Al models to predict the field during the DDM iterations, as schematically shown in FIGS. 12A-12D.
- FIG. 8 illustrates subdividing a domain into overlapping subdomains by merging the cells of a structured grid and leveraging its nodal connectivity to build the connectivity table between resulting subdomains. Note that, the grid cells do not need to conform to the curved or straight edges of the domain, meaning the process is even easier than generating a structured mesh in this case.
- any grid cell falling completely outside the domain boundaries is deleted.
- the main motivation behind initially using a structured grid to construct the overlapping subdomains is to leverage the nodal connectivity of this grid to subsequently build a connectivity table for the resulting subdomains.
- Such a connectivity table consists of identifying the overlapping subdomains and their relative location, i.e., BL, B, BR, RB, R, RT, etc., which is essential for updating BCs during the DDM iterations according to Equation (5).
- BL, B, BR, RB, R, RT i.e., BL, B, BR, RB, R, RT, etc.
- FIG. 8 we can easily identify neighboring subdomains overlapping with @ and their relative locations based on the location of cells in each subdomain. For example, the bottom left quadrant of @ is grid cell 14, which is the right quadrant of subdomain thus, the latter is identified as the subdomain overlapping with the bottom left quadrant of @.
- the FNO models used in the DLD 3 algorithm receive a subdomain geometry, its elastic moduli, and the applied BC as the input and predict the displacement field in this subdomain. Training such a model requires access to a diverse dataset with millions of entries encompassing various subdominant geometries (void volume fraction, different edge curvatures, etc.) and boundary conditions (Dirichlet vs Neumann). As noted previously, we opt out to break down this massive data set into several subsets, classified based on subdomains with no curved edge vs solid, square-shaped subdomains.
- the training dataset used for training the FNO models in this work is obtained from high-fidelity FE simulations conducted on fine conforming meshes.
- the key challenge is to ensure the diversity of entries in this dataset, encompassing various geometries and BCs for training the FNO models.
- constructing this dataset would not be feasible without the complete automation of the FE modeling process, i.e., reconstructing the geometrical models, constructing the conforming meshes, performing the FE simulation, and extracting the final labeled data (input: subdomain geometry + BC, output: displacement field).
- FIG. 9A shows an example of a virtually reconstructed geometrical model with embedded voids of various shapes, together with the Dirichlet BCs applied along the domain edges for the FE analysis.
- FIG. 9B depicts a small portion of the conforming mesh generated using CISAMR, corresponding to the inbox shown in FIG. 9A.
- the construction of the training dataset begins with building a shape library of more than 100 inclusions, representing voids with various shapes and sizes, involving different curvatures (concave vs convex).
- the morphology of voids in this shape library is represented using Non-Uniform Rational B-Splines (NURBS) [60], which facilitate the subsequent mesh generation when these voids are used to build a geometrical model for FE simulation.
- NURBS Non-Uniform Rational B-Splines
- the virtual packing algorithm introduced in [46] is then utilized to reconstruct a geometrical model by virtually embedding tens of these inclusions in a square-shaped domain, as in FIG. 9A of dimensions 100 pm * 100 pm.
- BBoxes hierarchical bounding boxes
- CISAMR Conforming to Interface Structured Adaptive Mesh Refinement
- the CISAMR technique implements a set of non-iterative operations involving h-adaptivity, r- adaptivity, element deletion, and element subdivision to transform a structured grid into a conforming mesh with low element aspect ratios.
- the algorithm ensures the element aspect ratios do not exceed 3, ensuring the construction of a high-quality mesh.
- FIG. 9B illustrates a small portion of the fine-conforming mesh generated using CISAMR for the porous domain shown in FIG. 9A.
- the conforming mesh generated for each geometrical model is then utilized to simulate its linear elastic FE response subjected to 15 different BCs.
- the BCs Dirichlet or Neumann
- the BCs are considered to be a fifth-order polynomial with arbitrary coefficients.
- E IGPa
- FIGS. 10A-10B illustrate the FE approximations of the displacement and strain fields in the domain shown in FIG.
- FIG. 10A shows the approximation of displacement and FIG. 10B shows the approximation of strain fields in the y-direction for the domain and the corresponding CIS AMR mesh shown in FIGS. 9A-9B.
- 2000 subdomains and the corresponding BCs and displacement fields are extracted from each domain to collect a total of 30M labeled data.
- the 2000 subdomains were chosen in a manner that ensures an equal distribution of subdomains with single void (i.e., subdomains intersecting only with a single void), multiple voids (i.e., subdomains intersecting with at most three voids), and no voids (i.e., subdomains completely inside the solid region).
- FIG. 11 illustrates extracting random subdomain geometries and the corresponding field/BC as entries into the training dataset.
- the subdomains are selected by cropping a small square sub-region of the domain at a random location and orientation. Note that the subdomains could have different sizes due to the linear elastic nature of the problem being analyzed, as the resulting field/BC is scalable within the subdomain.
- the subdomain sizes are selected within a range that intersects with at most three voids. Expanding the subdomain size beyond this range requires a higher-resolution (pixelated) representation of the geometry and field, resulting in a highly complex problem that requires many more million data entries for training the FNO models to achieve acceptable accuracy.
- the material properties (E and v) and the approximated displacement values (ux and uy) are extracted at the points of a 83 x 83 grid (total of 6889 points) to be used as a training data entry.
- To evaluate the displacement vector at each grid point we must first identify which element among > 3 million mesh elements holds this point, determine its local coordinate ⁇ p in that element, and then interpolate the field at this point.
- the massive volume of training data that must be extracted > 30 million
- the final data set consists of an equal number of subdomains intersecting with one, two, and three interfaces, as well as a uniform distribution of void volume fraction, V vo t d - For example, the volume number of subdomains with 0% ⁇ V void ⁇ 5% and those with 95% ⁇ V void ⁇ 100% is approximately the same in this dataset.
- FIGS. 12A-12D illustrate four different FNO models trained based on the subdomain geometry and applied BC.
- FIG. 12A shows a solid subdomain with Dirichlet and Neuman BC
- FIG. 12B shows a porous subdomain with Dirichlet BC along its straight edges
- FIG. 12C shows a porous subdomain with Dirichlet BC along its straight and curved edges
- FIG. 12D shows a porous subdomain with Dirichlet and Neumann BC.
- FIGS. 12A-12D four other case scenarios for training different FNO models based on the shape/BC of the subdomain are illustrated in FIGS. 12A-12D. Note that the subdomains subjected to pure Dirichlet BCs along their straight edges (cf. FIG. 6 and FIG. 12B) are the most important cases, as all interior subdomains of a DLD 3 model fall into this category.
- FIG. 13 is a flowchart diagram illustrating a method for extracting random subdomain geometries and the corresponding field/BC as entries into the training dataset.
- the input data f x) 1302 are first fed into a neural network lifting layer 1304, then fed into four Fourier layers (1306a, 1306b, 1306c, 1306d), and finally, another neural network (projection layer 1308) is deployed to map that into the output data u(x) 1310.
- the array containing the BC values has the corresponding BC values only at the boundary points and is padded in the interior with a chosen value of 3, which is outside the range of applied Dirichlet or Neumann BCs that is [-2, 2],
- the output u(x) is the displacement vector at all points of this array.
- the neural operator constructs an operator G e that learns an approximation of G by minimizing the following problem using a cost function C as [0125]
- a cost function C a cost function ⁇ [0125]
- f x)j refers to the input variables (E,v, and BC values)
- u(x) 7 refers to the predicted displacement field.
- the architecture of an FNO model consists of three main components [13]:
- GLU Gaussian error linear unit
- the kernel integral operator in (8) can be replaced by a convolution operator defined in the Fourier space as
- FIG. 14 is a graph depicting training and validation losses plotted against the number of epochs during the training of the FNO model.
- the training and validation losses were reduced to 2.4e-04 and 2.6e-04, respectively, where the latter is identical to the test loss (the test and validation datasets each have 150k entries).
- the similarity of test, validation, and training losses of the trained FNO model indicates no overfitting and thereby proper generalizability to predict the field in subdomains with various geometries, BCs, and material properties when utilized in the DLD 3 framework.
- FIG. 15 shows the FNO prediction of the magnitude of the displacement field in several subdomains (first row) and the corresponding distribution of the error vs FE simulation of the field using conforming meshes (second row).
- FIG. 15 shows the prediction of the field in several subdomains with various geometries and BCs using the trained FNO model, together with the corresponding error vs FE simulation results conducted on conforming meshes, i.e., the ground truth used for training the model.
- Example 1 Simple porous domain
- FIG. 16A illustrates SVE geometry and applied BC.
- FIG. 16B illustrates 11 x5 overlapping subdomains used for partitioning the domain to perform the DLD 3 simulation, where one of the subdomains is highlighted in red (1602).
- the domain has fixed BC along its bottom and ledge edges, while the following linear displacement BCs are applied along its top and right edges (The origin of the coordinate system in all examples is the bottom left corner of the domain):
- FIGS. 17A-17C illustrate an example simple porous domain problem where FIG. 17A shows results from applying DLD 3 , FIG. 17B shows FE approximations of the displacement field magnitude, and FIG. 17C shows the relative error between DLD 3 and FEM results.
- the DLD 3 simulation is carried out by partitioning the domain into 11 x5 overlapping subdomains (50% overlap), as shown in FIG. 16B.
- the DLD 3 approximation of the displacement field magnitude and its comparison with an FE simulation carried out on a fine, conforming mesh is illustrated in FIG. 17A and FIG. 17B, respectively.
- the field is initialized using a simple linear field corresponding to a domain with no porosity, resulting in 59 DDM iterations to achieve convergence.
- error
- is illustrated in FIG. 17C.
- This example shows the proposed DLD 3 yields a good accuracy for approximating the field in this problem after breaking down the original domain into smaller subdomains the response of which can accurately be predicted using the pre-trained FNO models.
- FIGS. 18A-18B illustrate a second example problem where FIG. 18A shows domain geometry and applied BC and FIG. 18B shows 217 overlapping subdomains extracted from a 19 x 19 partitioning of the domain to perform the DLD 3 simulation, where one of the subdomains is highlighted in red (1802).
- the bottom and right edges of the domain have fixed displacement BC, while the following quadratic displacement BCs are applied along the top and left edges: 0,45, ⁇ 12 ⁇ ⁇ 13 ⁇
- the domain is originally partitioned using 19 x 19 overlapping subdomains.
- the subdomains falling outside the domain in the bottom right regions are then removed, meaning only 217 subdomains are used for approximating the field, as shown in FIG. 18B.
- the subdomains used for partitioning the domain do not need to conform to the domain boundaries.
- FIGS. 19A-19C The resulting DLD 3 approximation of the displacement field, its comparison with FE results, and their relative error are shown in FIGS. 19A-19C.
- FIG. 19A shows results from applying DLD 3
- FIG. 19B shows FE approximations of the displacement field magnitude
- FIG. 19C shows relative error between DLD 3 and FEM results.
- Example 3 Porous aluminum microstructure
- FIGS. 20A-20B illustrate a third example problem where FIG. 20A shows porous aluminum SVE geometry and applied BC and FIG. 20B shows overlapping subdomains used for partitioning the domain to perform the DLD 3 simulation, where one of the subdomains is highlighted in red (2002).
- the SVE was fixed displacement BC along its left and bottom edges, whereas the following linear displacement BCs were applied along its top and right edges:
- FIGS. 21 A-21B show results from the third example problem.
- FIG. 21 A shows results from applying DLD 3
- FIG. 2 IB shows FE approximations of the displacement field magnitude.
- the DLD 3 simulation was carried out on 910 overlapping subdomains (35 * 26 partitioning), as shown in FIG. 21 A.
- the field was initialized using a linear displacement field corresponding to a domain with no porosity and the DDM solver converged after 705 iterations.
- the resulting DLD 3 approximation of the field, together with its comparison with FE simulation results, are illustrated in FIG. 21B.
- Embodiments of the present disclosure provide a generalizable Al-driven framework coined Deep Learning-Driven Domain Decomposition (DLD 3 ).
- DLD 3 was introduced as a surrogate for FEM to approximate the linear elastic response of two- dimensional problems with arbitrary geometry and BC.
- This method relies on a set of pretrained Al models (selected as the FNO model in this work) for predicting the displacement field (u x and Uy) in square-shaped subdomains of a larger domain, taking the geometry, BC, and material properties (E, v) as input.
- the pre-trained FNO models are then combined with the Schwarz overlapping domain decomposition method (DDM), enforcing 50% overlap between adjacent subdomains, to predict the field in the original domain by iteratively updating the subdomains BCs.
- DDM Schwarz overlapping domain decomposition method
- Several example problems were presented to show the versatility of the DLD 3 technique to accurately simulate the displacement field in singlematerial domains with various geometries and subject to different BCs.
- Our ongoing and future efforts entail extending the DLD 3 method to multi-material and three-dimensional elasticity problems, as well as other constitutive models such as transient diffusion and plasticity. While the overall algorithm remains unchanged, a large dataset and substantial training power are required to ensure the pre-trained Al models used for predicting the field in subdomains of such problems are truly generalizable, regardless of the domain geometry and applied BC.
- the present disclosure contemplates methods, systems, and program products on any machine-readable media for accomplishing various operations.
- the implementations of the present disclosure may be implemented using existing computer processors, or by a special purpose computer processor for an appropriate system, incorporated for this or another purpose, or by a hardwired system.
- Implementations within the scope of the present disclosure include program products including machine-readable media for carrying or having machine-executable instructions or data structures stored thereon.
- Such machine- readable media can be any available media that can be accessed by a general-purpose or special-purpose computer or other machines with a processor.
- machine-readable media can comprise RAM, ROM, EPROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code in the form of machineexecutable instructions or data structures, and which can be accessed by a general purpose or special purpose computer or other machines with a processor.
- Machine-executable instructions include, for example, instructions and data that cause a general-purpose computer, special-purpose computer, or special-purpose processing machine to perform a certain function or group of functions.
- F. Feyel A multilevel finite element method (FE2) to describe the response of highly non-linear structures using generalized continua, Computer Methods in applied Mechanics and engineering 192 (2003) 3233-3244.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- General Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computer Hardware Design (AREA)
- Geometry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Medical Informatics (AREA)
- Algebra (AREA)
- Computational Mathematics (AREA)
- Mathematical Analysis (AREA)
- Mathematical Optimization (AREA)
- Pure & Applied Mathematics (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202363495602P | 2023-04-12 | 2023-04-12 | |
| US63/495,602 | 2023-04-12 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2024215873A1 true WO2024215873A1 (en) | 2024-10-17 |
Family
ID=93060119
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/US2024/024050 Pending WO2024215873A1 (en) | 2023-04-12 | 2024-04-11 | System and methods for simulating the physical behavior of objects |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2024215873A1 (en) |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20200100871A1 (en) * | 2018-09-27 | 2020-04-02 | Align Technology, Inc. | Aligner damage prediction and mitigation |
| US20210049757A1 (en) * | 2019-08-14 | 2021-02-18 | Nvidia Corporation | Neural network for image registration and image segmentation trained using a registration simulator |
| US20210357555A1 (en) * | 2018-09-14 | 2021-11-18 | Northwestern University | Data-driven representation and clustering discretization method and system for design optimization and/or performance prediction of material systems and applications of same |
| WO2022235261A1 (en) * | 2021-05-04 | 2022-11-10 | Hewlett-Packard Development Company, L.P. | Object sintering states |
-
2024
- 2024-04-11 WO PCT/US2024/024050 patent/WO2024215873A1/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20210357555A1 (en) * | 2018-09-14 | 2021-11-18 | Northwestern University | Data-driven representation and clustering discretization method and system for design optimization and/or performance prediction of material systems and applications of same |
| US20200100871A1 (en) * | 2018-09-27 | 2020-04-02 | Align Technology, Inc. | Aligner damage prediction and mitigation |
| US20210049757A1 (en) * | 2019-08-14 | 2021-02-18 | Nvidia Corporation | Neural network for image registration and image segmentation trained using a registration simulator |
| WO2022235261A1 (en) * | 2021-05-04 | 2022-11-10 | Hewlett-Packard Development Company, L.P. | Object sintering states |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Zheng et al. | Deep learning in mechanical metamaterials: from prediction and generation to inverse design | |
| Li et al. | Non-iterative structural topology optimization using deep learning | |
| Sudakov et al. | Driving digital rock towards machine learning: Predicting permeability with gradient boosting and deep neural networks | |
| Lu et al. | DeepXDE: A deep learning library for solving differential equations | |
| Maurizi et al. | Predicting stress, strain and deformation fields in materials and structures with graph neural networks | |
| Oh et al. | A tutorial on quantum convolutional neural networks (QCNN) | |
| Yang et al. | Microstructural materials design via deep adversarial learning methodology | |
| Guirguis et al. | Evolutionary black-box topology optimization: Challenges and promises | |
| Li et al. | A deep adversarial learning methodology for designing microstructural material systems | |
| Shi et al. | Gnn-surrogate: A hierarchical and adaptive graph neural network for parameter space exploration of unstructured-mesh ocean simulations | |
| Mosavi et al. | Learning and intelligent optimization for material design innovation | |
| Qiu et al. | A deep learning approach for efficient topology optimization based on the element removal strategy | |
| Luo et al. | Physics-informed neural networks for PDE problems: A comprehensive review | |
| Sharpe et al. | Topology design with conditional generative adversarial networks | |
| KR20230065343A (en) | Physical Environment Simulation Using Mesh Representation and Graph Neural Networks | |
| US11501037B2 (en) | Microstructures using generative adversarial networks | |
| Kim et al. | A novel adaptive mesh refinement scheme for the simulation of phase‐field fracture using trimmed hexahedral meshes | |
| Chung et al. | Prediction of effective elastic moduli of rocks using graph neural networks | |
| Gunda | Accelerating Scientific Discovery With Machine Learning and HPC-Based Simulations | |
| Frankel et al. | Mesh-based graph convolutional neural networks for modeling materials with microstructure | |
| Zhao et al. | Review of empowering computer-aided engineering with artificial intelligence | |
| WO2024215873A1 (en) | System and methods for simulating the physical behavior of objects | |
| Viswanath et al. | Designing a TPMS metamaterial via deep learning and topology optimization | |
| Letov et al. | Volumetric cells: A framework for a bio-inspired geometric modelling method to support heterogeneous lattice structures | |
| Dittmer et al. | Selto: Sample-efficient learned topology optimization |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 24789440 Country of ref document: EP Kind code of ref document: A1 |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2024789440 Country of ref document: EP |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| ENP | Entry into the national phase |
Ref document number: 2024789440 Country of ref document: EP Effective date: 20251112 |
|
| ENP | Entry into the national phase |
Ref document number: 2024789440 Country of ref document: EP Effective date: 20251112 |
|
| ENP | Entry into the national phase |
Ref document number: 2024789440 Country of ref document: EP Effective date: 20251112 |
|
| ENP | Entry into the national phase |
Ref document number: 2024789440 Country of ref document: EP Effective date: 20251112 |
|
| ENP | Entry into the national phase |
Ref document number: 2024789440 Country of ref document: EP Effective date: 20251112 |