[go: up one dir, main page]

WO2024127318A1 - Débruitage de modèles de diffusion pour soins buccaux numériques - Google Patents

Débruitage de modèles de diffusion pour soins buccaux numériques Download PDF

Info

Publication number
WO2024127318A1
WO2024127318A1 PCT/IB2023/062713 IB2023062713W WO2024127318A1 WO 2024127318 A1 WO2024127318 A1 WO 2024127318A1 IB 2023062713 W IB2023062713 W IB 2023062713W WO 2024127318 A1 WO2024127318 A1 WO 2024127318A1
Authority
WO
WIPO (PCT)
Prior art keywords
tooth
oral care
representations
setups
mesh
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/IB2023/062713
Other languages
English (en)
Inventor
Jonathan D. Gandrud
Lois F. Duerst
Michael Starr
Mariah Sonja Pereira Penha
Donna J. STENBERG
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
3M Innovative Properties Co
Original Assignee
3M Innovative Properties Co
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 3M Innovative Properties Co filed Critical 3M Innovative Properties Co
Priority to EP23832829.8A priority Critical patent/EP4633528A1/fr
Priority to CN202380085976.0A priority patent/CN120322210A/zh
Publication of WO2024127318A1 publication Critical patent/WO2024127318A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/41Medical

Definitions

  • Patent Applications is incorporated herein by reference: 63/432,627; 63/366,492; 63/366,495; 63/352,850; 63/366,490; 63/366,494; 63/370,160; 63/366,507; 63/352,877; 63/366,514; 63/366,498; 63/366,514; 63/264,914; and 63/432,627.
  • This disclosure relates to configurations and training of denoising diffusion models (e.g., such as neural networks which have been trained for that purpose) to improve the accuracy and data precision of 3D oral care representations to be used dental or orthodontic treatments.
  • the present disclosure describes systems and techniques for training and using one or more denoising diffusion probabilistic models (e.g., neural networks such as U-Nets or autoencoders) to generate 3D oral care representations.
  • Denoising diffusion-based techniques are described for the placement of oral care articles in relation to one or more 3D representations of teeth.
  • Oral care articles which are to be placed may include: a 3D representation of a tooth (e.g., for the generation of an orthodontic setup), a dental restoration appliance component, oral care hardware (e.g., a lingual bracket, a labial bracket, an orthodontic attachment, a bite ramp, etc.), and the like.
  • denoising diffusion-based techniques are described for the generation of the geometry and/or structure of oral care articles based, at least in part, on one or more 3D representations of teeth.
  • Oral care articles which may be generated include: a dental restoration tooth design, a crown, a veneer, an archform, a clear tray aligner (CTA) trimline, and an appliance component (e.g., such as generated components for use in creating a dental restoration appliance) and the like.
  • CTA clear tray aligner
  • appliance component e.g., such as generated components for use in creating a dental restoration appliance
  • U-Nets are an example of models that may enable improvements to data precision.
  • Denoising diffusion models may be trained to automatically generate (or modify) 3D oral care representations, such as 3D point clouds (or other 3D representations described herein).
  • a denoising diffusion model may be trained to generate 3D polylines (e.g., for archforms and CTA trimlines), or sets of control points (e.g., control points through which a spline can be fitted for an archform).
  • a denoising diffusion model may be trained to predict an archform (e.g., taking dental arch and tooth data as input).
  • An archform may take the form of a surface, a 3D mesh, a 3D polyline or a set of control points (e.g., used to define a spline).
  • Such an archform may, in some instances, be given as an input to a setups prediction machine learning model, such as a setups prediction neural network.
  • a denoising diffusion model may be trained to label aspects of a 3D representation (or to generate one or more object masks which identify one or more objects within the input 3D representation), for use in segmentation or mesh cleanup of 3D oral care representations.
  • Techniques of this disclosure include methods of a generating a data structure for oral care treatment. Such methods may receive as input one or more attributes describing an intended output from a trained machine learning model. Such methods may generate one or more noisy representations of an intended output, and subsequently denoise the one or more noisy representations. The output of the method may include one or more generated denoised representations, which may be used to define one or more aspects of one or more digital oral care treatments.
  • a training dataset may be generated or refined by successively modifying one or more 3D oral care representations.
  • An untrained (or partially trained) machine learning model may be trained, at least in part, using the refined training dataset.
  • the resulting trained machine learning model may be deployed for clinical treatment of patients.
  • one or more representations of the training dataset may be modified, such as through the addition of noise.
  • one or more aspects of one or more representations of the training dataset may be encoded into a latent form before generating the one or more noisy representations.
  • the one or more attributes may include at least one of a real value, a categorical value, or a natural language text value.
  • the one or more attributes may include at least one of an oral care metric or an oral care parameter.
  • the trained machine learning model may include at least one neural network.
  • At least one neural network may include at least one encoder-decoder structure.
  • An encoder- decoder structure may comprise at least one encoder or at least one decoder.
  • Non-limiting examples of an encoder-decoder structure include a U-Net, a transformer, a pyramid encoder-decoder or an autoencoder, among others.
  • the denoising diffusion techniques may receive one or more 3D representations of the patient’s dentition (e.g., which may include at least one tooth). The one or more 3D representations of the patient’s dentition may be encoded into a latent form.
  • the one more denoised representations may include at least one of: one or more 3D oral care representations as defined herein, at least one label for a mesh element, at least one transform to place at least one tooth of a patient’s dentition into a setup pose, an appliance component, a trimline, an archform, a coordinate system, or a tooth restoration design (among others).
  • a label on a mesh element may be used to segment one or more 3D oral care representations (e.g., to segment at least one representation of a patient’s dentition).
  • a label may be used for the cleanup of a 3D oral care representation (e.g., the label may be used to modify one or more 3D oral care representations – such as a representation of a patient’s dentition).
  • a transform predicted using the techniques described herein may place a tooth in a setup pose.
  • a setup may correspond to a final setup.
  • a setup may correspond to an intermediate stage.
  • an appliance component e.g., that is placed or generated using a denoising diffusion model
  • the denoising diffusion techniques described herein may, in some instances, be practiced in combination.
  • denoising diffusion mesh cleanup may be performed on a dental arch from an intraoral scanner, after which denoising diffusion segmentation may be performed on that cleaned-up 3D representation of the patient’s dentition.
  • one or more of the segmented teeth may be identified for dental restorative treatment and subsequently provided to a denoising diffusion model for geometry generation (e.g., Geometry Generation Diffusion Model GGDM), to generate a dental restoration tooth design.
  • the segmented teeth of the dental arch may be provided to a Diffusion Setups model for setups prediction.
  • Techniques of this disclosure may generate data structures for oral care treatment using denoising diffusion probabilistic models.
  • the methods may receive oral care arguments (or oral care attributes) describing the intended output from a trained denoising diffusion model.
  • the methods may execute a forward pass to generate a series of increasingly noisy versions of each training example.
  • a denoising ML model e.g., a U-Net
  • U-Net a denoising ML model
  • the data structure may evolve into a 3D oral care representation which is suitable for use in treating the patient.
  • one or more denoised representations are generated, and based on these representations, one or more aspects of digital oral care treatments are automatically defined (e.g., a restoration design for a tooth is generated, or an orthodontic setup is generated, among examples).
  • the method may access a training dataset of 3D oral care representations and generate a refined training dataset by modifying representations of the training dataset (e.g., iteratively adding noise).
  • An untrained machine learning model may be trained using the refined training dataset, and the trained machine learning model is outputted based on the training.
  • the modifying of the representations of the training dataset involves adding noise to aspects of the representations.
  • the method encodes aspects of the representations into a latent form before generating the noisy representations.
  • the attributes describing the intended output can include real values, categorical values, or natural language text values. Additionally, the attributes can include oral care metrics or oral care parameters.
  • the trained machine learning model can include neural networks such as U-Nets, transformers, or autoencoder (or the like).
  • the method can receive 3D representations of a patient's dentition, and the denoised representations can include 3D oral care representations. These representations can be encoded into a latent form (e.g., when Stable Diffusion is used).
  • the methods may also be trained to label mesh elements and apply those mesh element labels to 3D representations of oral care data, to segment those 3D representations based on the labels.
  • the 3D oral care representations can include representations of a patient's dentition, among others described herein.
  • the denoised representations can include transforms (e.g., 4x4 transformation matrices, among others), which may be used to place a patient's tooth into setup poses (e.g., for final setups or intermediate stages).
  • the transforms may place other 3D representations of oral care data into poses which are suitable for oral care treatment.
  • Other 3D representations of oral care data include fixture model components, tooth restoration designs, appliance components, fixture model components, or archforms (or the like).
  • the denoised representations can include coordinate systems, or other 3D oral care representations described herein.
  • the disclosed methods utilize denoising diffusion probabilistic models to generate data structures for use in oral care treatment, allowing for automatic definition of various aspects of digital oral care treatments based on trained models and denoised representations.
  • methods of this disclosure may generate data structures for use in digital oral care treatment.
  • the methods may define one or more attributes (or oral care arguments) describing an intended output from a trained machine learning model.
  • the methods may generate training datasets by generating one or more noisy representations of the intended output (e.g., a series of increasingly noisy representations), and/or training machine learning models using the one or more noisy representations of the intended output.
  • the methods may use a trained model (e.g., a hierarchical neural network feature extraction model – such as a U-Net) to iteratively denoise a representation (e.g., a representation which initially has a random, Gaussian or other default value or distribution).
  • the methods may generate (or modify) one or more denoised representations of the intended output, after one or more iterations of denoising.
  • the methods may generate one or more oral care appliances using the denoised representation, or otherwise use the denoised representation for one or more aspects of one or more digital oral care treatments.
  • a refined dataset may be generated by successively modifying one or more representations of the training dataset.
  • An initially untrained machine learning model may be trained using the refined training dataset, and outputted for use in oral care appliance generation.
  • the refined training dataset may be generated by adding noise to aspects of the one or more representations of the training dataset.
  • one or more aspects of one or more representations of the training dataset may be encoded into a latent form before generating the one or more noisy representations.
  • each example of the training dataset may first undergo latent encoding (e.g., using an encoder) before the refined dataset (of iteratively noisier examples) is generated.
  • the one or more oral care attributes may describe the shape, structure, layout, or other properties of an intended output. Oral care arguments may include oral care metrics, and/or oral care parameters.
  • the one or more oral care attributes may include at least one of a real value, a categorical value, or a natural language text value.
  • the trained machine learning model may include at least one neural network, such as a hierarchical neural network feature extraction model (e.g., a U-Net), an autoencoder, or a transformer (e.g., 3D SWIN transformer).
  • One or more 3D representations of the patient’s dentition e.g., one or more teeth of the patient
  • the methods may denoise a 3D oral care representation (e.g., a 3D oral care representation which is to undergo modification).
  • the 3D oral care representation (e.g., 3D representations of the patient’s dentition) may be encoded into a latent form.
  • the one or more denoised representations includes at least one or more labels for a mesh element (e.g., for use in mesh segmentation or mesh cleanup).
  • the methods may segment a 3D representation (e.g., a mesh, point cloud, voxelized representation, etc.) using the one or more labels.
  • One or more 3D representations of oral care data may be segmented or undergo mesh cleanup (e.g., 3D representations of a patient’s dentition).
  • Mesh cleanup may involve using the one or more labels to modify the one or more 3D oral care representations (e.g., to remove mesh elements which have a particular label value, or to modify mesh elements which have a particular label value).
  • the methods may, in some implementations, denoise one or more representations of transforms (e.g., transforms which may place teeth into setups poses – such as final setups or intermediate stages, place appliance components for appliance generation, or place fixture model components for fixture model generation).
  • the methods may, in some implementations, denoise one or more 3D representations of a tooth restoration design, to modify the shape and/or structure of a pre-restoration tooth and make that tooth suitable for use in digital oral care treatment (e.g., to use in the generation of a dental restoration appliance).
  • the methods may, in some implementations, denoise one or more 3D representations of appliance components, and make the one or more appliance components suitable for use in digital oral care treatment (e.g., to use in the generation of a dental restoration appliance).
  • the methods may, in some implementations, denoise one or more 3D representations of fixture model components, and make the one or more fixture model components suitable for use in digital oral care treatment (e.g., to use in the generation of a digital fixture model).
  • the digital fixture model may be 3D printed, resulting in a physical fixture model.
  • Orthodontic aligner trays may be thermoformed using the physical fixture model and used in orthodontic treatment of the patient.
  • the methods may, in some implementations, denoise one or more 3D representations of clear tray aligner trimlines, and make the one or more clear tray aligner trimlines suitable for use in digital oral care treatment (e.g., to use in trimming an aligner tray off of a physical fixture model).
  • the methods may, in some implementations, denoise one or more 3D representations of an archform, and make the one or more archforms suitable for use in digital oral care treatment (e.g., to use in the generation of an oral care appliance).
  • FIG.1 shows a method of using a denoising diffusion probabilistic model to generate (or modify) a 3D oral care representation.
  • FIG.2 shows a method of training a denoising diffusion probabilistic model to generate (or modify) a 3D oral care representation.
  • FIG.3 shows a method of generating orthodontic setups transforms using denoising diffusion probabilistic models.
  • FIG.4 shows a method of using a hierarchical neural network feature extraction module to perform mesh element labeling (e.g., for segmentation or mesh cleanup) of a 3D representation, according to the denoising diffusion probabilistic models of this disclosure.
  • FIG.5 shows a U-Net structure, which may be used to extract hierarchical features from a 3D representation.
  • FIG.6 shows a pyramid encoder-decoder structure, which may be used to extract hierarchical features from a 3D representation.
  • Diffusion models may be applied to 2D image generation, or the generation of 3D representations, among others.
  • such implementations may take input from natural language text (e.g., natural language text, real values, categorical values, reference images or reference 3D representations which describe an intended outcome).
  • natural language text e.g., natural language text, real values, categorical values, reference images or reference 3D representations which describe an intended outcome.
  • the techniques described herein expand diffusion models into the digital oral care space.
  • the techniques described herein use a diffusion model to generate 3D oral care representations (e.g., as 3D point clouds, 3D meshes, 3D voxelized representations, 3D surfaces and the like) based on inputted arguments (e.g., which may include natural language text, integer arguments, real-valued arguments, categorical arguments, and the like).
  • Such arguments may contain one or more attributes describing an intended output from a trained machine learning model.
  • Techniques described herein may, in some instances, take as input representations of the patient’s dentition (e.g., 3D point clouds, 2D views or 2D digital photographs), which are to be used as guides for the generation of artifacts which are to be used in digital oral care.
  • Techniques described herein may, in some instances, take as input representations of appliances, appliance components, trimlines, archforms or other 3D oral care representations (e.g., in the form of 3D point clouds, 2D views or 2D digital photographs) which are to be modified or used as templates or references for artifacts which are to be generated for use in digital oral care.
  • these inputs may be encoded into a latent (e.g., information-rich reduced dimensionality) form by autoencoders or by other encoder-decoder neural networks.
  • a trimline may be generated which is used to cut a thermoformed tray off of a fixture model (e.g., a fixture model used in the manufacture of either an indirect bonding tray or an orthodontic aligner).
  • techniques of this disclosure may realize improved data precision through the use of such latent representations.
  • a denoising diffusion model may involve a forward pass 206 over the input data 200 (e.g., a 3D point cloud of a tooth, a transform, or another 3D oral care representation described herein) which may generate a Markov chain of steps 204 which may introduce successively more noise (e.g., Gaussian noise) to the input data 200.
  • the forward pass 206 may generate training data.
  • the Markov chain may generate a set of successively noisier training data examples.
  • Input oral care arguments (or attributes) 202 may influence the functioning of the denoising diffusion model, causing the denoising diffusion model to generate output to the specification of the clinician (e.g., enabling the customization of the output that is generated by the denoising diffusion model).
  • Oral care arguments may include oral care parameters, oral care metrics, among other examples.
  • an optional latent encoding module 214 may encode 3D oral care representations 200 into latent form.
  • a latent encoding module may be trained to encode data into a reduced dimensionality latent form. Examples of data which may be encoded include a transform (e.g., a tooth transform, etc.), a point cloud or mesh describing a tooth, a set of mesh element labels.
  • Such data may be encoded into a latent vector or latent capsule.
  • an optional latent encoding module 212 may, in some implementations, encode one or more of oral care arguments 202 into latent form.
  • the Markov chain may generate a succession of increasingly noisy versions of the input data 200, which may then be used to train, at least in part, a denoising ML module 210 (e.g., which may function as a part of the reverse pass 208).
  • the denoising ML model which is trained for use in reverse pass may comprise one or more neural networks 210 (e.g., U-Net, VAE, 3D SWIN transformer, pyramid encoder- decoder, or the like) and may be trained to de-noise a highly noisy version of the data (e.g., starting with a completely randomized version of the input data structure and de-nosing that data structure until the data structure takes on aspects of a suitable generated output).
  • the reverse pass 208 may be used during model deployment (e.g., after deployment of the trained denoising ML model 210).
  • the reverse pass 208 may start with a point cloud (or other raw data structure – such as a transforms or mesh element labels) that has a Gaussian distribution, and over successive applications of the reverse pass denoising ML model 210 gradually shape that point cloud (or other data structure) into a suitable example of the target 3D oral care representation (e.g., a tooth design which is suitable for use in creating a dental restoration appliance).
  • a loss function e.g., cross-entropy or MSE, among others
  • MSE ground truth
  • the loss function may be used to train, at least in part, the denoising ML model 210.
  • oral care metrics may be computed on the generated 3D oral care representation 220.
  • the oral care metrics may be used, at least in part, to assess the quality of fitness for use of the generated 3D oral care representation (e.g., to measure whether the generated 3D oral care representation meets the specification of the oral care arguments 202 and is ready for use in generating an oral care appliance).
  • the forward pass 206 of the denoising diffusion model may generate a training dataset of increasingly noisy examples of the input data.
  • Noise may be introduced to disfigure the input data (e.g., an image, a 3D point cloud, a transform, or latent representations of one or more of these inputs, etc.) and those noisy examples may be used, at least in part, to train a denoising diffusion machine learning model 210 to reverse of this noise-introducing process (e.g., in model deployment).
  • the reverse pass 208 may be trained to reconstruct the pristine input data by removing the noise from a noisy example of that input data.
  • the denoising diffusion model e.g., a U-Net
  • the model is capable to generate new 3D oral care representations by passing a noisy data example (e.g., a randomly generated noisy example) through the denoising diffusion process (aka the reverse process).
  • a cohort patient case may include a set of tooth crown meshes, a set of tooth root meshes, or a data file containing attributes of the case (e.g., a JSON file).
  • a typical example of a cohort patient case may contain up to 32 crown meshes (e.g., which may each contain tens of thousands of vertices or tens of thousands of faces), up to 32 root meshes (e.g., which may each contain tens of thousands of vertices or tens of thousands of faces), multiple gingiva mesh (e.g., which may each contain tens of thousands of vertices or tens of thousands of faces) or one or more JSON files which may each contain tens of thousands of values (e.g., objects, arrays, strings, real values, Boolean values or Null values).
  • values e.g., objects, arrays, strings, real values, Boolean values or Null values
  • the denoising ML model 210 of the reverse pass 208 may modify an existing 3D oral care representation (e.g., modify a pre-restoration tooth design).
  • the existing 3D oral care representation e.g., an example of optional instant patient case data 2166
  • Optional instant patient case data 216 may include data pertaining to the dentition of a patient.
  • the instant patient case data may, in some implementations, be introduced to customize the functioning of the denoising diffusion ML model to the anatomy of the patient.
  • the instant patient case data may be encoded 218 into a latent or embedded form.
  • the instant patient case data may be provided to the denoising diffusion ML model 210.
  • the instant data 216 may comprise an appliance component or a fixture model component which requires modification.
  • which data structure that undergoes iterative denoising by the denoising ML model 210 may be initialized, at least in part, according to a stochastic process (e.g., using random noise or random configurations), or by aspects of the instant patient case data 216, or by a combination of the two.
  • a stochastic process e.g., using random noise or random configurations
  • the instant patient case data 216 may include 3D oral care representations described herein, including tooth transforms (e.g., maloccluded transforms for one or more teeth during setups prediction), tooth meshes with transforms already applied, tooth meshes without transforms applied, one or more 3D representations of pre-restoration tooth designs, pre-segmentation dental arch mesh(es), pre-cleanup dental arch mesh(es), one or more mesh element labels (e.g., for segmentation or mesh cleanup), one or more segmented teeth for use in coordinate system prediction, one or more coordinate systems (e.g., each of which may be described by a transform), one or more 3D representations of appliance components (e.g., parting surfaces, etc.), one or more fixture model components (e.g., digital pontic teeth or interproximal webbing, etc.), or the like.
  • tooth transforms e.g., maloccluded transforms for one or more teeth during setups prediction
  • tooth meshes with transforms already applied es
  • the denoising diffusion ML model 210 may be trained to generate one or more generated 3D oral care representations 220 (e.g., as defined herein) for the pat’ent's treatment.
  • a generated 3D oral care representation 220 (aka a denoised representation) may be generated over one or more iterations of denoising by the denoising diffusion ML model 210.
  • Generated 3D oral care representations 220 may include setups transforms for one or more teeth, transforms for the placement of one or more appliance components (e.g., for the generation of a dental restoration appliance), one or more 3D representations of post-restoration tooth designs, one or more generated (or modified) appliance components (e.g., a parting surface or gingival ribbon for use in generating a dental restoration appliance), one or more generated (or modified) fixture model components (e.g., one or more trimlines, etc.), post-segmentation dental arch mesh(es), post-cleanup dental arch mesh(es), one or more mesh element labels for use in segmentation or mesh cleanup, one or more object masks (e.g., masks to be applied to mesh elements) for use in segmentation or mesh cleanup, one or more coordinate axes for one or more predicted coordinate systems, one or more archforms, or other of the 3D oral care representations described herein.
  • appliance components e.g., for the generation of a dental restoration appliance
  • the machine learning techniques described herein may receive a variety of input data, as described herein, including tooth meshes for one or both dental arches of the patient.
  • the tooth data, appliance components, fixture model components, to name a few may be provided in the form of 3D representations, such as meshes, point clouds or voxelized geometries. These data may be preprocessed, for example, by arranging the constituent mesh elements into lists and computing an optional mesh element feature vector for each mesh element. Such vectors may impart valuable information about the shape and/or structure of an oral care mesh to the machine learning models described herein. Additional inputs may be received as the input to the machine learning models described herein, such as one or more oral care metrics.
  • Oral care metrics may be used for measuring one or more physical aspects of an oral care mesh (e.g., physical relationships within a tooth or between different teeth).
  • an oral care metric may be computed for either or both of a malocclusion oral care mesh example and/or a ground truth oral care mesh example which is then used in the training of the machine learning models described herein.
  • the metric value may be received as the input to the machine learning models described herein, as a way of training that model or those models to encode a distribution of such a metric over the several examples of the training dataset.
  • the network may then receive metric value(s) as input, to assist in training the network to link that inputted metric value to the physical aspects of the ground truth oral care mesh which is used in loss calculation.
  • Such a loss calculation may quantify the difference between a prediction and a ground truth example (e.g., between a predicted oral care mesh and a ground truth oral care mesh).
  • the neural network techniques of this disclosure may, through the course of loss calculation and subsequent backpropagation, learn train the neural network to encode a distribution of a given metric.
  • one or more oral care arguments may be defined to specify one or more aspects of an intended 3D oral care representation (e.g., a 3D mesh, a polyline, a 3D point cloud or a voxelized geometry), which is to be generated using the machine learning models described herein (e.g., denoising diffusion models) that have been trained for that purpose.
  • oral care arguments may be defined to specify one or more aspects of a customized vector, matrix or any other numerical representation (e.g., to describe 3D oral care representations such as control points for a spline, an archform, a transform to place a tooth or appliance component relative to another 3D oral care representation, or a coordinate system), which is to be generated using the machine learning models described herein (e.g., denoising diffusion models) that have been trained for that purpose.
  • a customized vector, matrix or other numerical representation may describe a 3D oral care representation which conforms to the intended outcome of the treatment of the patient.
  • Oral care arguments may include oral care metrics or oral care parameters, among others.
  • Oral care arguments may specify one or more aspects of an oral care procedure – such as orthodontic setups prediction or restoration design generation, among others.
  • one or more oral care parameters may be defined which correspond to respective oral care metrics.
  • Oral care arguments can be provided as the input to the machine learning models described herein and be taken as an instruction to that module to generate an oral care mesh with the specified customization, to place an oral care mesh for the generation of an orthodontic setup (or appliance), to segment an oral care mesh, or to clean up an oral care mesh, to name a few examples.
  • This interplay between oral care metrics and oral care parameters may also apply to the training and deployment of other predictive models in oral care as well.
  • the predictive models of the present disclosure may, in some implementations, may produce more accurate results by the incorporation of one or more of the following inputs: archform information V, interproximal reduction (IPR) information U, tooth dimension information P, tooth gap information Q, latent capsule representations of oral care meshes T, latent vector representations of oral care meshes A, procedure parameters K (which may describe a clinician’s intended treatment of the patient), doctor preferences L (which may describe the typical procedure parameters chosen by a doctor), flags regarding tooth status M (such as for fixed or pinned teeth), tooth position information N, tooth orientation information O, tooth name/dental notation R, oral care metrics S (comprising of at least one of oral care metrics and restoration design metrics).
  • IPR interproximal reduction
  • Systems of this disclosure may, in some instances, be deployed at a clinical context (such as a dental or orthodontic office) for use by clinicians (e.g., doctors, dentists, orthodontists, nurses, hygienists, oral care technicians).
  • clinicians e.g., doctors, dentists, orthodontists, nurses, hygienists, oral care technicians.
  • Such systems which are deployed at a clinical context may enable clinicians to process oral care data (such as dental scans) in the clinic environment, or in some instances, in a “chairside” context (where the patient is present in the clinical environment).
  • a non-limiting list of examples of techniques may include: segmentation (e.g., such as diffusion segmentation), mesh cleanup, coordinate system prediction, CTA trimline generation, restoration design generation (e.g., such as using diffusion models for 3D point cloud generation), appliance component generation or placement or assembly (e.g., such as using diffusion models for 3D point cloud generation), generation of other oral care meshes, the validation of oral care meshes, setups prediction (e.g., such as diffusion setups), removal of hardware from tooth meshes, hardware placement on teeth, imputation of missing values, clustering on oral care data, oral care mesh classification, setups comparison, metrics calculation, or metrics visualization.
  • segmentation e.g., such as diffusion segmentation
  • mesh cleanup e.g., coordinate system prediction
  • CTA trimline generation e.g., restoration design generation (e.g., such as using diffusion models for 3D point cloud generation)
  • appliance component generation or placement or assembly e.g., such as using diffusion models for 3D point cloud generation
  • Some existing techniques may use diffusion models for dental crown restoration. Such techniques may convert a 3D point cloud of a tooth to a 2D depth map, use a diffusion model to process the depth map, and then attempt to convert the resulting modified depth map back into a 3D point cloud of the tooth. Error may be introduced by attempting to convert between the 3D representations and 2D representations in this fashion.
  • the denoising diffusion probabilistic models of this disclosure can be trained to directly generate a 3D tooth restoration design (or other 3D representations described herein, such as fixture model components or appliance components) without the intervening step of depth map generation.
  • the denoising ML model 210 of this disclosure may be directly trained on a series of increasingly noisy point clouds 204 of tooth restoration designs (e.g., either pre-restoration or post-restoration tooth designs).
  • the fully trained denoising ML model 210 may then be used to denoise a noisy (or randomly initialized) 3D point cloud (or other 3D representation) directly to generate a tooth restoration design, without the need to first convert the 3D tooth data into a 2D depth map or any other 2D representation, and then back into a 3D representation.
  • a noisy (or randomly initialized) 3D point cloud or other 3D representation
  • the techniques of this disclosure avoid the computational cost (e.g., compute cycles) and storage (e.g., computer memory) requirements of converting between representations of different dimensionalities (e.g., as pertain to the depth map approach).
  • Systems of this disclosure may automate operations in digital orthodontics (e.g., setups prediction, hardware placement, setups comparison), in digital dentistry (e.g., restoration design generation) or in combinations thereof. Some techniques may apply to either or both of digital orthodontics and digital dentistry.
  • a non-limiting list of examples is as follows: segmentation, mesh cleanup, coordinate system prediction, oral care mesh validation, imputation of oral care parameters, oral care mesh generation or modification (e.g., using autoencoders, transformers, continuous normalizing flows or denoising diffusion models), metrics visualization, appliance component placement or appliance component generation or the like.
  • systems of this disclosure may enable a clinician or technician to process oral care data (such as scanned dental arches).
  • the systems of this disclosure may enable orthodontic treatment planning, which may involve setups prediction as at least one operation.
  • Systems of this disclosure may also enable restoration design generation (e.g., using a denoising diffusion model), where one or more restored tooth designs are generated and processed in the course of creating oral care appliances.
  • Systems of this disclosure may enable either or both of orthodontic or dental treatment planning, or may enable automation steps in the generation of either or both of orthodontic or dental appliances. Some appliances may enable both of dental and orthodontic treatment, while other appliances may enable one or the other.
  • This disclosure pertains to digital oral care, which encompasses the fields of digital dentistry and digital orthodontics.
  • This disclosure generally describes methods of processing three-dimensional (3D) representations of oral care data.
  • 3D representation is a 3D geometry.
  • a 3D representation may include, be, or be part of one or more of a 3D polygon mesh, a 3D point cloud (e.g., such as derived from a 3D mesh), a 3D voxelized representation (e.g., a collection of voxels – for sparse processing), or 3D representations which are described by mathematical equations.
  • a 3D representation may describe elements of the 3D geometry and/or 3D structure of an object.
  • Dental arches S1, S2, S3 and S4 all contain the exact same tooth meshes, but those tooth meshes are transformed differently, according to the following description.
  • a first arch S1 includes a set of tooth meshes arranged (e.g., using transforms) in their positions in the mouth, where the teeth are in the mal positions and orientations.
  • a second arch S2 includes the same set of tooth meshes from S1 arranged (e.g., using transforms) in their positions in the mouth, where the teeth are in the ground truth setup positions and orientations.
  • a third arch S3 includes the same meshes as S1 and S2, which are arranged (e.g., using transforms) in their positions in the mouth, where the teeth are in the predicted final setup poses (e.g., as predicted by one or more of the techniques of this disclosure).
  • S4 is a counterpart to S3, where the teeth are in the poses corresponding to one of the several intermediate stages of orthodontic treatment with clear tray aligners.
  • GDL geometric deep learning
  • RL reinforcement learning
  • VAE variational autoencoder
  • MLP multilayer perceptron
  • PT pose transfer
  • FDG force directed graphs
  • MLP Setups, VAE Setups and Capsule Setups each fall within the scope of Autoencoder Setups. Some implementations of MLP Setups may fall within the Scope of Transformer Setups.
  • Representation Setups refers to any of MLP Setups, VAE Setups, Capsule Setups and any other setups prediction machine learning model which uses an autoencoder to create the representation for at least one tooth.
  • Each of the setups prediction techniques of this disclosure is applicable to the fabrication of clear tray aligners and/or indirect bonding trays.
  • the setups predictions techniques may also be applicable to other products that involve final teeth poses, also.
  • a pose may comprise a position (or location) and a rotation (or orientation).
  • a 3D mesh is a data structure which may describe the geometry or shape of an object related to oral care, including but not limited to a tooth, a hardware element, or a patient’s gum tissue.
  • a 3D mesh may include one or more mesh elements such as one or more of vertices, edges, faces and combinations thereof.
  • mesh elements may include voxels, such as in the context of sparse mesh processing operations.
  • Various spatial and structural features may be computed for these mesh elements and be provided to the predictive models of this disclosure, with the predictive models of this disclosure providing the technical advantage of improving data precision in the form of the models of this disclosure outputting more accurate predictions.
  • a patient’s dentition may include one or more 3D representations of the patient’s teeth (e.g., and/or associated transforms), gums and/or other oral anatomy.
  • An orthodontic metric may, in some implementations, quantify the relative positions and/or orientations of at least one 3D representation of a tooth relative to at least one other 3D representation of a tooth.
  • a restoration design metric may, in some implementations, quantify at least one aspect of the structure and/or shape of a 3D representation of a tooth.
  • An orthodontic landmark (OL) may, in some implementations, locate one or more points or other structural regions of interest on a 3D representation of a tooth.
  • An OL may, in some implementations, be used in the generation of an orthodontic or dental appliance, such as a clear tray aligner or a dental restoration appliance.
  • a mesh element may, in some implementations, comprise at least one constituent element of a 3D representation of oral care data.
  • mesh elements may include at least: vertices, edges, faces and voxels.
  • a mesh element feature may, in some implementations, quantify some aspect of a 3D representation in proximity to or in relation with one or more mesh elements, as described elsewhere in this disclosure.
  • Orthodontic procedure parameters may, in some implementations, specify at least one value which defines at least one aspect of planned orthodontic treatment for the patient (e.g., specifying desired target attributes of a final setup in final setups prediction).
  • Orthodontic Doctor preferences may, in some implementations, specify at least one typical value for an OPP, which may, in some instances, be derived from past cases which have been treated by one or more oral care practitioners.
  • Restoration Design Parameters may, in some implementations, specify at least one value which defines at least one aspect of planned dental restoration treatment for the patient (e.g., specifying desired target attributes of a tooth which is to undergo treatment with a dental restoration appliance).
  • Doctor Restoration Design Preferences may, in some implementations, specify at least one typical value for an RDP, which may, in some instances, be derived from past cases which have been treated by one or more oral care practitioners.
  • 3D oral care representations may include, but are not limited to: 1) a set of mesh element labels which may be applied to the 3D mesh elements of teeth/gums/hardware/appliance meshes (or point clouds) in the course of mesh segmentation or mesh cleanup; 2) 3D representation(s) for one or more teeth/gums/hardware/appliances for which shapes have been modified (e.g., trimmed, distorted, or filled-in) in the course of mesh segmentation or mesh cleanup; 3) one or more coordinate systems (e.g., describing one, two, three or more coordinate axes) for a single tooth or a group of teeth (such as a full arch – as with the LDE coordinate system); 4) 3D representation(s) for one or more teeth for which shapes have been modified or otherwise made suitable for use in
  • the Setups Comparison tool may be used to compare the output of the GDL Setups model against ground truth data, compare the output of the RL Setups model against ground truth data, compare the output of the VAE Setups model against ground truth data and compare the output of the MLP Setups model against ground truth data.
  • the Metrics Visualization tool can enable a global view of the final setups and intermediate stages produced by one or more of the setups prediction models, with the advantage of enabling the selection of the best setups prediction model.
  • the Metrics Visualization tool furthermore, enables the computation of metrics which have a global scope over a set of intermediate stages. These global metrics may, in some implementations, be consumed as inputs to the neural networks for predicting setups (e.g., GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, PT Setups, Similarity Setups, among others). The global metrics may also be provided to FDG Setups.
  • GDL Setups e.g., GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, PT Setups, Similarity Setups, among others.
  • the global metrics may also be provided to FDG Setups.
  • the local metrics from this disclosure may, in some implementations, be consumed by the neural networks herein for predicting setups, with the advantage of improving predictive results.
  • the metrics described in this disclosure may, in some implementations, be visualized using the Metric Visualization tool.
  • the VAE and MAE models for mesh element labelling and mesh in-filling can be advantageously combined with the setups prediction neural networks, for the purpose of mesh cleanup ahead of or during the prediction process.
  • the VAE for mesh element labelling may be used to flag mesh elements for further processing, such as metrics calculation, removal, or modification.
  • such flagged mesh elements may be provided as inputs to a setups prediction neural network, to inform that neural network about important mesh features, attributes or geometries, with the advantage of improving the performance of the resulting setups prediction model.
  • mesh in-filling may cause the geometry of a tooth to become more nearly complete, enabling the better functioning of a setups prediction model (i.e., improved correctness of prediction on account of better-formed geometry).
  • a neural network to classify a setup i.e., the Setups Classifier
  • the setups classifier tells that setups prediction neural network when the predicted setup is acceptable for use and can be provided to a method for aligner tray generation.
  • a Setups Classifier may aid in the generation of final setups and also in the generation of intermediate stages.
  • a Setups Classifier neural network may be combined with the Metrics Visualization tool.
  • a Setups Classification neural network may be combined with the Setups Comparison tool (e.g., the Setup Comparison tool may output an indication of how a setup produced in part by the Setups Classifier compares to a setup produced by another setups prediction method).
  • the VAE for mesh element labelling may identify one or more mesh elements for use in a metrics calculation.
  • the resulting metrics outputs may be visualized by the Metrics Visualization tool.
  • the Setups Classifier neural network may aid in the setups prediction technique described in U.S. Patent Application No. US20210259808A1 (which is incorporated herein by reference in its entirety) or the setups prediction technique described in PCT Application with Publication No. WO2021245480A1 (which is incorporated herein by reference in its entirety) or in PCT Application No. PCT/IB2022/057373 (which is incorporated herein by reference in its entirety).
  • the Setups Classifier would help one or more of those techniques to know when the predicted final setup is most nearly correct.
  • the Setups Classifier neural network may output an indication of how far away from final setup a given setup is (i.e., a progress indicator).
  • the latent space embedding vector(s) from the reconstruction VAE can be concatenated with the inputs to the setups prediction neural network described in WO2021245480A1.
  • the latent space vectors can also be incorporated as inputs to the other setups prediction models: GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups and Diffusion Setups, among others.
  • the various setups prediction neural networks of this disclosure may work together to produce the setups required for orthodontic treatment.
  • the GDL Setups model may produce a final setup, and the RL Setups model may use that final setup as input to produce a series of intermediate stages setups.
  • the VAE Setups model (or the MLP Setups model) may create a final setup which may be used by an RL Setups model to produce a series of intermediate stages setups.
  • a setup prediction may be produced by one setups prediction neural network, and then taken as input to another setups prediction neural network for further improvements and adjustments to be made. In some implementations, such improvements may be performed in iterative fashion.
  • a setups validation model such as the model disclosed in US Provisional Application No. US63/366495, may be involved in this iterative setups prediction loop.
  • a setup may be generated (e.g., using a model trained for setups prediction, such as GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, PT Setups, Similarity Setups and FDG Setups, among others), then the setup undergoes validation. If the setup passes validation, the setup may be outputted for use. If the setup fails validation, the setup may be sent back to one or more of the setups prediction models for corrections, improvements and/or adjustments. In some instances, the setups validation model may output an indication of what is wrong with the setup, enabling the setups generation model to make an improved version upon the next iteration. The process iterates until done.
  • a model trained for setups prediction such as GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, PT Setups, Similarity Setups and FDG Setup
  • two or more of the following techniques of the present disclosure may be combined in the course of orthodontic and/or dental treatment: GDL Setups, Setups Classification, Reinforcement Learning (RL) Setups, Setups Comparison, Autoencoder Setups (VAE Setups or Capsule Setups), VAE Mesh Element Labeling, Masked Autoencoder (MAE) Mesh In- filling, Multi-Layer Perceptron (MLP) Setups, Metrics Visualization, Imputation of Missing Oral Care Parameters Values, Tooth Classification Using Latent Vector, FDG Setups, Pose Transfer Setups, Restoration Design Metrics Calculation, Neural Network Techniques for Dental Restoration and/or Orthodontics (e.g., 3D Oral Care Representation Generation or Modification Using Transformers), Landmark-based (LB) Setups, Diffusion Setups, Imputation of Tooth Movement Procedures, Capsule Autoencoder Segment
  • Oral care parameters may include one or more values that specify orthodontic procedure parameters, or restoration design parameters (RDP), as described herein.
  • Oral care parameters may define one or more intended aspects of a 3D oral care representation and may be provided to an ML model to promote that ML model to generate output which may be used in the generation of oral care appliances that are suitable for the treatment of a patient.
  • Other types of values include doctor preferences and restoration design preferences, as described herein.
  • Doctor preferences and restoration design preferences may define the typical treatment choices or practices of a particular clinician. Restoration design preferences are subjective to a particular clinician, and so differ from restoration design parameters.
  • doctor preferences or restoration design preferences may be computed by unsupervised means, such as clustering, which may determine the typical values that a clinician uses in patient treatment.
  • Those typical values may be stored in a datastore and recalled to be provided to an automated ML model as default values (e.g., default values which may be modified before execution of the model).
  • RDP restoration design parameter
  • one clinician may prefer one value for a restoration design parameter (RDP), while another clinician may prefer a different value for that RDP, when faced with a similar diagnosis or treatment protocol.
  • RDP restoration design parameter
  • Procedure parameters and/or doctor preferences may, in some implementations, be provided to a setups prediction model for orthodontic treatment, for the purpose of improving the customization of the resulting orthodontic appliance.
  • Restoration design parameters and doctor restoration preferences may in some implementations be used to design tooth geometry for use in the creation of a dental restoration appliance, for the purpose of improving the customization of that appliance.
  • ML prediction models of this disclosure in orthodontic treatment, may also take as input a setup (e.g., an arrangement of teeth).
  • an ML prediction model of this disclosure may take as input a final setup (i.e., final arrangement of teeth), such as in the case of a prediction model trained to generate intermediate stages.
  • these preferences are referred to as doctor restoration preferences, but it is intended to be used in a non-limiting sense. Specifically, it should be appreciated that these preferences may be specified by any treating or otherwise appropriate medical professional and are not intended to be limited to doctor preferences per se (i.e., preferences from someone in possession of an M.D. or equivalent degree).
  • An oral care professional or clinician such as a dentist or orthodontist, may specify information about patient treatment in the form of a patient-specific set of procedure parameters.
  • an oral care professional may specify a set of general preferences (aka doctor preferences) for use over a broad range of cases, to use as default values in the set of procedure parameters specification process.
  • Oral care parameters may in some implementations be incorporated into the techniques described in this disclosure, such as one or more of GDL Setups, VAE Setups, RL Setups, Setups Comparison, Setups Classification, VAE Mesh Element Labelling, MAE Mesh In-Filling, Validation Using Autoencoders, Imputation of Missing Procedure Parameters Values, Metrics Visualization, or FDG Setups.
  • GDL Setups e.g., VAE Setups, RL Setups, Setups Comparison, Setups Classification, VAE Mesh Element Labelling, MAE Mesh In-Filling, Validation Using Autoencoders, Imputation of Missing Procedure Parameters Values, Metrics Visualization, or FDG Setups.
  • One or more of these models may take as input one or more procedure parameters vector K and/or one or more doctor preference vectors L.
  • one or more of these models may introduce to one or more of a neural network’s hidden layers one or more
  • one or more of these models may introduce either or both of K and L to a mathematical calculation, such as a force calculation, for the purpose of improving that calculation and the ultimate customization of the resulting appliance to the patient.
  • a neural network for predicting a setup may incorporate information from an oral care professional (aka doctor). This information may influence the arrangement of teeth in the final setup, bringing the positions and orientations of the teeth into conformance with a specification set by the doctor, within tolerances.
  • oral care parameters may be fed directly into the generator network as a separate input alongside the mesh data.
  • oral care parameters may be incorporated into the feature vector which is computed for each mesh element before the mesh elements are provided to the generator for processing.
  • a setups prediction model e.g., VAE setups or diffusion setups
  • the procedure parameters K and/or the doctor preference information L may be concatenated with the latent space vector C.
  • a doctor’s preferences (e.g., in an orthodontic context) and/or doctor’s restoration preferences may be indicated in a treatment form, or they could be based upon characteristics in treatment plans such as final setup characteristics (e.g., amount of bite correction or midline correction in planned final setups), intermediate staging characteristics (e.g., treatment duration, tooth movement protocols, or overcorrection strategies), or outcomes (e.g., number of revisions/refinements).
  • final setup characteristics e.g., amount of bite correction or midline correction in planned final setups
  • intermediate staging characteristics e.g., treatment duration, tooth movement protocols, or overcorrection strategies
  • outcomes e.g., number of revisions/refinements.
  • Orthodontic procedure parameters may specify one or more of the following (with possible values shown in ⁇ ). Non-limiting categorical values for some example OPP are described below. In some implementations, a real value may be specified for one or more of these OPP.
  • the Overbite OPP may specify a quantity of overbite (e.g., in millimeters) which is desired in a setup, and may be received as input of a setups prediction model to provide that setups prediction model information about the amount of overbite which is desired in the setup.
  • Some implementations may specify a numerical value for the Overjet OPP, or other OPP.
  • one or more OPP may be defined which correspond to one or more orthodontic metrics (OM).
  • OM orthodontic metrics
  • a numerical value may be specified for such an OPP, for the purpose of controlling the output of a setups prediction model.
  • one or more orthodontic procedure parameters may be defined to specify at least one of the 2 nd and 3 rd order rotation angles to be applied to a tooth (i.e., angulation and torque, respectively), which may enable a target setup arrangement where crown landmarks lie within a threshold distance of a common occlusal plane, for example.
  • one or more orthodontic procedure parameters may be defined to specify the position in global coordinates where at least one landmark (e.g., a centroid) of a tooth crown (or root) is to be placed in a setup arrangement of teeth.
  • an oral care parameter may be defined which corresponds to an oral care metric.
  • an orthodontic procedure parameter may be defined which corresponds to an orthodontic metric (e.g., to specify at the input of a setups prediction model an amount of a certain metric which is desired to appear in a predicted setup).
  • Doctor preferences may differ from orthodontic procedure parameters in that doctor preferences pertain to an oral care provider and may comprise of the means, modes, medians, minimums, or maximums (or some other statistic) of past settings associated with an oral care provider’s treatment decisions on past orthodontic cases.
  • Procedure parameters may pertain to a specific patient, and describe the needs of a particular patient’s treatment. Doctor preferences may pertain to a doctor and the doctor’s past treatment practices, whereas procedure parameters may pertain to the treatment of a particular patient.
  • Doctor preferences may specify one or more of the following (with some non-limiting possible values shown in ⁇ ). Other possible values are found elsewhere in this disclosure. [0049] Doctor preferences may specify one or more of the following. Deep Bite Cases (Amount of Bite Correction) - Final Overbite: [real value in millimeters, e.g., 0.5 mm] Option – Intrude Upper Anteriors: ⁇ yes, no ⁇ Option – Include lower canines in vertical overcorrection: ⁇ yes, no ⁇ Midline Correction in Planned Final Setup: ⁇ MaintainInitialMidline, ImproveMidlineWithIPR, AsIndicated ⁇ Deep Bite Cases – Reverse Curve of Speed: ⁇ yes, no ⁇ Anterior Open Bite Cases – Final Overbite: [real value in millimeters, e.g., 2 mm] Is Arch Expansion a Priority for Your Cases?: ⁇ Yes, No ⁇ If yes, specify acceptable expansion
  • archform information V may be provided as an input to any of the GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups and Diffusion Setups prediction neural networks.
  • archform information V may be provided directly to one or more internal neural network layers in one or more of those setups applications.
  • the additional procedure parameters may include text descriptions of the patient’s medical condition and of the intended treatment.
  • Such text descriptions may be analyzed via natural language processing operations, including tokenization, stop word removal, stemming, n-gram formation, text data vectorization, bag of words analysis, term frequency inverse document frequency (TF-IDF) analysis, sentiment analysis, naive Bayes classification, and/or logistic regression classification.
  • the outputs of such analysis techniques may be used as input to one or more of the neural networks of this disclosure with the advantage of customizing and improving the predicted outputs (e.g., the predicted setups or predicted mesh geometries).
  • TF-IDF term frequency inverse document frequency
  • a dataset used for training one or more of the neural network models of this disclosure may be filtered conditionally on one or more of the orthodontic procedure parameters described in this section.
  • patient cases which exhibit outlier values for one or more of these procedure parameters may be omitted from a dataset (alternatively used to form a dataset) for training one or more of the neural networks of this disclosure.
  • One or more procedure parameters and/or doctor preferences may be provided to a neural network during training. In this manner the neural network may be conditioned on the one or more procedure parameters and/or doctor preferences.
  • tooth shape-based inputs may be provided to a neural network for setups predictions.
  • non-shape-based inputs can be used, such as a tooth name or designation, as it pertains to dental notation.
  • a vector R of flags may be provided to the neural network, where a ‘1’ value indicates that the tooth is present and a ‘0’ value indicates that the tooth is absent from the patient case (though other values are possible).
  • the vector R may comprise a 1- hot vector, where each element in the vector corresponds to a tooth type, name or designation. Identifying information about a tooth (e.g., the tooth’s name) can be provided to the predictive neural networks of this disclosure, with the advantage of enabling the neural network to become trained to handle different teeth in tooth-specific ways.
  • the setups prediction model may learn to make setups transformations predictions for a specific tooth designation (e.g., upper right central incisor, or lower left cuspid, etc.).
  • the autoencoder may be trained to provide specialized treatment to a tooth according to that tooth’s designation, in this manner.
  • Tooth designation/name may be defined, for example, according to the Universal Numbering System, Palmer System, or the FDI World Dental Federation notation (ISO 3950).
  • a vector R may be defined as an optional input to the setups prediction neural networks of this disclosure, where there is a 0 in the vector element corresponding to each of the wisdom teeth, and a 1 in the elements corresponding to the following teeth: UR7, UR6, UR5, UR4, UR3, UR2, UR1, UL1, UL2, UL3, UL4, UL5, UL6, UL7, LL7, LL6, LL5, LL4, LL3, LL2, LL1, LR1, LR2, LR3, LR4, LR5, LR6, LR7 [0057]
  • the position of the tooth tip may be provided to a neural network for setups predictions.
  • one or more vectors S of the orthodontic metrics described elsewhere in this disclosure may be provided to a neural network for setups predictions.
  • the advantage is an improved capacity for the network to become trained to understand the state of a maloccluded setup and therefore be able to predict a more accurate final setup or intermediate stage.
  • the neural networks may take as input one or more indications of interproximal reduction (IPR) U, which may indicate the amount of enamel that is to be removed from a tooth during the course orthodontic treatment (either mesially or distally).
  • IPR interproximal reduction
  • IPR information (e.g., quantity of IPR that is to be performed on one or more teeth, as measured in millimeters, or one or more binary flags to indicate whether or not IPR is to be performed on each tooth identified by flagging) may be concatenated with a latent vector A which is produced by a VAE or a latent capsule T autoencoder.
  • the vector(s) and/or capsule(s) resulting from such a concatenation may be provided to one or more of the neural networks of the present disclosure, with the technical improvement or added advantage of enabling that predictive neural network to account for IPR.
  • IPR is especially relevant to setups prediction methods, which may determine the positions and poses of teeth at the end of treatment or during one or more stages during treatment.
  • one or more procedure parameters K and/or doctor preferences vectors L may be introduced to a setups prediction model.
  • one or more optional vectors or values of tooth position N e.g., XYZ coordinates, in either tooth local or global coordinates
  • tooth orientation O e.g., pose, such as in transformation matrices or quaternions, Euler angles or other forms described herein
  • dimensions of teeth P e.g., length, width, height, circumference, diameter, diagonal measure, volume - any of which dimensions may be normalized in comparison to another tooth or teeth
  • tooth dimensions P may in some instances be used to describe the intended dimensions of a tooth for dental restoration design generation.
  • tooth dimensions P e.g., length, width, height, or circumference
  • the tooth dimension of height may be measured as the distance from gums to incisal edge.
  • the tooth dimension of width may be measured as the distance from the mesial extent to the distal extent of the tooth.
  • the circularity or roundness of the tooth cross-section may be measured and included in the vector P. Circularity or roundness may be defined as the ratio of the radii of inscribed and circumscribed circles.
  • the distance Q between adjacent teeth can be implemented in different ways (and computed using different distance definitions, such as Euclidean or geodesic).
  • a distance Q1 may be measured as an averaged distance between the mesh elements of two adjacent teeth.
  • a distance Q2 may be measured as the distance between the centers or centroids of two adjacent teeth.
  • a distance Q3 may be measured between the mesh elements of closest approach between two adjacent teeth.
  • a distance Q4 may be measured between the cusp tips of two adjacent teeth.
  • Teeth may, in some implementations, be considered adjacent within an arch. Teeth may, in some implementations, also be considered adjacent between opposing arches.
  • any of Q1, Q2, Q3 and Q4 may be divided by a term for the purpose of normalizing the resulting value of Q.
  • the normalizing term may involve one or more of: the volume of a tooth, the count of mesh elements in a tooth, the surface area of a tooth, the cross-sectional area of a tooth (e.g., as projected into the XY plane), or some other term related to tooth size.
  • the vector M may contain flags which apply to one or more teeth.
  • M contains at least one flag for each tooth to indicate whether the tooth is pinned.
  • M contains at least one flag for each tooth to indicate whether the tooth is fixed.
  • M contains at least one flag for each tooth to indicate whether the tooth is pontic.
  • a flag that is set to a value that indicates that a tooth should be fixed is a signal to the network that the tooth should not move over the course of treatment.
  • the neural network loss function may be designed to be penalized for any movement in the indicated teeth (and in some particular cases, may be heavily penalized).
  • a flag to indicate that a tooth is pontic informs the network that the tooth gap is to be maintained, although that gap is allowed to move.
  • M may contain a flag indicating that a tooth is missing.
  • the presence of one or more fixed teeth in an arch may aid in setups prediction, because the one or more fixed teeth may provide an anchor for the poses of the other teeth in the arch (i.e., may provide a fixed reference for the pose transformations of one or more of the other teeth in the arch).
  • one or more teeth may be intentionally fixed, so as to provide an anchor against which the other teeth may be positioned.
  • a 3D representation (such as a mesh) which corresponds to the gums may be introduced, to provide a reference point against which teeth can be moved.
  • one or more of the optional input vectors K, L, M, N, O, P, Q, R, S, U and V described elsewhere in this disclosure may also be provided to the input or into an intermediate layer of one or more of the predictive models of this disclosure.
  • these optional vectors may be provided to the MLP Setups, GDL Setups, RL Setups, VAE Setups, Capsule Setups and/or Diffusion Setups, with the advantage of enabling the respective model to generate setups which better meet the orthodontic treatment needs of the patient.
  • such inputs may be provided, for example, by being concatenated with one or more latent vectors A which are also provided to one or more of the predictive models of this disclosure.
  • such inputs may be introduced, for example, by being concatenated with one or more latent capsules T which are also provided to one or more of the predictive models of this disclosure.
  • K, L, M, N, O, P, Q, R, S, U and V may be introduced to the neural network (e.g., MLP or Transformer) directly in a hidden layer of the network.
  • a setups prediction model (such as GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, PT Setups, Similarity Setups and Diffusion Setups) may take as input one or more latent vectors A which correspond to one or more input oral care meshes (e.g., such as tooth meshes).
  • a setups prediction model (such as GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups and Diffusion Setups) may take as input one or more latent capsules T which correspond to one or more input oral care meshes (e.g., such as tooth meshes).
  • a setups prediction method may take as input both of A and T.
  • Various loss calculation techniques are generally applicable to the techniques of this disclosure (e.g., GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, PT Setups, Similarity Setups, Setups Classification, Tooth Classification, VAE Mesh Element Labelling, MAE Mesh In-Filling and the imputation of procedure parameters).
  • GDL Setups RL Setups
  • VAE Setups Capsule Setups
  • MLP Setups Diffusion Setups
  • PT Setups Diffusion Setups
  • Similarity Setups Setups Classification, Tooth Classification, VAE Mesh Element Labelling, MAE Mesh In-Filling and the imputation of procedure parameters.
  • MSE mean squared error
  • Losses may be computed and used in the training of neural networks, such as multi-layer perceptron’s (MLP), U-Net structures, generators and discriminators (e.g., for GANs), autoencoders, variational autoencoders, regularized autoencoders, masked autoencoders, transformer structures, or the like. Some implementations may use either triplet loss or contrastive loss, for example, in the learning of sequences. [0069] Losses may also be used to train encoder structures and decoder structures.
  • MLP multi-layer perceptron’s
  • U-Net structures generators and discriminators (e.g., for GANs)
  • autoencoders variational autoencoders
  • regularized autoencoders regularized autoencoders
  • masked autoencoders transformer structures, or the like.
  • Losses may also be used to train encoder structures and decoder structures.
  • a KL- Divergence loss may be used, at least in part, to train one or more of the neural networks of the present disclosure, such as a mesh reconstruction autoencoder or the generator of GDL Setups, which the advantage of imparting Gaussian behavior to the optimization space.
  • This Gaussian behavior may enable a reconstruction autoencoder to produce a better reconstruction (e.g., when a latent vector representation is modified and that modified latent vector is reconstructed using a decoder, the resulting reconstruction is more likely to be a valid instance of the inputted representation).
  • There are other techniques for computing losses which may be described elsewhere in this disclosure. Such losses may be based on quantifying the difference between two or more 3D representations.
  • MSE loss calculation may involve the calculation of an average squared distance between two sets, vectors or datasets. MSE may be generally minimized. MSE may be applicable to a regression problem, where the prediction generated by the neural network or other machine learning model may be a real number.
  • a neural network may be equipped with one or more linear activation units on the output to generate an MSE prediction.
  • Mean absolute error (MAE) loss and mean absolute percentage error (MAPE) loss can also be used in accordance with the techniques of this disclosure.
  • Cross entropy may, in some implementations, be used to quantify the difference between two or more distributions. Cross entropy loss may, in some implementations, be used to train the neural networks of the present disclosure.
  • Cross entropy loss may, in some implementations, involve comparing a predicted probability to a ground truth probability. Other names of cross entropy loss include “logarithmic loss,” “logistic loss,” and “log loss”. A small cross entropy loss may indicate a better (e.g., more accurate) model. Cross entropy loss may be logarithmic. Cross entropy loss may, in some implementations, be applied to binary classification problems. In some implementations, a neural network may be equipped with a sigmoid activation unit at the output to generate a probability prediction. In the case of multi-class classifications, cross entropy may also be used.
  • a neural network trained to make multi-class predictions may, in some implementations, be equipped with one or more softmax activation functions at the output (e.g., where there is one output node for class that is to be predicted).
  • Other loss calculation techniques which may be applied in the training of the neural networks of this disclosure include one or more of: Huber loss, Hinge loss, Categorical hinge loss, cosine similarity, Poisson loss, Logcosh loss, or mean squared logarithmic error loss (MSLE). Other loss calculation methods are described herein and may be applied to the training of any of the neural networks described in the present disclosure.
  • One or more of the neural networks of the present disclosure may, in some implementations, be trained, at least in part by a loss which is based on at least one of: a Point-wise Mesh Euclidean Distance (PMD) and an Earth Mover’s Distance (EMD).
  • PMD Point-wise Mesh Euclidean Distance
  • EMD Earth Mover’s Distance
  • Some implementations may incorporate a Hausdorff Distance (HD) calculation into the loss calculation.
  • HD Hausdorff Distance
  • Computing the Hausdorff distance between two or more 3D representations may provide one or more technical improvements, in that the HD not only accounts for the distances between two meshes, but also accounts for the way that those meshes are oriented, and the relationship between the mesh shapes in those orientations (or positions or poses).
  • Hausdorff distance may improve the comparison of two or more tooth meshes, such as two or more instances of a tooth mesh which are in different poses (e.g., such as the comparison of predicted setup to ground truth setup which may be performed in the course of computing a loss value for training a setups prediction neural network).
  • Reconstruction loss may compare a predicted output to a ground truth (or reference) output.
  • all_points_target is a 3D representation (e.g., a 3D mesh or point cloud) corresponding to ground truth data (e.g., a ground truth tooth restoration design, or a ground truth example of some other 3D oral care representation).
  • all_points_predicted is a 3D representation (e.g., a 3D mesh or point cloud) corresponding to generated or predicted data (e.g., a generated tooth restoration design, or a generated example of some other kind of 3D oral care representation).
  • Other implementations of reconstruction loss may additionally (or alternatively) involve L2 loss, mean absolute error (MAE) loss or Huber loss terms.
  • Reconstruction error may compare reconstructed output data (e.g., as generated by a reconstruction autoencoder, such as a tooth design which has been generated for use in generating a dental restoration appliance) to the original input data (e.g., the data which were provided to the input of the reconstruction autoencoder, such as a pre-restoration tooth).
  • all_points_input is a 3D representation (e.g., a 3D mesh or point cloud) corresponding to input data (e.g., the pre-restoration tooth design, which was provided to a reconstruction autoencoder, or another 3D oral care representation which is provided to the input of an ML model).
  • all_points_reconstructed is a 3D representation (e.g., 3D mesh or point cloud) corresponding to reconstructed (or generated) data (e.g., a reconstructed tooth restoration design, or another example of a generated 3D oral care representation).
  • reconstruction loss is concerned with computing a difference between a predicted output and a reference output
  • reconstruction error is concerned with computing a difference between a reconstructed output and an original input from which the reconstructed data are derived.
  • the techniques of this disclosure may include operations such as 3D convolution, 3D pooling, 3D unconvolution and 3D unpooling.3D convolution may aid segmentation processing, for example in down sampling a 3D mesh.3D un-convolution undoes 3D convolution, for example, in a U- Net.3D pooling may aid the segmentation processing, for example in summarized neural network feature maps.3D un-pooling undoes 3D pooling, for example in a U-Net.
  • These operations may be implemented by way of one or more layers in the predictive or generative neural networks described herein. These operations may be applied directly on mesh elements, such as mesh edges or mesh faces. These operations provide for technical improvements over other approaches because the operations are invariant to mesh rotation, scale, and translation changes. In general, these operations depend on edge (or face) connectivity, therefore these operations remain invariant to mesh changes in 3D space as long as edge (or face) connectivity is preserved. That is, the operations may be applied to an oral care mesh and produce the same output regardless of the orientation, position or scale of that oral care mesh which may lead to data precision improvement.
  • MeshCNN is a general-purpose deep neural network library for 3D triangular meshes, which can be used for tasks such as 3D shape classification or mesh element labelling (e.g., for segmentation or mesh cleanup). MeshCNN implements these operations on mesh edges. Other toolkits and implementations may operate on edges or faces. [0077] In some implementations of the techniques of this disclosure, neural networks may be trained to operate on 2D representations (such as images). In some implementations of the techniques of this disclosure, neural networks may be trained to operate on 3D representations (such as meshes or point clouds). An intraoral scanner may capture 2D images of the patient's dentition from various views.
  • An intraoral scanner may also (or alternatively) capture 3D mesh or 3D point cloud data which describes the patient's dentition.
  • autoencoders (or other neural networks described herein) may be trained to operate on either or both of 2D representations and 3D representations.
  • a 2D autoencoder (comprising a 2D encoder and a 2D decoder) may be trained on 2D image data to encode an input 2D image into a latent form (such as a latent vector or a latent capsule) using the 2D encoder, and then reconstruct a facsimile of the input 2D image using the 2D decoder.
  • 2D images may be readily captured using one or more of the onboard cameras. In other examples, 2D images may be captured using an intraoral scanner which is configured for such a function.
  • 2D autoencoder or other 2D neural network
  • 2D image convolution may involve the “sliding” of a kernel across a 2D image and the calculation of elementwise multiplications and the summing of those elementwise multiplications into an output pixel.
  • the output pixel that results from each new position of the kernel is saved into an output 2D feature matrix.
  • neighboring elements e.g., pixels
  • a 2D pooling layer may be used to down sample a feature map and summarize the presence of certain features in that feature map.
  • 2D reconstruction error may be computed between the pixels of the input and reconstructed images. The mapping between pixels may be well understood (e.g., the upper pixel [23,134] of the input image is directly compared to pixel [23,134] of the reconstructed image, assuming both images have the same dimensions).
  • Modern mobile devices may also have the capability of generating 3D data (e.g., using multiple cameras and stereophotogrammetry, or one camera which is moved around the subject to capture multiple images from different views, or both), which in some implementations, may be arranged into 3D representations such as 3D meshes, 3D point clouds and/or 3D voxelized representations.
  • 3D representations such as 3D meshes, 3D point clouds and/or 3D voxelized representations.
  • the analysis of a 3D representation of the subject may in some instances provide technical improvements over 2D analysis of the same subject.
  • a 3D representation may describe the geometry and/or structure of the subject with less ambiguity than a 2D representation (which may contain shadows and other artifacts which complicate the depiction of depth from the subject and texture of the subject).
  • 3D processing may enable technical improvements because of the inverse optics problem which may, in some instances, affect 2D representations.
  • the inverse optics problem refers to the phenomenon where, in some instances, the size of a subject, the orientation of the subject and the distance between the subject and the imaging device may be conflated in a 2D image of that subject. Any given projection of the subject on the imaging sensor could map to an infinite count of ⁇ size, orientation, distance ⁇ pairings.
  • 3D representations enable the technical improvement in that 3D representations remove the ambiguities introduced by the inverse optics problem.
  • a device that is configured with the dedicated purpose of 3D scanning such as a 3D intraoral scanner (or a CT scanner or MRI scanner), may generate 3D representations of the subject (e.g., the patient's dentition) which have significantly higher fidelity and precision than is possible with a handheld device.
  • 3D intraoral scanner or a CT scanner or MRI scanner
  • 3D representations of the subject e.g., the patient's dentition
  • the use of a 3D autoencoder offers technical improvements (such as increased data precision), to extract the best possible signal out of those 3D data (i.e., to get the signal out of the 3D crown meshes used in tooth classification or setups classification).
  • a 3D autoencoder (comprising a 3D encoder and a 3D decoder) may be trained on 3D data representations to encode an input 3D representation into a latent form (such as a latent vector or a latent capsule) using the 3D encoder, and then reconstruct a facsimile of the input 3D representation using the 3D decoder.
  • a 3D autoencoder for the analysis of a 3D representation (e.g., 3D mesh or 3D point cloud) are 3D convolution, 3D pooling and 3D reconstruction error calculation.
  • a 3D convolution may be performed to aggregate local features from nearby mesh elements.
  • a 3D pooling operation may enable the combining of features from a 3D mesh (or other 3D representation) at multiple scales.
  • 3D pooling may iteratively reduce a 3D mesh into mesh elements which are most highly relevant to a given application (e.g., for which a neural network has been trained). Similarly to 3D convolution, 3D pooling may benefit from special processing beyond that entailed in 2D convolution, to account for the differing count and locations of neighboring mesh elements (relative to a particular mesh element). In some instances, the order of neighboring mesh elements may be less relevant to 3D pooling than to 3D convolution. [0088] 3D reconstruction error may be computed using one or more of the techniques described herein, such as computing Euclidean distances between corresponding mesh elements, between the two meshes. Other techniques are possible in accordance with aspects of this disclosure.
  • 3D reconstruction error may generally be computed on 3D mesh elements, rather than the 2D pixels of 2D reconstruction error.
  • 3D reconstruction error may enable technical improvements over 2D reconstruction error, because a 3D representation may, in some instances, have less ambiguity than a 2D representation (i.e., have less ambiguity in form, shape and/or structure). Additional processing may, in some implementations, be entailed for 3D reconstruction which is above and beyond that of 2D reconstruction, because of the complexity of mapping between the input and reconstructed mesh elements (i.e., the input and reconstructed meshes may have different mesh element counts, and there may be a less clear mapping between mesh elements than there is for the mapping between pixels in 2D reconstruction).
  • the technical improvements of 3D reconstruction error calculation include data precision improvement.
  • a 3D representation may be produced using a 3D scanner, such as an intraoral scanner, a computerized tomography (CT) scanner, ultrasound scanner, a magnetic resonance imaging (MRI) machine or a mobile device which is enabled to perform stereophotogrammetry.
  • a 3D representation may describe the shape and/or structure of a subject.
  • a 3D representation may include one or more 3D mesh, 3D point cloud, and/or a 3D voxelized representation, among others.
  • a 3D mesh includes edges, vertices, or faces. Though interrelated in some instances, these three types of data are distinct. The vertices are the points in 3D space that define the boundaries of the mesh.
  • An edge is described by two points and can also be referred to as a line segment.
  • a face is described by a number of edges and vertices. For instance, in the case of a triangle mesh, a face comprises three vertices, where the vertices are interconnected to form three contiguous edges.
  • Some meshes may contain degenerate elements, such as non-manifold mesh elements, which may be removed, to the benefit of later processing.
  • 3D meshes are commonly formed using triangles, but may in other implementations be formed using quadrilaterals, pentagons, or some other n-sided polygon.
  • a 3D mesh may be converted to one or more voxelized geometries (i.e., comprising voxels), such as in the case that sparse processing is performed.
  • the techniques of this disclosure which operate on 3D meshes may receive as input one or more tooth meshes (e.g., arranged in one or more dental arches).
  • Each of these meshes may undergo pre-processing before being input to the predictive architecture (e.g., including at least one of an encoder, decoder, pyramid encoder-decoder and U-Net).
  • This pre-processing may include the conversion of the mesh into lists of mesh elements, such as vertices, edges, faces or in the case of sparse processing - voxels.
  • feature vectors may be generated. In some examples, one feature vector is generated per vertex of the mesh.
  • Each feature vector may contain a combination of spatial and/or structural features, as specified in the following table: Element Spatial Features Structural Features Edges XYZ position of an edge Edge curvature (depends on a midpoint, XYZ positions of the connectivity neighborhood, edge vertices, or the normal average curvature of two vector at an edge midpoint vertices), dihedral angles, edge (average of the normal vectors length, density measure such as of two vertices). a count of incident edges (i.e., a count of the other neighboring edges which share the vertices of that edge). Faces XYZ position of a face centroid, Face curvature (average surface normal vector.
  • color may be considered as a mesh element feature in addition to the spatial or structural mesh element features described in Table 1.
  • a point differs from a vertex in that a point is part of a 3D point cloud, whereas a vertex is part of a 3D mesh and may have incident faces or edges.
  • a dihedral angle (which may be expressed in either radians or degrees) may be computed as the angle (e.g., a signed angle) between two connected faces (e.g., two faces which are connected along an edge).
  • a sign on a dihedral angle may reveal information about the convexity or concavity of a mesh surface.
  • a positively signed angle may, in some implementations, indicate a convex surface.
  • a negatively signed angle may, in some implementations, indicate a concave surface.
  • directional curvatures may first be calculated to each adjacent vertex around the vertex. These directional curvatures may be sorted in circular order (e.g., 0, 49, 127, 210, 305 degrees) in proximity to the vertex normal vector and may comprise a subsampled version of the complete curvature tensor. Circular order means: sorted in by angle around an axis.
  • the sorted directional curvatures may contribute to a linear system of equations amenable to a closed form solution which may estimate the two principal curvatures and directions, which may characterize the complete curvature tensor.
  • a voxel may also have features which are computed as the aggregates of the other mesh elements (e.g., vertices, edges and faces) which either intersect the voxel or, in some implementations, are predominantly or fully contained within the voxel. Rotating the mesh may not change structural features but may change spatial features.
  • the term “mesh” should be considered in a non- limiting sense to be inclusive of 3D mesh, 3D point cloud and 3D voxelized representation.
  • mesh element features apart from mesh element features, there are alternative methods of describing the geometry of a mesh, such as 3D keypoints and 3D descriptors. Examples of such 3D keypoints and 3D descriptors are found in “TONIONI A, et al. in ‘Learning to detect good 3D keypoints.’, Int J Comput. Vis.2018 Vol .126, pages 1-20.”.3D keypoints and 3D descriptors may, in some implementations, describe extrema (either minima or maxima) of the surface of a 3D representation.
  • one or more mesh element features may be computed, at least in part, via deep feature synthesis (DFS), e.g. as described in: J. M. Kanter and K.
  • DFS deep feature synthesis
  • Mesh element feature vectors may, in some implementations, be computed for a 3D representation which is provided to a Latent Encoding Module 214 or 218 (e.g., a representation generation neural network).
  • a mesh element feature vector may be computed for one or more of the 3D mesh’s mesh elements, to improve the accuracy of the resulting latent representation (e.g., the latent representation generated by the Latent Encoding Module 214 or 218).
  • Representation generation neural networks based on autoencoders, U-Nets, transformers, 3D SWIN transformers, other types of encoder-decoder structures, convolution and/or pooling layers, or other models may benefit from the use of mesh element features.
  • Mesh element features may convey aspects of a 3D representation’s surface shape and/or structure to the neural network models of this disclosure.
  • Each mesh element feature describes distinct information about the 3D representation that may not be redundantly present in other input data that are provided to the neural network. For example, a vertex curvature may quantify aspects of the concavity or convexity of the surface of a 3D representation which would not otherwise be understood by the network. Stated differently, mesh element features may provide a processed version of the structure and/or shape of the 3D representation, data that would not otherwise be available to the neural network. This processed information is often more accessible, or more amenable for encoding by the neural network.
  • a system implementing the techniques disclosed herein has been utilized to run a number of experiments on 3D representations of teeth.
  • mesh element features have been provided to a representation generation neural network which is based on a U-Net model, and also to a representation generation model based on a variational autoencoder with continuous normalizing flows.
  • a full complement of mesh element features e.g., “XYZ” coordinates tuple, “Normal vector”, “Vertex Curvature”, Points-Pivoted, and Normals-Pivoted
  • Points-Pivoted describes “XYZ” coordinates tuples that have local coordinate systems (e.g., at the centroid of the respective tooth).
  • Normals-Pivoted describes “Normal Vectors” which have local coordinate systems (e.g., at the centroid of the respective tooth). Furthermore, training converges more quickly when the full complement of mesh element features are used. Stated another way, the machine learning models trained using the full complement of mesh element features tended to be more accurate more quickly (at earlier epochs) than systems which did not. For an existing system observed to have a historical accuracy rate of 91% accurate, an improvement in accuracy of 3% reduces the actual error rate by more than 30%.
  • Predictive models which may operate on feature vectors of the aforementioned features include but are not limited to: GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, other Denoising Diffusion Models, PT Setups, Similarity Setups, Tooth Classification, Setups Classification, Setups Comparison, VAE Mesh Element Labeling, MAE Mesh In-filling, Mesh Reconstruction Autoencoder, Validation Using Autoencoders, Mesh Segmentation, Coordinate System Prediction, Mesh Cleanup, Restoration Design Generation, Appliance Component Generation and/or Placement, or Archform Prediction.
  • tooth movements may specify one or more tooth transformations that can be encoded in various ways to specify tooth positions and orientations within the setup and are applied to 3D representations of teeth.
  • the tooth positions can be cartesian coordinates of a tooth's canonical origin location which is defined in some semantic context.
  • Tooth orientations can be represented as rotation matrices, unit quaternions, or other 3D rotation representations such as Euler angles with respect to a frame of reference (either global or local).
  • tooth rotations may be described by 3x3 matrices (or by matrices of other dimensions). Tooth position and rotation information may, in some implementations, be combined into the same transform matrix, for example, as a 4x4 matrix, which may reflect homogenous coordinates.
  • affine spatial transformation matrices may be used to describe tooth transformations, for example, the transformations which describe the maloccluded pose of a tooth, an intermediate pose of a tooth and/or a final setup pose of a tooth.
  • Some implementations may use relative coordinates, where setup transformations are predicted relative to malocclusion coordinate systems (e.g., a malocclusion-to- setup transformation is predicted instead of a setup coordinate system directly).
  • Other implementations may use absolute coordinates, where setup coordinate systems are predicted directly for each tooth. In the relative mode, transforms can be computed with respect to the centroid of each tooth mesh (vs the global origin), which is termed “relative local.”
  • relative local coordinates Some of the advantages of using relative local coordinates include eliminating the need for malocclusion coordinate systems (landmarking data) which may not be available for all patient case datasets.
  • Some of the advantages of using absolute coordinates include simplifying the data preprocessing as mesh data are originally represented as relative to the global origin.
  • tooth position encoding and tooth orientation encoding may, in some implementations, also apply one or more of the neural networks models of the present disclosure, including but not limited to: GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, other Denoising Diffusion Models, PT Setups, Similarity Setups, FDG Setups, Setups Classification, Setups Comparison, VAE Mesh Element Labeling, MAE Mesh In-filling, Mesh Reconstruction VAE, and Validation Using Autoencoders.
  • convolution layers in the various 3D neural networks described herein may use edge data to perform mesh convolution.
  • the convolution layers may use vertex data to perform mesh convolution.
  • vertex information is advantageous in that there are typically fewer vertices than edges or faces, so vertex-oriented processing may lead to a lower processing overhead and lower computational cost.
  • the convolution layers may use face data to perform mesh convolution.
  • the convolution layers may use voxel data to perform mesh convolution.
  • voxel information is advantageous in that, depending on the granularity chosen, there may be significantly fewer voxels to process compared to the vertices, edges or faces in the mesh. Sparse processing (with voxels) may lead to a lower processing overhead and lower computational cost (especially in terms of computer memory or RAM usage).
  • Representation generation neural networks based on autoencoders, U-Nets, transformers, other types of encoder-decoder structures, convolution and/or pooling layers, or other models may benefit from the use of oral care arguments (e.g., oral care metrics or oral care parameters).
  • oral care metrics may convey aspects of the shape and/or structure of the patient’s dentition (e.g., the shape and/or structure of an individual tooth, or the special relationships between two or more teeth) to the neural network models of this disclosure.
  • Each oral care metric describes distinct information about the patient’s dentition that may not be redundantly present in other input data that are provided to the neural network.
  • an “Overbite” metric may quantify the overlap between the upper and lower central incisors along the vertical Z-axis, information which may not otherwise, in some implementations, be readily ascertainable by a traditional neural network.
  • the oral care metrics provide refined information about the patient’s dentition that a traditional neural network (e.g., a representation generation neural network) may not be adequately trained or configured to extract the oral care metrics described herein.
  • a neural network which is specifically trained to generate oral care metrics may overcome such a shortcoming, because, for example loss may be computed in such a way as to facilitate accurate oral care metrics prediction.
  • Mesh oral care metrics may provide a processed version of the structure and/or shape of the patient’s dentition, data which may not otherwise be available to the neural network. This processed information is often more accessible, or more amenable for encoding by the neural network.
  • a system implementing the techniques disclosed herein has been utilized to run a number of experiments on 3D representations of teeth.
  • oral care metrics have been provided to a representation generation neural network which is based on a U-Net model. Based on experiments, it was found that systems using oral care metrics (e.g., “Overbite”, “Overjet” and “Canine Class Relationship” metrics) were at least 2.5% more accurate than systems that did not. Furthermore, training converges more quickly when the oral care metrics are used. Stated another way, the machine learning models trained using oral care metrics tended to be more accurate more quickly (at earlier epochs) than systems which did not. For an existing system observed to have a historical accuracy rate of 91%, an improvement in accuracy of 2.5% reduces the actual error rate by almost 30%. [0097] Oral care arguments may include oral care parameters, or oral care metrics.
  • RDM may describe the shape and/or form of one or more 3D representations of teeth for use in dental restoration.
  • One use case example is in the creation of one or more dental restoration appliances.
  • Another use case example is in the creation of one or more veneers (such as a zirconia veneer).
  • Some RDM may quantify the shape and/or other characteristics of a tooth.
  • Other RDM may quantify relationships (e.g., spatial relationships) between two or more teeth.
  • RDM differ from restoration design parameters (RDP) in that restoration design metrics define a current state of a patient's dentition, whereas restoration design parameters serve as specifications to a machine learning or other optimization model to generate desired tooth shapes and/or forms.
  • RDM describe the shapes of the teeth currently (e.g., in a starting or mal condition).
  • Restoration design parameters specify how an oral care provider (such as a dentist or dental technician) intends for the teeth to look after the completion of restoration treatment.
  • Either or both of RDM and RDP may be provided a neural network or other machine learning or optimization algorithm for the purpose of dental restoration.
  • RDM may be computed on the pre- restoration dentition of the patient (i.e., the primary implementation).
  • RDM may be computed on the post-restoration dentition of the patient.
  • a restoration design may comprise one or more teeth and may be referred to as a restoration arch. Restoration design generation may involve the generation of an improved geometry and/or structure of one or more teeth in a restoration arch.
  • RDM may be measured, for example, through locating landmarks in the teeth (or gums, hardware and/or other elements of the patient's dentition), and the measurements of distances between those landmarks, or otherwise made in relation to those landmarks.
  • one or more neural networks or other machine learning models may be trained to identify or extract one or more RDM from one or more 3D representations of teeth (or gums, hardware and/or other elements of the patient's dentition). Techniques of this disclosure may use RDM in various ways.
  • one or more neural networks or other machine learning models may be trained to classify or label one or more setups, arches, dentitions or other sets of teeth based at least in part on RDM.
  • RDMs form a part of the training data used for training these models.
  • This autoencoder (e.g., a variational autoencoder or VAE) takes as input a tooth mesh (or other 3D representation) that reflects a mal state (i.e., the pre-restoration tooth shape).
  • the encoder component of the autoencoder encodes that tooth mesh to a latent form (e.g., a latent vector). Modifications may be applied to this latent vector (e.g., based on a mapping of the latent space through prior experiments), for the purpose of altering the geometry and/or structure of the eventual reconstructed mesh.
  • Additional vectors may, in some implementations, be included with the latent vector (e.g., through concatenation), and the resulting concatenation of vectors may be reconstructed by way of the decoder component of the autoencoder into a reconstructed tooth mesh which is a facsimile of the input tooth mesh.
  • RDM and RDP may also be used as neural network inputs in the execution phase, in accordance with aspects of this disclosure.
  • one or more RDM may be concatenated with the input to the encoder, for the purpose of telling the encoder specific information about the input 3D tooth representation.
  • one or more RDM may be concatenated with the latent vector, before reconstruction, for the purpose of providing the decoder component with specific information about the input 3D tooth representation.
  • one or more restoration design parameters (RDP) may be concatenated with the input to the encoder component, for the purpose of providing the encoder specific information about the input 3D tooth representation.
  • one or more restoration design parameters (RDP) may be concatenated with the latent vector, before reconstruction, for the purpose of providing the decoder specific information about the input 3D tooth representation.
  • either or both of RDM and RDP may be introduced to the functioning of an autoencoder (e.g., a tooth reconstruction autoencoder), and serve to influence the geometry and/or structure of the reconstructed restoration design (i.e., influence the shape of the tooth on the output of the autoencoder).
  • an autoencoder e.g., a tooth reconstruction autoencoder
  • the variational autoencoder of US Provisional Application No. US63/366514 may be replaced by a capsule autoencoder (e.g., instead of encoding the tooth mesh to a latent vector, the tooth mesh is encoded to one or more latent capsules).
  • clustering or other unsupervised techniques may be performed on RDM to cluster one or more setups, arches, dentitions, or other sets of teeth based on the restoration characteristics of the teeth.
  • Such clusters may be useful in treatment planning, as the clusters provide insight into categories of patients with different treatment needs. This information may be instructive to clinicians as they learn about possible treatment options.
  • best practices may be identified (such as default RDP values) for patient cases that fall into one or another cluster (e.g., as determined by a similarity measure, as in k-NN). After a new case is classified into a particular cluster, information about the relevant best practices may be provided to the clinician who is responsible for processing the case. Such default values may, in some instances, undergo further tuning or modifications.
  • Case Assignment Such clusters may be used to gain further insight into the kinds of patient cases which exist in a dataset. Analysis of such clusters may reveal that patient treatment cases with certain RDM values (or ranges of values) may take less time to treat (or alternatively more time to treat). Cases which take more time to treat (or are otherwise more difficult) may be assigned to experienced or senior technicians for processing. Cases which take less time to treat may be assigned to newer or less- experienced techniques for processing. Such an assignment may be further aided by finding correlations between RDM values for certain cases and the known processing durations associated with those cases.
  • the following RDM may be measured and used in the creation of either or both of dental restoration appliances and veneers ⁇ veneers are a type of dental restoration appliance ⁇ , with the objective of making the resulting teeth natural looking. Symmetry is generally a preferred facet. There may be differences between patients based on demographic differences. The generation of dental restoration appliances may benefit from some or all of the following RDM. Shade and translucency may pertain, in particular, to the creation of veneers, though some implementations of dental restoration appliances may also consider this information. [00105] Examples of inter-tooth RDM are enumerated below. [00106] 1) Bilateral Symmetry and/or Ratios: A measure of the symmetry between one or more teeth and one or more other teeth on opposite sides of the dental.
  • a measure of the width of each tooth For example, for a pair of corresponding teeth, a measure of the width of each tooth. In one instance, the one tooth is of normal width, and the other tooth is too narrow. In another instance, both teeth are of normal width.
  • the following is a list of attributes that can be measured for a tooth, and compared to the corresponding measurement for one or more corresponding teeth: a) width - mesial to distal distance; b) length - gingival to incisal distance; c) diagonal - distance across the tooth, e.g., from the mesial gingival corner to the distal incisal corner (this measure is one of many that can be used to quantify the shape of teeth beyond length and width).
  • Ratios between a and b may be computed, such as a/b or b/a. Such ratios can be indicative of whether spatial symmetry exists (e.g., by measuring the ratio a/b on the left side and measuring the ratio a/b on the right side, then compare the left and right ratios). In some implementations, where spatial symmetry is "off", the length, width and/or ratios may not match. Such a ratio may, in some implementations, be computed relative to a standard. A number of esthetic standards are available in the dental literature. Examples include Golden Proportion and Recurring Esthetic Dental Proportion.
  • spatial symmetry may be measured on a pair of teeth, where one tooth is on the right side of the arch, and the other tooth is on the left side of the arch.
  • Proportions of Adjacent Teeth Measure the width proportions of adjacent teeth as measured as a projection along an arch onto a plane (e.g., a plane that is situated in front of the patient's face).
  • the ideal proportions for use in the final restoration design can be, for example, the so-called golden proportions.
  • the golden proportions relate adjacent teeth, such as central incisors and lateral incisors. This metric pertains to the measuring of these proportions as the proportions exist in the pre- restoration mal dentition.
  • the ideal golden proportions are 1.6, 1, 0.6, for the central incisor, lateral incisor and cuspid, on a particular side (either left or right) for a particular arch (e.g., the upper arch). If one or more of these proportion values is off (e.g., in the case of "peg laterals"), the patient may wish for dental restoration treatment to correct the proportions.
  • Arch Discrepancies A measure of any size discrepancies between the upper arch and lower arch, for example, pertaining to the widths of the teeth, for the purpose of dental restoration. For example, techniques of this disclosure may make adjacent tooth width proportion measurements in the upper arch and in the lower arch.
  • Bolton analysis measurements may be made by measuring upper widths, lower widths, and proportions between those quantities. Arch discrepancies may be described in absolute measurements (e.g., in mm or other suitable units) or in terms of proportions or ratios, in various implementations.
  • Midline A measure of the midline of the maxillary incisors, relative to the midline of the mandibular incisors. Techniques of this disclosure may measure the midline of the maxillary incisors, relative to the midline of the nose (if data about nose location is available).
  • Proximal Contacts A measure of the size (area, volume, circumference, etc.) of the proximal contact between adjacent teeth.
  • the teeth touch along the mesial/distal surfaces and the gums fill in gingivally to where the teeth touch.
  • Black triangles may form if the gum tissue fails to fill the space below the proximal contact.
  • the size of the proximal contact may get progressively shorter for teeth located farther towards the posterior of the arch.
  • the proximal contact would be long enough so that there is an appropriately sized incisal embrasure and the gum tissue fills in the area below or gingival to the contact.
  • Embrasure In some implementations, techniques of this disclosure may measure the size (area, volume, circumference, etc.) of an embrasure, the gap between teeth at either of the gingival or incisal edge. In some implementations, techniques of this disclosure may measure the symmetry between embrasures on opposite sides of the arch. An embrasure is based at least in part on the length of the length of the contact between teeth, and/or at least in part on the shape of the tooth. In some instances, the size of the embrasure may get progressively longer for teeth located farther towards the posterior of the arch. [00112] Examples of Intra-tooth RDM are enumerated below, continuing with the numbering of other RDM listed above.
  • Length and/or Width A measure of the length of a tooth relative to the width of that tooth. This metric may reveal, for example, that a patient has long central incisors. Width and length are defined as: a) width - mesial to distal distance; b) length - gingival to incisal distance; c) other dimensions of tooth body - the portions of tooth between the gingival region and the incisal edge. In some implementations, either or both of a length and a width may be measured for a tooth and compared to the length and/or width of one or more teeth.
  • Tooth Morphology A measure of the primary anatomy of the tooth shape, such as line angles, buccal contours, and/or incisal angles and/or embrasures.
  • the frequency and/or dimensions may be measured.
  • the observed primary tooth shape aspects may be matched to one or more known styles.
  • Techniques of this disclosure may measure secondary anatomy of the tooth shape, such as mamelon grooves. For instance, the frequency and/or dimensions may be measured.
  • the observed secondary tooth shape aspects may be matched to one or more known styles.
  • techniques of this disclosure may measure tertiary anatomy of the tooth shape, such as perikymata or striations. For instance, the frequency and/or dimensions may be measured.
  • the observed tertiary tooth shape aspects may be matched to one or more known styles.
  • Shade and/or Translucency A measure of tooth shade and/or translucency. Tooth shade is often described by the Vita Classical or 3D Master shade guide. Tooth translucency is described by transmittance or a contrast ratio. Tooth shade and translucency may be evaluated (or measured) based on one or more of the following kinds of data pertaining to teeth: the incisal edge, incisal third, body and gingival third. The enamel layer translucency is general higher than the dentin or cementum layer. Shade and translucency may, in some implementations, be measured on a per-voxel (local) basis.
  • Shade and translucency may, in some implementations, be measured on a per-area basis, such as an incisal area, tooth body area, etc. Tooth body may pertain to the portions of the tooth between the gingival region and the incisal edge.
  • Height of Contour A measure of the contour of a tooth. When viewed from the proximal view, all teeth have a specific contour or shape, moving from the gingival aspect to the incisal. This is referred to as the facial contour of the tooth. In each tooth, there is a height of contour, where that shape is the most pronounced. This height of contour changes from the teeth in the anterior of the arch to the teeth in the posterior of the arch.
  • this measurement may take the form of fitting against a template of known dimensions and/or known proportions. In some implementations, this measurement may quantify a degree of curvature along the facial tooth surface. In some implementations, measure the location along the contour of the tooth where the height of the curvature is most pronounced. This location may be measured as a distance away from the gingival margin or a distance away from the incisal edge, or a percentage along the length of the tooth.
  • the orthodontic metrics may be used to quantify the physical arrangement of an arch of teeth for the purpose of orthodontic treatment (as opposed to restoration design metrics – which pertain to dentistry and describe the shape and/or form of one or more pre-restoration teeth, for the purpose of supporting dental restoration). These orthodontic metrics can measure how badly maloccluded the arch is, or conversely the metrics can measure how correctly arranged the teeth are.
  • the GDL Setups model (or RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, PT Setups, Similarity Setups and FDG Setups) may incorporate one or more of these orthodontic metrics, or other similar or related orthodontic metrics.
  • such orthodontic metrics may be incorporated into the feature vector for a mesh element, where these per- element feature vectors are provided to the setups prediction network as inputs.
  • such orthodontic metrics may be directly consumed by a generator, an MLP, a transformer, or other neural network as direct inputs (such as presented in one or more input vectors of real numbers S, such as described elsewhere in this disclosure.
  • the use of such orthodontic metrics in the training of the generator may improve the performance (i.e., correctness) of the resulting generator, resulting in predicted transforms which place teeth more nearly in the correct final setups poses than would otherwise be possible.
  • Such orthodontic metrics may be provided to an encoder structure or by a U-Net structure (in the case of GDL Setups).
  • Such orthodontic metrics may be consumed by an autoencoder, variational autoencoder, masked autoencoder or regularized autoencoder (in the case of the VAE Setups, VAE Mesh Element Labelling, MAE Mesh In-Filling).
  • Such orthodontic metrics may be consumed by a neural network which generates action predictions as a part of a reinforcement learning RL Setups model.
  • Such orthodontic metrics may be consumed by a classifier which applies a label to a setup arch (e.g., labels such as mal, staging or final setup). This description is non-limiting, as the orthodontic metrics may also be incorporated in other ways into the various techniques of this disclosure.
  • the various loss calculations of the present disclosure may, in some examples, incorporate one or more orthodontic metrics, with the advantage of improving the correctness of the resulting neural network.
  • An orthodontic metric may be used to directly compare a predicted example to the corresponding ground truth example (such as is done with the metrics in the Setups Comparison description).
  • one or more orthodontic metrics may be taken from this section and incorporated into a loss computation.
  • Such an orthodontic metric may be computed on the predicted example, and then the orthodontic metric would also be computed on the ground truth example.
  • one or more orthodontic metrics pertaining to the alignment of two or more adjacent teeth may be computed and incorporated into a loss function, for example, to train, at least in part, a setups prediction neural network.
  • such an orthodontic metric may influence the network to align the mesial surface of a tooth with the distal surface of an adjacent tooth.
  • Backpropagation is an example algorithm by which a neural network may be trained using one or more loss values.
  • one or more orthodontic metrics may be used to evaluate the predicted output of a neural network, such as a setups prediction. Such a metric(s) may enable the training algorithm to determine how close the predicted output is to an acceptable output, for example, in a quantified sense.
  • this use of an orthodontic metric may enable a loss value to be computed which does not depend entirely on a comparison to a ground truth.
  • such a use of an orthodontic metric may enable loss calculation and network training to proceed without the need for a comparison against a ground truth example.
  • the advantage of such an approach is that loss may be computed based on a general principle or specification for the predicted output (such as a setup) rather than tying loss calculation to a specific ground truth example (which may have been defined by a particular doctor, clinician, or technician, whose treatment philosophy may differ from that of other technicians or doctors).
  • such an orthodontic metric may be defined based on a FID (Frechet Inception Distance) score.
  • An orthodontic metric that can be computed using tensors may be especially advantageous when training one of the neural networks of the present disclosure, because tensor operations may promote efficient computations. The more efficient (and faster) the computation, the faster the rate at which training can proceed.
  • an error pattern may be identified in one or more predicted outputs of an ML model (e.g., a transformation matrix for a predicted tooth setup, a labelling of mesh elements for mesh cleanup, an addition of mesh elements to a mesh for the purpose of mesh in-filling, a classification label for a setup, a classification label for a tooth mesh, etc.).
  • One or more orthodontic metrics may be selected to become an input to the next round of ML model training, to address any pattern of errors or deficiencies which may be identified in the one or more predicted outputs.
  • Some OM may be defined relative to an archfrom coordinate frame, the LDE coordinate system.
  • a point may be described using an LDE coordinate frame relative to an archform, where L, D and E correspond to: 1) Length along the curve of the archform, 2) Distance away from the archform, and 3) distance in the direction perpendicular to the L and D axes (which may be termed Eminence), respectively.
  • L, D and E correspond to: 1) Length along the curve of the archform, 2) Distance away from the archform, and 3) distance in the direction perpendicular to the L and D axes (which may be termed Eminence), respectively.
  • Various of the OM and other techniques of the present disclosure may compute collisions between 3D representations (e.g., of oral care objects, such as teeth). Such collisions may be computed as at least one of: 1) penetration distance between 3D tooth representations, 2) count of overlapping mesh elements between 3D tooth representations, and 3) volume of overlap between 3D tooth representations.
  • an OM may be defined to quantify the collision of two or more 3D representations of oral care structures, such as teeth.
  • Some optimization algorithms such as setups prediction techniques, may seek to minimize collisions between oral care structures (such as teeth).
  • Between-arch orthodontic metrics are as follows. [00126] Six (6) metrics for the comparison of two or more arches are listed below. Other suitable comparison orthodontic metrics are found elsewhere in this disclosure, such as in the section for the Setups Comparison technique. 1. Rotation geodesic distance (rotation between predicted example and ground truth setup example) 2. Translation distance (gap between predicted example and ground truth setup example) 3. Normalized translation distance 4. 3D alignment error that measures the distance between predicted mesh elements and ground truth mesh elements, in units of mm. 5. Normalized 3D alignment 6.
  • Percent overlap (% overlap) by volume (alternatively % overlap by mesh elements) of predicted example and corresponding ground truth example [00127]
  • a 3D tooth orientation vector may be calculated using the tooth's mesial-distal axis.
  • a 3D vector which may be tangent vector to the archform at the position of the tooth may also be calculated.
  • the XY components i.e., which may be 2D vectors
  • Cosine similarity may be used to calculate the 2D orientation difference (angle) between the archform tangent and the tooth's mesial-distal axis.
  • the absolute difference may be calculated between each tooth’s X-coordinate and the global coordinate reference frame’s X-axis.
  • This delta may indicate the arch asymmetry for a given tooth pair.
  • the result of such a calculation may be the mean X-axis delta of one or more tooth-pairs from the arch. This calculation may, in some implementations, be performed relative to the Y-axis with y-coordinates (and/or relative to the Z axis with Z-coordinates).
  • Archform D-axis Differences – May compute the D dimension difference (i.e., the positional difference in the facial-lingual direction) between two arch states, for one or more teeth. May, in some implementations, return a dictionary of the D-direction tooth movement for each tooth, with tooth UNS number as the key. May use the LDE coordinate system relative to an archform.
  • Archform (Lower) Length Ratio – May compute the ratio between the current lower arch length and the arch length as it was in the original maloccluded lower arch.
  • Archform (Upper) Length Ratio – May compute the ratio between the current upper arch length and the arch length as it was in the original maloccluded upper arch.
  • Archform Parallelism (Full arch) - For at least one local tooth coordinate system origin in the upper arch, the one or more nearest origins (e.g., tooth local coordinate system origins) in the lower arch.
  • the two nearest origins may be used. May compute the straight line distance from the upper arch point to the line formed between the origins of the two teeth in the opposing (lower) arch. May return the standard deviation of the set of “point-to-line” distances mentioned above, where the set may be composed of the point-to-line distances for each tooth in the arch.
  • This metric may share some computational elements with the archform_parallelism_global orthodontic metric, except that this metric may input the mean distance from a tooth origin to the line formed by the neighboring teeth in opposing arches (e.g., a tooth in the upper arch and the corresponding tooth in the lower arch). The mean distance may be computed for one or more such pairs of teeth. In some implementations, this may be computed for all pairs of teeth. Then the mean distance may be subtracted from the distance that is computed for each tooth pair. This OM may yield the deviation of a tooth from a “typical” tooth parallelism in the arch.
  • Buccolingual Inclination For at least one molar or premolar, find the corresponding tooth on the opposite side of the same arch (i.e., for a tooth on the left side of the arch, find the same type of tooth on the right side and vice versa).
  • This OM may compute an n-element list for each tooth (e.g. n may equal 2).
  • Such an n-element vector may be computed for each molar and each premolar in the upper and lower arches.
  • the buccal cusps may be identified on the molars and premolars on each of the left and right sides of the arch. Draw a line between the buccal cusps of the left tooth and the buccal cusps on the right tooth. Make a plane using this line and the z-axis of the arch. The lingual cusps may be projected onto the plane (i.e., at this point the angle of inclination may be determined). By performing an additional projection, the approximate vertical distance between the lingual cusps and the buccal cusps may be computed. This distance may be used as the buccolingual inclination OM. Canine Overbite - The upper and lower canines may be identified. The first premolar for the given side of the mouth may be identified.
  • a distance may be computed between the upper canine and the lower canine, and also between the upper pre-molar and the lower pre-molar.
  • the average (or median, or mode or some other statistic) may be computed for the measured distances.
  • the z- component of this result indicates the degree of overbite.
  • Overbite may be computed between any tooth in one arch and the corresponding tooth in the other arch.
  • Canine Overjet Contact – May calculate the collisions (e.g., collision distances) between pairs of canines on opposing arches.
  • Canine Overjet Contact KDE – May take an orthodontic metric score for the current patient case as input, and may convert that score into to a log-likelihood using a previously trained kernel density estimation (KDE) model or distribution. This operation may yield information about where in the distribution of "typical" values this patient case lies.
  • Canine Overjet – This OM may share some computational steps with the canine overbite OM. In some implementations, average distances may be computed. In some implementations, the distance calculation may compute the Euclidean distance of the XY components of a tooth in the upper arch and a tooth in the lower arch, to yield overjet (i.e., as opposed to computing the difference in Z-components, as may be performed for canine overbite).
  • Overjet may be computed between any tooth in one arch and the corresponding tooth in the other arch.
  • Canine Class Relationship also applies to first, second and third molars
  • This OM may, in some implementations comprise two functions (e.g., written in Python).
  • get_canine_landmarks() Get landmarks for each tooth which may be used to compute the class relationship, and then, in some implementations, map those landmarks onto the global coordinate space so that measurements may be made between teeth.
  • class_relationship_score_by_side() May compute the average position of at least one landmark on at least one tooth in the lower arch, and may compute the same for the upper arch.
  • This OM may compute how far forward or behind the tooth is positioned on the l-axis relative to the tooth or teeth of interest in the opposing arch.
  • Crossbite - Fossa in at least one upper molar may be located by finding the halfway point between distal and mesial marginal ridge saddles of the tooth.
  • a lower molar cusp may lie between the marginal ridges of the corresponding upper molar.
  • This OM may compute a vector from the upper molar fossa midpoint to the lower molar cusp. This vector may be projected onto the d-axis of the archform, yielding a lateral measure of distance from the cusp to the fossa. This distance may define the crossbite magnitude.
  • Edge Alignment – This OM may identify the leftmost and rightmost edges of a tooth, and may identify the same for that tooth’s neighbor. The OM may then draw a vector from the leftmost edge of the tooth to the leftmost edge of the tooth’s neighbor. The OM may then draw a vector from the rightmost edge of the tooth to the rightmost edge of the tooth’s neighbor. The OM may then calculates the linear fit error between the two vectors.
  • Incisor Interarch Contact KDE May identify the deviation of the IncisorInterarchContact from the mean of a modeled distribution of such statistics across a dataset of one or more other patient cases.
  • This OM may calculate the difference in height between two or more neighboring teeth. For molars, this OM may use the midpoint between the mesial and distal saddle ridges as the height of the molar. For non-molar teeth, this OM may use the length of the crown from gums to tip. In some implementations, the tip may be the origin of the local coordinate space of the tooth. Other implementations may place the origin in other locations. A simple subtraction between the heights of neighboring teeth may yield the leveling delta between the teeth (e.g., by comparing Z components).
  • Midline – May compute the position of the midline for the upper incisors and/or the lower incisors, and then may compute the distance between them.
  • Molar Interarch Contact KDE – May compute a molar interarch contact score (i.e., a collision depth or other type of collision), and then may identify where that score lies in a pre-defined KDE (distribution) built from representative cases.
  • the cusp may be scored according to how well the cusp contacts the neighboring (corresponding) tooth in the opposite arch.
  • a vector may be found from the cusp of the tooth in question to the vertical intersection point in the corresponding tooth of the opposing arch.
  • the distance and/or direction (i.e., up or down) to the opposing arch may be computed.
  • a list may be returned that contains the resulting signed distances, one for each cusp on the tooth in question.
  • Overbite The upper and lower central incisors may be compared along the z-axis. The difference along the z-axis may be used as the overbite score.
  • Overjet The upper and lower central incisors may be compared along the y-axis. The difference along the y-axis may be used as the overjet score.
  • Molar Interarch Contact — May calculate the contact score between molars and may use collision measurement(s) (such as collision depth).
  • Root Movement d – The tooth transforms for an initial state and a next state may be recieved.
  • the archform axes at a point L along the archform may be computed. This OM may return a distance moved along the d-axis. This may be accomplished by projecting the root pivot point onto the d-axis.
  • Root Movement l – The tooth transforms for an initial state and a next state may be received.
  • the archform axes at a point L along the archform may be computed.
  • This OM may return a distance moved along the l-axis. This may be accomplished by projecting the root pivot point onto the l-axis. Spacing – May compute the spacing between each tooth and its neighbor.
  • the transforms and meshes for the arch may be received.
  • the left and right edges of each tooth mesh may be computed.
  • One or more points of interest may be transformed from local coordinates into the global arch coordinate frame.
  • the spacing may be computed in a plane (e.g., the XY plane) between each tooth and its neighbor to the "left".
  • Torque – May return an array of one or more Euclidean distances (e.g., such as in the XY plane) which may represent the spacing between each tooth and its neighbor to the left.
  • Torque – May compute torque (i.e., rotation around and axis, such as the x-axis).
  • torque i.e., rotation around and axis, such as the x-axis.
  • For one or more teeth one or more rotations may be converted from Euler angles into one or more rotation matrices.
  • a component (such as a x-component) of the rotations may be extracted and converted back into Euler angles. This x- component may be interpreted as the torque for a tooth.
  • a list may be returned which contains the torque for one or more teeth, and may be indexed by the UNS number of the tooth.
  • the neural networks of this disclosure may exploit one or more benefits of the operation of parameter tuning, whereby the inputs and parameters of a neural network are optimized to produce more data-precise results.
  • One parameter which may be tuned is neural network learning rate (e.g., which may have values such as 0.1, 0.01, 0.001, etc.).
  • Data augmentation schemes may also be tuned or optimized, such as schemes where “shiver” is added to the tooth meshes before being input to the neural network (i.e., small random rotations, translations and/or scaling may be applied to vary the dataset and make the neural network robust to variations in data).
  • a subset of the neural network model parameters available for tuning are as follows: o Learning rate (LR) decay rate (e.g., how much the LR decays during a training run) o Learning rate (LR). The floating-point value (e.g., 0.001) that is used by the optimizer.
  • LR schedule e.g., cosine annealing, step, exponential
  • Voxel size for cases with sparse mesh processing operations
  • Dropout % e.g., dropout which may be performed in a linear encoder
  • LR decay step size e.g., decay every 10 or 20 or 30 epochs
  • Model scaling which may increase or decrease the count of layers and/or the count of parameters per layer.
  • Parameter tuning may be advantageously applied to the training of a neural network for the prediction of final setups or intermediate staging to provide data precision-oriented technical improvements. Parameter tuning may also be advantageously applied to the training of a neural network for mesh element labeling or a neural network for mesh in-filling. In some examples, parameter tuning may be advantageously applied to the training of a neural network for tooth reconstruction. In terms of classifier models of this disclosure, parameter tuning may be advantageously applied to a neural network for the classification of one or more setups (i.e., classification of one or more arrangements of teeth). The advantage of parameter tuning is to improve the data precision of the output of a predictive model or a classification model.
  • Parameter tuning may, in some instances, provide the advantage of obtaining the last remaining few percentage points of validation accuracy out of a predictive or classification model.
  • Various neural network models of this disclosure may draw benefits from data augmentation. Examples include models of this which are trained on 3D meshes, such as GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, PT Setups, Similarity Setups, FDG Setups, Setups Classification, Setups Comparison, VAE Mesh Element Labeling, MAE Mesh In-filling, Mesh Reconstruction VAE, and Validation Using Autoencoders.
  • Data augmentation may increase the size of the training dataset of dental arches.
  • Data augmentation can provide additional training examples by adding random rotations, translations, and/or rescaling to copies of existing dental arches.
  • data augmentation may be carried out by perturbing or jittering the vertices of the mesh, in a manner similar to that described in (“Equidistant and Uniform Data Augmentation for 3D Objects”, IEEE Access, Digital Object Identifier 10.1109/ACCESS.2021.3138162).
  • the position of a vertex may be perturbed through the addition of Gaussian noise, for example with zero mean, and 0.1 standard deviation. Other mean and standard deviation values are possible in accordance with the techniques of this disclosure.
  • Some techniques of the present disclosure may benefit from a processing step which may align (or register) arches of teeth (e.g., where a tooth may be represented by a 3D point cloud, or some other type of 3D representation described herein).
  • a processing setup may, for example, be used to register a ground truth setup arch from a patient case with the maloccluded arch from that same case, before these mal and ground truth setup arches are used to train a setups prediction neural network model.
  • Such a step may aid in loss calculation, because the predicted arch (e.g., an arch outputted by a generator) may be in better alignment with the ground truth setup arch, a condition which may facilitate the calculation of reconstruction loss, representation loss, L1 loss, L2 loss, MSE loss and/or other kinds of losses described herein.
  • an iterative closest point (ICP) technique may be used for such registration. ICP may minimize the squared errors between corresponding entities, such as 3D representations.
  • linear least squares calculations may be performed.
  • non-linear least squares calculations may be performed.
  • Various registration models may incorporate portions of the following algorithms, in whole or in part: Levenberg-Marquardt ICP, Least Square Rigid transformation, Robust Rigid transformation, random sample consensus (RANSAC) ICP, K-means based RANSAC ICP and Generalized ICP (GICP).
  • Registration may, in some instances, help decrease the subjectivity and/or randomness that may, in some instances, occur in reference ground truth setup designs which have been designed by technicians (i.e., two technicians may produce different but valid final setups outputs for the same case) or by other optimization techniques.
  • the ground truth (or reference) setup was registered to the malocclusion (or maloccluded setup).
  • the maloccluded teeth were provided to the setups prediction model, which generated final setup transforms for the maloccluded teeth. Loss was computed between the resulting predicted setup and the pre-registered ground truth setup, so that corresponding aspects of the two setups would line-up. The result was a more accurate loss calculation. This pre-registration operation resulted in a 6% improvement in absolute accuracy (e.g., as measured by ADD10 score), which amounts to a reduction in error rate of nearly 50% compared with conventional techniques. [00133] Because generator networks of this disclosure can be implemented as one or more neural networks, the generator may contain an activation function.
  • an activation function When executed, an activation function outputs a determination of whether or not a neuron in a neural network will fire (e.g., send output to the next layer).
  • Some activation functions may include: binary step functions, or linear activation functions.
  • Other activation functions impart non-linear behavior to the network, including: sigmoid/logistic activation functions, Tanh (hyperbolic tangent) functions, rectified linear units (ReLU), leaky ReLU functions, parametric ReLU functions, exponential linear units (ELU), softmax function, swish function, Gaussian error linear unit (GELU), or scaled exponential linear unit (SELU).
  • a linear activation function may be well suited to some regression applications (among other applications), in an output layer.
  • a sigmoid/logistic activation function may be well suited to some binary classification applications (among other applications), in an output layer.
  • a softmax activation function may be well suited to some multiclass classification applications (among other applications), in an output layer.
  • a sigmoid activation function may be well suited to some multilabel classification applications (among other applications), in an output layer.
  • a ReLU activation function may be well suited in some convolutional neural network (CNN) applications (among other applications), in a hidden layer.
  • a Tanh and/or sigmoid activation function may be well suited in some recurrent neural network (RNN) applications (among other applications), for example, in a hidden layer.
  • optimization algorithms which can be used in the training of the neural networks of this disclosure (such as in updating the neural network weights), including gradient descent (which determines a training gradient using first-order derivatives and is commonly used in the training of neural networks), Newton's method (which may make use of second derivatives in loss calculation to find better training directions than gradient descent, but may require calculations involving Hessian matrices), and conjugate gradient methods (which may yield faster convergence than gradient descent, but do not require the Hessian matrix calculations which may be required by Newton's method).
  • additional methods may be employed to update weights, in addition to or in place of the techniques described above. These additional methods include the Levenberg-Marquardt method and/or simulated annealing.
  • Neural networks contribute to the functioning of many of the applications of the present disclosure, including but not limited to: GDL Setups, RL Setups, VAE Setups, Capsule Setups, MLP Setups, Diffusion Setups, PT Setups, Similarity Setups, Tooth Classification, Setups Classification, Setups Comparison, VAE Mesh Element Labeling, MAE Mesh In-filling, Mesh Reconstruction Autoencoder, Validation Using Autoencoders, imputation of oral care parameters, 3D mesh segmentation (3D representation segmentation), Coordinate System Prediction, Mesh Cleanup, Restoration Design Generation, Appliance Component Generation and/or Placement, or Archform Prediction.
  • the neural networks of the present disclosure may embody part or all of a variety of different neural network models. Examples include the U-Net architecture, multi-later perceptron (MLP), transformer, pyramid architecture, recurrent neural network (RNN), autoencoder, variational autoencoder, regularized autoencoder, conditional autoencoder, capsule network, capsule autoencoder, stacked capsule autoencoder, denoising autoencoder, sparse autoencoder, conditional autoencoder, long/short term memory (LSTM), gated recurrent unit (GRU), deep belief network (DBN), deep convolutional network (DCN), deep convolutional inverse graphics network (DCIGN), liquid state machine (LSM), extreme learning machine (ELM), echo state network (ESN), deep residual network (DRN), Kohonen network (KN), neural Turing machine (NTM), or generative adversarial network (GAN).
  • U-Net architecture multi-later perceptron (MLP), transformer, pyramid architecture, recurrent
  • an encoder structure or a decoder structure may be used.
  • Each of these models provides one or more of its own particular advantages.
  • a particular neural networks architecture may be especially well suited to a particular ML technique.
  • autoencoders are particularly suited to the classification of 3D oral care representations, due to the ability to encode the 3D oral care representation into a form which is more easily classifiable.
  • the neural networks of this disclosure can be adapted to operate on 3D point cloud data (alternatively on 3D meshes or 3D voxelized representation).
  • Numerous neural network implementations may be applied to the processing of 3D representations and may be applied to training predictive and/or generative models for oral care applications, including: PointNet, PointNet++, SO-Net, spherical convolutions, Monte Carlo convolutions and dynamic graph networks, PointCNN, ResNet, MeshNet, DGCNN, VoxNet, 3D-ShapeNets, Kd-Net, Point GCN, Grid-GCN, KCNet, PD-Flow, PU-Flow, MeshCNN and DSG-Net.
  • Oral care applications include, but are not limited to: setups prediction (e.g., using VAE, RL, MLP, GDL, Capsule, Diffusion, etc.
  • Autoencoders that can be used in accordance with aspects of this disclosure include but are not limited to: AtlasNet, FoldingNet and 3D-PointCapsNet. Some autoencoders may be implemented based on PointNet.
  • Representation learning may be applied to setups prediction techniques of this disclosure by training a neural network to learn a representation of the teeth, and then using another neural network to generate transforms for the teeth.
  • Some implementations may use a VAE or a Capsule Autoencoder to generate a representation of the reconstruction characteristics of the one or more meshes related to the oral care domain (including, in some instances, information about the structures of the tooth meshes).
  • that representation (either a latent vector or a latent capsule) may be used as input to a module which generates the one or more transforms for the one or more teeth.
  • These transforms may in some implementations place the teeth into final setups poses.
  • These transforms may in some implementations place the teeth into intermediate staging poses.
  • a transform may be described by a 9x1 transformation vector (e.g., that specifies a translation vector and a quaternion). In other implementations, a transform may be described by a transformation matrix (e.g., a 4x4 affine transformation matrix).
  • systems of this disclosure may implement a principal components analysis (PCA) on an oral care mesh and use the resulting principal components as at least a portion of the representation of the oral care mesh in subsequent machine learning and/or other predictive or generative processing.
  • PCA principal components analysis
  • An autoencoder may be trained to generate a latent form of a 3D oral care representation.
  • An autoencoder may contain a 3D encoder (which encodes a 3D oral care representation into a latent form), and/or a 3D decoder (which reconstructs that latent from into a facsimile of the inputted 3D oral care representation).
  • 3D encoders and 3D decoders the term 3D should be interpreted in a non-limiting fashion to encompass multi-dimensional modes of operation.
  • systems of this disclosure may train multi-dimensional encoders and/or multi-dimensional decoders.
  • Systems of this disclosure may implement end-to-end training.
  • Some of the end-to-end training-based techniques of this disclosure may involve two or more neural networks, where the two or more neural networks are trained together (i.e., the weights are updated concurrently during the processing of each batch of input oral care data).
  • End-to-end training may, in some implementations, be applied to setups prediction by concurrently training a neural network which learns a representation of the teeth, along with a neural network which generates the tooth transforms.
  • a neural network e.g., a U-Net
  • a first task e.g., such as coordinate system prediction
  • the neural network trained on the first task may be executed to provide one or more of the starting neural network weights for the training of another neural network that is trained to perform a second task (e.g., setups prediction).
  • the first network may learn the low-level neural network features of oral care meshes and be shown to work well at the first task.
  • the second network may exhibit faster training and/or improved performance by using the first network as a starting point in training.
  • Certain layers may be trained to encode neural network features for the oral care meshes that were in the training dataset. These layers may thereafter be fixed (or be subjected to minor changes over the course of training) and be combined with other neural network components, such as additional layers, which are trained for one or more oral care tasks (such as setups prediction).
  • a portion of a neural network for one or more of the techniques of the present disclosure may receive initial training on another task, which may yield important learning in the trained network layers. This encoded learning may then be built upon with further task-specific training of another network.
  • Systems of this disclosure may train ML models with representation learning.
  • representation learning include the fact that the generative network (e.g., neural network that predicts a transform for use in setups prediction) can be configured to receive input with a known size and/or standard format, as opposed to receiving input with a variable size or structure.
  • Representation learning may produce improved performance over other techniques, because noise in the input data may be reduced (e.g., because the representation generation model extracts the important aspects of an inputted representation (e.g., a mesh or point cloud) through loss calculations or network architectures chosen for that purpose). Such loss calculation methods may include KL-divergence loss, reconstruction loss or other losses disclosed herein.
  • Representation learning may reduce the size of the dataset required for training a model, because the representation model learns the representation, enabling the generative network to focus on learning the generative task. The result may be improved model generalization because meaningful features of the input data (e.g., local and/or global features) are made available to the generative network.
  • transfer learning may first train a representation generation model.
  • That representation generation model (in whole or in part) may then be used to pre- train a subsequent model, such as a generative model (e.g., that generates transform predictions).
  • a representation generation model may benefit from taking mesh element features as input, to improve the understanding of the structure and/or shape of the inputted 3D oral care representations in the training dataset.
  • transfer learning may be used for setups prediction, as well as for other oral care applications, such as mesh classification (e.g., tooth or setups classification), mesh element labeling, mesh element in-filling, procedure parameter imputation, mesh segmentation, coordinate system prediction, restoration design generation, mesh validation (for any of the applications disclosed herein).
  • a neural network trained to output predictions based on oral care meshes may first be partially trained on one of the following publicly available datasets, before being further trained on oral care data: Google PartNet dataset, ShapeNet dataset, ShapeNetCore dataset, Princeton Shape Benchmark dataset, ModelNet dataset, ObjectNet3D dataset, Thingi10K dataset (which is especially relevant to 3D printed parts validation), ABC: A Big CAD Model Dataset For Geometric Deep Learning, ScanObjectNN, VOCASET, 3D-FUTURE, MCB: Mechanical Components Benchmark, PoseNet dataset, PointCNN dataset, MeshNet dataset, MeshCNN dataset, PointNet++ dataset, PointNet dataset, or PointCNN dataset.
  • a neural network which was previously trained on a first dataset may subsequently receive further training on oral care data and be applied to oral care applications (such as setups prediction). Transfer learning may be employed to further train any of the following networks: GCN (Graph Convolutional Networks), PointNet, ResNet or any of the other neural networks from the published literature which are listed above.
  • GCN Graph Convolutional Networks
  • PointNet PointNet
  • ResNet any of the other neural networks from the published literature which are listed above.
  • a first neural network may be trained to predict coordinate systems for teeth (such as by using the techniques described in WO2022123402A1 or US Provisional Application No. US63/366492).
  • a second neural network may be trained for setups prediction, according to any of the setups prediction techniques of the present disclosure (or a combination of any two or more of the techniques described herein).
  • Transfer learning may transfer at least a portion of the knowledge or capability of the first neural network to the second neural network. As such, transfer learning may provide the second neural network an accelerated training phase to reach convergence.
  • the training of the second network may, after being augmented with the transferred learning, then be completed using one or more of the techniques of this disclosure.
  • Systems of this disclosure may train ML models with representation learning.
  • representation learning includes that the generative network (e.g., neural network that predicts a transform for use in setups prediction) can be configured to receive input with a known size and/or standard format, as opposed to receiving input with a variable size or structure.
  • Representation learning may produce improved performance over other techniques, because noise in the input data may be reduced (e.g., because the representation generation model extracts hierarchical neural network features and/or reconstruction characteristics of an inputted representation (e.g., a mesh or point cloud) through loss calculations or network architectures chosen for that purpose).
  • Reconstruction characteristics may comprise values in of a latent representation (e.g., a latent vector) that describe aspects of the shape and/or structure of the 3D representation that was provided to the representation generation module that generated the latent representation.
  • the weights of the encoder module of a reconstruction autoencoder may be trained to encode a 3D representation (e.g., a 3D mesh, or others described herein) into a latent vector representation (e.g., a latent vector).
  • the capability to encode a large set (e.g., hundreds, thousands or millions) of mesh elements into a latent vector may be learned by the weights of the encoder.
  • Each dimension of that latent vector may contain a real number which describes some aspect of the shape and/or structure of the original 3D representation.
  • the weights of the decoder module of the reconstruction autoencoder may be trained to reconstruct the latent vector into a close facsimile of the original 3D representation.
  • the capability to interpret the dimensions of the latent vector, and to decode the values within those dimensions may be learned by the decoder.
  • the encoder and decoder neural network modules are trained to perform the mapping of a 3D representation into a latent vector, which may then be mapped back (or otherwise reconstructed) into a 3D representation that is substantially similar to an original 3D representation for which the latent vector was generated.
  • examples of loss calculation may include KL-divergence loss, reconstruction loss or other losses disclosed herein.
  • Representation learning may reduce the size of the dataset required for training a model, because the representation model learns the representation, enabling the generative network to focus on learning the generative task. The result may be improved model generalization because meaningful neural network features of the input data (e.g., local and/or global features) are made available to the generative network.
  • a first network may learn the representation, and a second network may make the predictive decision.
  • each of the networks may generate more accurate results for their respective tasks than with a single network which is trained to both learn a representation and make a decision.
  • transfer learning may first train a representation generation model. That representation generation model (in whole or in part) may then be used to pre-train a subsequent model, such as a generative model (e.g., that generates transform predictions).
  • a representation generation model may benefit from taking mesh element features as input, to improve the capability of a second ML module to encode the structure and/or shape of the inputted 3D oral care representations in the training dataset.
  • One or more of the neural networks models of this disclosure may have attention gates integrated within. Attention gate integration provides the enhancement of enabling the associated neural network architecture to focus resources on one or more input values.
  • an attention gate may be integrated with a U-Net architecture, with the advantage of enabling the U-Net to focus on certain inputs, such as input flags which correspond to teeth which are meant to be fixed (e.g,. prevented from moving) during orthodontic treatment (or which require other special handling).
  • An attention gate may also be integrated with an encoder or with an autoencoder (such as VAE or capsule autoencoder) to improve predictive accuracy, in accordance with aspects of this disclosure.
  • attention gates can be used to configure a machine learning model to give higher weight to aspects of the data which are more likely to be relevant to correctly generated outputs.
  • attention gates or mechanisms
  • the ultimate predictive accuracy of those machine learning models is improved.
  • the quality and makeup of the training dataset for a neural network can impact the performance of the neural network in its execution phase.
  • Dataset filtering and outlier removal can be advantageously applied to the training of the neural networks for the various techniques of the present disclosure (e.g., for the prediction of final setups or intermediate staging, for mesh element labeling or a neural network for mesh in-filling, for tooth reconstruction, for 3D mesh classification, etc.), because dataset filtering and outlier removal may remove noise from the dataset.
  • dataset filtering and outlier removal may remove noise from the dataset.
  • this approach allows for the machine learning model to focus on relevant aspects of the dataset, and may lead to improvements in accuracy similar to improvements in accuracy realized vis-à-vis attention gates.
  • a patient case may contain at least one of a set of segmented tooth meshes for that patient, a mal transform for each tooth, and/or a ground truth setup transform for each tooth.
  • a patient case may contain at least one of a set of segmented tooth meshes for that patient, a mal transform for each tooth, and/or a set of ground truth intermediate stage transforms for each tooth.
  • a training dataset may exclude patient cases which contact passive stages (i.e., stages where the teeth of an arch do not move).
  • the dataset may exclude cases where passive stages exist at the end of treatment.
  • a dataset may exclude cases where overcrowding is present at the end of treatment (i.e., where the oral care provider, such as an orthodontist or dentist) has chosen a final setup where the tooth meshes overlap to some degree.
  • the dataset may exclude cases of a certain level (or levels) of difficulty (e.g., easy, medium and hard).
  • the dataset may include cases with zero pinned teeth (or may include cases where at least one tooth is pinned).
  • a pinned tooth may be designated by a technician as they design the treatment to stop the various tools from moving that particular tooth.
  • a dataset may exclude cases without any fixed teeth (conversely, where at least one tooth is fixed).
  • a fixed tooth may be defined as a tooth that shall not move in the course of treatment.
  • a dataset may exclude cases without any pontic teeth (conversely, cases in which at least one tooth is pontic).
  • a pontic tooth may be described as a “ghost” tooth that is represented in the digital model of the arch but is either not actually present in the patient’s dentition or where there may be a small or partial tooth that may benefit from future work (such as the addition of composite material through a dental restoration appliance).
  • the advantage of including a pontic tooth in a patient case is to leave space in the arch as a part of a plan for the movements of other teeth, in the course of orthodontic treatment.
  • a pontic tooth may save space in the patient’s dentition for future dental or orthodontic work, such as the installation of an implant or crown, or the application of a dental restoration appliance, such as to add composite material to an existing tooth that is too small or has an undesired shape.
  • the dataset may exclude cases where the patient does not meet an age requirement (e.g., younger than 12).
  • the dataset may exclude cases with interproximal reduction (IPR) beyond a certain threshold amount (e.g., more than 1.0 mm).
  • IPR interproximal reduction
  • the dataset to train a neural network to predict setups for clear tray aligners (CTA) may exclude patient cases which are not related to CTA treatment.
  • the dataset to train a neural network to predict setups for an indirect bonding tray product may exclude cases which are not related to indirect bonding tray treatment.
  • the dataset may exclude cases where only certain teeth are treated.
  • a dataset may comprise of only cases where at least one of the following are treated: anterior teeth, posterior teeth, bicuspids, molars, incisors, and/or cuspids.
  • the mesh comparison module may compare two or more meshes, for example for the computation of a loss function or for the computation of a reconstruction error. Some implementations may involve a comparison of the volume and/or area of the two meshes.
  • Some implementations may involve the computation of a minimum distance between corresponding vertices/faces/edges/voxels of two meshes. For a point in one mesh (vertex point, mid-point on edge, or triangle center, for example) compute the minimum distance between that point and the corresponding point in the other mesh. In the case that the other mesh has a different number of elements or there is otherwise no clear mapping between corresponding points for the two meshes, different approaches can be considered.
  • the open-source software packages CloudCompare and MeshLab each have mesh comparison tools which may play a role in the mesh comparison module for the present disclosure.
  • a Hausdorff Distance may be computed to quantify the difference in shape between two meshes.
  • the open-source software tool Metro developed by the Visual Computing Lab, can also play a role in quantifying the difference between two meshes.
  • the following paper describes the approach taken by Metro, which may be adapted by the neural networks applications of the present disclosure for use in mesh comparison and difference quantification: "Metro: measuring error on simplified surfaces” by P. Cignoni, C. Rocchini and R. Scopigno, Computer Graphics Forum, Blackwell Publishers, vol. 17(2), June 1998, pp 167-174.
  • Some techniques of this disclosure may incorporate the operation of, for one or more points on the first mesh, projecting a ray normal to the mesh surface and calculating the distance before that ray is incident upon the second mesh.
  • Restoration treatment of the patient may involve the specification of one or more of the following: restoration guidelines, restoration design parameters, and/or restoration rules for modifying one or more aspects of a patient’s dentition.
  • restoration guidelines restoration design parameters, and/or restoration rules for modifying one or more aspects of a patient’s dentition.
  • restoration rules for modifying one or more aspects of a patient’s dentition.
  • One or more of many possible factors may be considered in designing a 3D restoration, whether from an esthetic standpoint and/or from a technical standpoint.
  • the dental and facial midlines and angulation may provide overall guidance, as does the amount of tooth seen by others when the lips are at rest and/or when smiling.
  • a set of “golden proportions” may also inform the esthetic design of overall tooth sizes. Tooth-to-tooth proportion may be configured to reflect these “golden proportions,” which are: 1.618:1.0:0.618 for the central incisor, lateral incisor and the canine, respectively.
  • a real value may be specified for one or more of the RDP and be received at the input of a dental restoration design prediction model (e.g., a machine learning model to predict the final tooth shape at the completion of restoration design).
  • one or more RDP may be defined which correspond to one or more restoration design metrics (RDM).
  • RDM restoration design metrics
  • Tooth length, width, and width-to-length esthetic relationships may be specified for one or more teeth.
  • the length for a maxillary central incisor may be set to 11 mm, and the width- to-length esthetic relationship may be set to either 70% or 80%.
  • the lateral incisors may be between 1.0 mm and 2.5 mm shorter than the central incisors.
  • Canine teeth may, in some instances, be between 0.5 mm and 1.0 mm shorter than the central incisors. Other proportions and measurements are possible for various teeth.
  • restorations produced from a given material must be of sufficient thickness to have the necessary mechanical strength required for long term use.
  • the tooth width and shape must be designed to provide a suitable contact with the adjacent teeth.
  • the example Style options in the list below are from the LVI standards, from the Las Vegas Institute (LVI) of Advanced Dental Studies. Other style guides are commercially or freely available.
  • a neural network engine of this disclosure may incorporate as an input one or more of accepted “golden proportion” guidelines for the size of teeth, accepted “ideal” tooth shapes, patient preferences, practitioner preferences, etc.
  • Restoration design parameters may be used to encode aspects of smile design guidelines described herein, such as parameters which pertain to the intended dimensions of a restored tooth. Restoration design parameters are intended as instructions and/or specifications which describe the shape and/or form that one or more teeth should assume after the completion of dental restoration treatment.
  • One or more RDP may be received by a neural network or other machine learning or optimization algorithm for dental restoration design, with the advantage of providing guidance to that optimization algorithm.
  • a dental restoration design may be used to define the target shapes of one or more teeth for the generation of a dental restoration appliance.
  • a dental restoration design may be used to define the target teeth shapes for the generation of one or more veneers.
  • a partial list of tooth dimensions may include: length, width, height, circumference, diameter, diagonal measure, volume—any of which dimensions may be normalized in comparison to another tooth or teeth.
  • one or more restoration design parameters may be defined which pertain to a gap between two or more teeth, and the amount, if any, of the gap which the patient wishes to remain after treatment (e.g., such as when a patient wishes to retain a small gap between the upper central incisors).
  • Additional restoration design parameters may include the parameters specified in Table 2. In the event that one of these parameters contradicts another, the following order may determine precedence (i.e., let the first parameter in the following list be considered authoritative). If a parameter value is not specified, then that parameter may be ignored.
  • default values may be introduced for one or more parameters. Such default values may be determined, for example, through clustering of prior patient cases.
  • a golden proportion guideline may specify one or more numbers pertaining to the widths of adjacent teeth, such as: ⁇ 1.6, 1, 0.6 ⁇ .
  • Parameter Name Parameter Value of Unit of Measure Golden proportion guideline ⁇ guideline01, guideline02, guideline03, guideline04, etc. ⁇ Tooth width at base (mesial to distal distance) [millimeters] Tooth width at incisal edge (mesial to distal [millimeters] distance) Tooth height (gingival to incisal distance) [millimeters] Width-to-length esthetic relationship [percentage] Tooth-to-tooth proportion – upper right central [real number] incisor width to upper lateral incisor width Tooth-to-tooth proportion – upper right lateral [real number] incisor width to upper cuspid width Tooth-to-tooth proportion – lower right central [real number] incisor width to lower lateral incisor width Tooth-to-tooth proportion – lower right
  • Proportions may be made relative to tooth widths, heights, diagonals, etc. Angle lines, incisal angles and buccal contours may describe primary aspects of tooth macro shape. Mamelon grooves may be vertical macro textures on the front of a tooth, and sometimes may take a V-shape. Striations or perikymata may be a horizontal micro texture on the teeth. Symmetry may be generally desired. There may be differences between male and female patients.
  • Parameters may be defined to encode doctor restoration design preferences (DRDP), as pertains to various use case scenarios. These use case scenarios, may reflect information about the treatment preferences of one or more doctors, and directly affect the characteristics of one or more teeth in a dental restoration design or a veneer.
  • DRDP doctor restoration design preferences
  • DRDP may describe preferred or habitually involved values or ranges of values of RDP for a doctor or other treating medical professional. In some instances, such value or ranges of values may be derived from historical patient cases that were treated by that doctor or medical professional. In some instances, a DRDP may be defined which is derived from a RDP (e.g., such as Width-to-length esthetic relationship) or from a RDM.
  • Machine learning models such as those described herein (e.g., denoising diffusion models), may be trained to generated designs for tooth crowns or tooth roots (or both).
  • a dental restoration design may describe a tooth shape which is intended at the end of dental restoration.
  • a neural network (such as a generative neural network) may be trained to generate a dental restoration design which is to be used to produce either a veneer (e.g., a zirconia veneer) or a dental restoration appliance.
  • a veneer e.g., a zirconia veneer
  • a dental restoration appliance e.g., a dental restoration appliance.
  • Such a model make take as input data from cohort patient cases, including pre-restoration tooth meshes and corresponding ground truth examples of completed restorations (e.g., tooth meshes with restored shapes and/or structures).
  • Such a model may be trained, at least in part through the calculation of a loss function, which may quantify the difference between a generated crown restoration design and a ground truth crown restoration design.
  • the resulting loss may be used to update the weights of the generative neural network model (e.g., a denoising diffusion model – which may include a U-Net), thereby training the model (at least in part).
  • a reconstruction loss may be computed to compare a predicted tooth mesh to a ground truth tooth mesh, or to compare pre-restoration tooth mesh to a completed restoration design tooth mesh.
  • Reconstruction loss may be computed as the sum of pairwise distances between corresponding mesh elements and may be computed to quantify the difference between two tooth crown designs. Other losses disclosed herein may also be in the training.
  • a veneer may be created using the generated restoration design. Such as veneer may be 3D printed.
  • Machine learning models such as those described herein (e.g., denoising diffusion models), may be trained to generate components for use in creating a dental restoration appliance.
  • a dental restoration appliance may be used to shape dental composite in the patient’s mouth while that composite is cured (e.g., using a curing light), to ultimately produce veneers on one or more of the patient’s teeth.
  • the 3M FILTEK Matrix is an example of such a product.
  • a machine learning model for generating an appliance component may take inputs which are operable to customize the shape and/or structure of the appliance component, including inputs such as oral care parameters.
  • One or more oral care parameters may, in some instances, be defined based on oral care metrics.
  • An oral care metric may describe physical and/or spatial relationships between two or more teeth or may describe physical and/or dimensional characteristics of an individual tooth.
  • An oral care parameter may be defined which is intended to provide a machine learning model with guidance on generating a 3D oral care representation with a particular physical characteristic (e.g., pertaining to shape and/or structure). For example, physical characteristics may be measured with an oral care metric to which that oral care parameter corresponds.
  • Such oral care parameters may be defined to customize the generation of mold parting surfaces, gingival trim meshes or other generated appliance components, to tailor those appliance components to the patient’s dental anatomy.
  • the 3D representation generation techniques described herein may be trained to generate the custom appliance components by determining the characteristics of the custom appliance components, such as a size, shape, position, and/or orientation of the custom appliance components.
  • custom appliance components include a mold parting surface, a gingival trim surface, a shell, a facial ribbon, a lingual shelf (also referred to as a “stiffening rib”), a door, a window, an incisal ridge, a case frame sparing, or a diastema matrix wrapping, a spline, among others.
  • a spline refers to a curve that passes through a plurality of points or vertices, such as a piecewise polynomial parametric curve.
  • a mold parting surface refers to a 3D mesh that bisects two sides of one or more teeth (e.g., separates the facial side of one or more teeth from the lingual side of the one or more teeth).
  • a gingival trim surface refers to a 3D mesh that trims an encompassing shell along the gingival margin.
  • a shell refers to a body of nominal thickness. In some examples, an inner surface of the shell matches the surface of the dental arch and an outer surface of the shell is a nominal offset of the inner surface.
  • the facial ribbon refers to a stiffening rib of nominal thickness that is offset facially from the shell.
  • a window refers to an aperture that provides access to the tooth surface so that dental composite can be placed on the tooth.
  • a door refers to a structure that covers the window.
  • An incisal ridge provides reinforcement at the incisal edge of dental appliance and may be derived from the archform.
  • the case frame sparing refers to connective material that couples parts of a dental appliance (e.g., the lingual portion of a dental appliance, the facial portion of a dental appliance, and subcomponents thereof) to the manufacturing case frame.
  • a mold parting surface refers to a 3D mesh that bisects two sides of one or more teeth (e.g., by separating the facial side of one or more teeth from the lingual side of the one or more teeth).
  • a gingival trim surface refers to a 3D mesh that trims an encompassing shell along the gingival margin.
  • a shell refers to a body of nominal thickness.
  • an inner surface of the shell matches the surface of the dental arch and an outer surface of the shell is a nominal offset of the inner surface.
  • the facial ribbon refers to a stiffening rib of nominal thickness that is offset facially from the shell.
  • a window refers to an aperture that provides access to the tooth surface so that dental composite can be placed on the tooth.
  • a door refers to a structure that covers the window.
  • An incisal ridge provides reinforcement at the incisal edge of a dental restoration appliance and may be derived from the archform.
  • the case frame sparing refers to connective material that couples parts of a dental restoration appliance (e.g., the lingual portion of a dental restoration appliance, the facial portion of a dental restoration appliance, and subcomponents thereof) to the manufacturing case frame.
  • the case frame sparing may tie the parts of a dental restoration appliance to the case frame during manufacturing, protect the various parts from damage or loss, and/or reduce the risk of mixing up parts.
  • Additional 3D oral care representations which may be generated by a denoising diffusion model (such as denoising diffusion models trained as described herein) including tooth restoration designs (e.g., which may include tooth interproximal surfaces, fossae, incisal edges, cusp tips, tooth roots or the like).
  • the denoising diffusion models (DDM) described herein may be trained to perform 3D mesh element labeling (e.g., labeling vertices, edges, faces, voxels or points) in 3D oral care representations.
  • Those labeled mesh elements may be used for mesh cleanup or mesh segmentation.
  • the labeled aspects of a scanned tooth mesh may be used for appliance erasure (removal + replacement) or be used to modify (e.g., by smoothing) one or more aspects of the tooth to remove aspects of attached hardware (or other aspects of the mesh which may be unwanted for certain processing and appliance creation - such as extraneous material).
  • Mesh element features may be computed for one or more mesh elements in a 3D representation of oral care data (e.g., 3D representation of the patient’s dentition).
  • a vector of such mesh element features may be computed for each mesh element and then be received by a DDM which has been trained to label mesh elements in a 3D oral care representation for the purpose of either mesh segmentation or mesh cleanup.
  • Such mesh element features may confer valuable information about the shape and/or structure of the input mesh to the labeling DDM.
  • training data 200 may contain 3D representations of the patient’s pre-restoration dentition (e.g., from an intraoral scanner, CT scanner, or the like), and ground truth data which includes ground truth mesh element labels.
  • the ground truth mesh element labels may describe ground truth (or reference) segmentations of the patient’s dentition (e.g., each mesh element which occupies a lower left central incisor may have the same label, each mesh element in the gums of the upper arch may have the same label, each mesh element which appears in an upper right cuspid may have the same label, etc.).
  • Either or both of the patient’s dentition and the ground truth mesh element labels may undergo latent encoding (214).
  • Either the original training data 200 or the corresponding latent-encoded training data may be provided to the Markov chain 204, as a part of the forward pass 206.
  • the forward pass may iteratively add noise to the mesh element labels, generating a series of tens, hundreds, or thousands of successively noisier versions of the set of ground truth mesh element labels.
  • the successive addition of noise may comprise randomizing the value of one or more mesh element labels.
  • the successive addition of noise may comprise adjusting the labels of the neighbors of a randomly chosen mesh element (e.g., such as a mesh element which lies along the boundary of a tooth). In this manner, the boundaries between teeth, or between a tooth and the gums man be adjusted to become increasingly noisy or distorted.
  • the Markov chain may generate a succession of increasingly noisy sets of mesh element labels which may be used to train, at least in part, the denoising ML model 210.
  • the 3D representation(s) of the patient’s detention may also be provided, along with the mesh element labels, to the training of the denoising ML model 210.
  • the reverse pass 208 may be executed.
  • Oral care arguments 202 may be provided, to customize the functioning of the segmentation operation.
  • instant patient case data 216 e.g., a pre-segmentation 3D representation of the patient’s dentition
  • the instant patient case data 216 may undergo latent encoding (218).
  • an initial set of mesh element labels may be generated which correspond to the 3D representation of the patient’s detention 216.
  • the mesh element labels may, in some implementations, be initialized at random, or according to heuristics.
  • An example of a heuristic is that a set of random mesh elements are chosen, and the neighbors of those mesh elements (e.g., neighbors with a particular distance or number of edges of the chosen mesh element) may be grouped together in cliques (e.g., which each mesh element in a clique has the same mesh element label).
  • the reverse pass 208 of the denoising ML model 210 may cause the denoising ML model 210 to iteratively refine the set of mesh element labels, relative to the 3D representation of the patient’s dentition 216.
  • the iteratively refined mesh element labels may be set to the output 220.
  • These generated mesh element labels may be used to segment the patient’s detention or to perform mesh cleanup of the patient’s dentition.
  • the patient’s dentition may become segmented through the application of the generated mesh element labels 220.
  • the generated mesh element labels may enable tooth cutting (e.g., the generation of a new mesh for each distinct tooth), according to the generated mesh element labels 220.
  • Some implementations of the DDM-based mesh cleanup techniques described herein may train the DDMs to remove (or modify) generic triangle mesh defects (e.g., via mesh element labeling), such as: degenerate triangle with zero surface area; redundant triangle that covers the same surface area as another triangle; non-manifold edge with more than two adjacent triangles, also referred to as a “fin”; non-manifold vertex with more than one adjacent sequence of connected triangles (triangle fans); intersecting triangles – where two triangles penetrate each other; spikes - sharp features composed of multiple triangles, often conical, caused by one or more vertices being displaced from the actual surface; folds - sharp features composed of multiple triangles, often Z-shaped with a small undercut area, caused by one or more vertices being displaced from the actual surface; islands/small components - disconnected objects in a scan which should only contain a single object (e.g., typically the smaller objects are deleted); small holes in the mesh surface, either from the original scan or from
  • Some implementations of the DDM-based mesh cleanup techniques described herein may train the DDM models to remove (or modify) aspects of meshes (e.g., via mesh element labeling) which are unwanted under certain circumstances and/or domain-specific defects, such as: extraneous material – portions of the intraoral scan outside the anatomical area of interest, e.g., non-tooth surfaces that are not within some distance of tooth surfaces, or scan artifacts that do not represent actual anatomy; divots - concave depressions in surfaces (e.g., may be scan artifacts, which should be fixed, or anatomical features, which are generally left intact); undercuts - sides of a tooth of lower radius than the crown, such that physical impressions or aligners may become difficult to remove or emplace.
  • extraneous material – portions of the intraoral scan outside the anatomical area of interest e.g., non-tooth surfaces that are not within some distance of tooth surfaces, or scan artifacts that do not represent actual anatomy
  • Undercuts may be a natural feature or due to damage such as an abfraction. Abfractions are associated with the erosion of a tooth near the gumline, causing or exacerbating an undercut.
  • Appliances that the DDM-based models of this disclosure may process include orthodontic hardware such as attachments, brackets, wires, buttons, lingual bars, Carriere appliances, or the like, and may be present in intraoral scans. Digital removal and replacement with synthetic tooth/gingiva surfaces may, in some circumstances, be beneficial if performed before appliance creation steps proceed.
  • Denoising diffusion may train a neural network (or other machine learning model) to iteratively remove noise from (or refine) a data structure.
  • Techniques of this disclosure remove noise from a 3D point cloud (or other 3D representation), from a transformation matrix, a vector of labels (e.g., labels which are applied to mesh elements and are used for segmentation or mesh cleanup, and which may be used to define object masks) or from other kinds of 3D oral care representations.
  • the techniques may start out with randomly generated data structures (e.g., 3D point clouds or transformation matrices), and the techniques may transform those random (or noisy) data structures into artifacts which are usable in digital oral care.
  • Denoising models may be trained to perform these operations because of the kind of training data that is generated using the techniques described herein. This training data may comprise a series of increasingly noisy examples of the relevant data structure.
  • the denoising neural network (of which a U-Net is an example) may be trained to gradually, in small steps, transform a random (or noisy) data structure into an artifact which is useable in digital oral care treatment (e.g., the generation of oral care appliances).
  • Digital oral care involves many different kinds of 3D representations which are customized to the anatomy of the patient. Diffusion models are particularly applicable to digital oral care because of the customizable nature of the generated output representations.
  • a denoising diffusion model may operate on input data (e.g., 3D oral care representations) in the data’s original format (e.g., training on and/or generating 3D representations – such as 3D point clouds, and the like).
  • a denoising diffusion model may operate on a latent form of the input data (e.g., such as in Stable Diffusion), which may preserve the 3D nature of the input data.
  • a latent form may be information-rich (describing the structure and/or shape of the input data - such as a 3D oral care representation) and/or have a small data footprint that is easier for an ML model to be trained on.
  • an autoencoder may be trained to generate a latent representation. The information-rich nature of the latent representation may be demonstrated when the latent representation is reconstructed into a close facsimile of the original 3D representation (e.g., as measured by reconstruction error defined herein).
  • the original 3D representation may comprise hundreds or thousands of mesh elements, which may be encoded by the autoencoder into a a vector of a few hundred real values, in some instances.
  • This vector of a few hundred real values may, in some instances, subsequently be reconstructed into a close facsimile of the original 3D representation.
  • the accuracy of the reconstruction e.g., the fidelity with which the reconstructed form matches the original form
  • An oral care diffusion model may, in some implementations, be trained to modify the 3D oral care representations described herein (e.g., an appliance component, a tooth restoration design, a trimline, a transform for a 3D oral care representations – such as a tooth in an orthodontic setup or an appliance component, etc.).
  • a template or reference 3D oral care representation may be received at the input and subsequently modified by the OCDM.
  • an initial tooth representation (e.g., pre-restoration tooth) may be received at the input and subsequently be modified by the OCDM to generate a tooth which is ready for restoration treatment (e.g., for using making a crown, bridge, or a dental restoration appliance such as the FILTEK Matrix).
  • An OCDM may, in some implementations, be trained to generate the 3D oral care representations described herein (e.g., an appliance component, a tooth restoration design, a trimline, an orthodontic setup, etc.).
  • a diffusion model may be trained to generate a 3D oral care representation (e.g., a point cloud or 3D mesh representing a 3D oral care representation such as a trimline, an appliance component or a tooth design for dental restoration – such as the generation of a veneer, the generation of a dental restoration appliance).
  • a 3D oral care representation e.g., a point cloud or 3D mesh representing a 3D oral care representation such as a trimline, an appliance component or a tooth design for dental restoration – such as the generation of a veneer, the generation of a dental restoration appliance.
  • GGDM Geometry Generation Diffusion neural network
  • GGDM Geometry Generation Diffusion Model
  • the input data for a diffusion model may be pre-processed using an encoder (e.g., to generate a reduced dimensionality representation of the input data structure).
  • Such a latent representation may, in some instances, be reconstructed using a decoder.
  • the decoder may reconstruct the latent representation into a facsimile of the input data (e.g., such as with a reconstruction autoencoder).
  • the decoder may reconstruct the latent representation into a modified version of the Input data (e.g., when one or more aspects of the latent representation are modified).
  • Point-Voxel CNNs (PVCNN) or encoder-decoder networks e.g., such as Variational Autoencoders, among others
  • a 3D oral care representation 102 may be received at the input to the Oral Care Diffusion Model (OCDM) and be encoded to a latent representation (e.g., a latent vector) by an encoder.
  • a latent representation e.g., a latent vector
  • an optional encoder 104 e.g., as a structure latent encoder – which may encoder aspects of the shape and/or structure of the input data
  • may encode an input 3D oral care representation 102 into a latent representation 108 of that shape e.g., termed a structure latent representation).
  • an optional latent mesh element encoder 106 may encode an input 3D oral care representation 102 into a latent representation 110 of mesh elements (e.g., termed a latent mesh element representation).
  • mesh elements may comprise points in a point cloud, voxels in a sparse representation, or vertices/edges/faces in a 3D mesh, among others.
  • each of these two representations may be generated.
  • Such encoders may, in some implementations, incorporate mesh element features at the input, to improve the latent representations generated by the encoders.
  • Mesh element feature vectors may be computed by mesh element feature modules 130 and 128.
  • a Markov chain of diffusion steps may be applied to generate a set of training data from such a latent representation.
  • the forward process may generate increasingly noisy versions of the input 3D oral care representation 102 which may be used in training the one or more denoising diffusion neural network models described herein.
  • a diffusion neural network (e.g., which may be implemented with a U-Net, variational autoencoder, etc.) may be trained to reverse the forward process, starting with the 3D oral care representation at time T, and performing a denoising operation to convert the 3D oral care representation to a version which corresponds to time T-1. This is called the reverse process.
  • the diffusion neural network Given the latent representation at time t, the diffusion neural network may compute the version of the latent representation at time t-1, in an iterative fashion.
  • the diffusion neural network may introduce changes to the latent representation (e.g., to remove noise or to introduce geometrical or structural aspects which are desired in the output).
  • the diffusion model may consume instructions on such changes in the form of oral care arguments.
  • Oral care arguments may comprise of one or more natural language text string arguments, one or more categorical arguments, one or more integer arguments, one or more real-valued arguments, one or more image arguments (e.g., a reference image describing aspects of an intended outcome of model execution), one or more 3D representation arguments (e.g., a reference 3D point cloud describing aspects of an intended out of model execution) or a combination of multiple of those previously mentioned.
  • Such arguments may first undergo latent encoding (e.g., using the encoder portion of an autoencoder or using a text or numerical encoder).
  • the encoded arguments may be introduced to the diffusion neural network (e.g., by appending the arguments to the input data, such as concatenating the arguments with a latent representation of the input data that was generated using an encoder).
  • the encoded arguments may be introduced to one or more of the successive resolution layers of the U-Net (e.g., the U-Net may have sets of layers which correspond to increasingly lower levels of resolution, followed by sets of layers which correspond to increasingly higher levels of resolution).
  • the U-Net may employ skip-connections to share information from input to output at each level of resolution.
  • the U-Net may incorporate attention layers.
  • the encoded argument information may be introduced to any or all of such layers, with the advantage of guiding the generated output of the diffusion neural network.
  • the denoising diffusion neural networks of this disclosure may include U-Nets, ResNets, transformers or autoencoders (e.g., variational autoencoders), among others.
  • the latent representation of the shape of a 3D point cloud 102 (or other 3D representation described herein) is called a structure latent representation 108, which may encode aspects of the structure of the input 3D point cloud 102.
  • Oral care arguments 100 (or input parameters) may comprise one or more attributes describing an intended output from a trained machine learning model.
  • a structure latent representation 108 may undergo denoising and/or modification by a structure latent denoising diffusion model 112.
  • the latent representation of mesh elements e.g., such as points
  • a latent mesh elements representation 110 e.g., which may encode aspects of the shape of the input 3D point cloud 102
  • a latent mesh elements representation may undergo denoising and/or modification by a latent mesh element denoising diffusion model 114.
  • the denoised/modified structure latent representations 116 may be provided to a final decoder module 120 (e.g., implemented as a PVCNN).
  • the latent mesh elements representation 118 may also be provided to the final decoder module 120.
  • the final decoder module may reconstruct a generated geometry 122 (e.g., or other 3D oral care representation) which is either a modified version of the input 3D oral care representation 102 or is an entirely new 3D oral care representation with aspects as specified by the oral care arguments 100.
  • the generated 3D oral care representation 122 may be embodied by 3D point cloud points.
  • Such a point cloud may undergo an optional subsequent processing step 124 to produce a mesh, for example, using Shape As Points (SAP) surface reconstruction (e.g., as shown in FIG. 1). If SAP is applied, then a 3D mesh 126 is outputted.
  • SAP Shape As Points
  • FIG.1 shows a method of using a fully trained GGDM.
  • the GGDM has been trained to modify an input 3D oral care representation according to (optional) arguments (e.g., as text instructions or according to oral care parameters described herein).
  • a diffusion model or a Text-to-Image (TTI) system e.g., such as DALL-E 2, MidJourney or Stable Diffusion
  • 2D representations such as 2D images
  • denoising ML model 210 may be first trained on 2D data, and subsequently be further trained (via transfer learning) on 3D data. Such a transfer-learned model may, in some implementations, be trained to generate 3D oral care representations. In some instances, a stable diffusion model may operate on latent representations of trial 3D oral care representations 216, for the purpose of modifying those 3D oral care representations 216. [00185] In some non-limiting implementations, the denoising diffusion probabilistic models of this disclosure may operate on 3D representations of teeth (e.g., to generate tooth restoration designs).
  • a 3D representation of the pre-restoration tooth 102 may be provided as the input, and processed by denoising diffusion probabilistic neural networks in a manner that preserves information about the 3D shape and/or 3D structure of that pre-restoration tooth.
  • the pre-restoration tooth may undergo optional mesh element feature vector generation (128), and optionally undergo encoding into latent form (106) which preserves information about the 3D qualities of the pre-restoration tooth.
  • the resulting latent mesh elements representation 110 may represent the 3D mesh elements (e.g., 3D points, voxel, edges, faces, or vertices) of the pre-restoration tooth in latent form.
  • the latent mesh elements representations 110 may be provided to the latent mesh element denoising diffusion model 114, which may generate a modified latent mesh elements representation 118.
  • the pre-restoration tooth may undergo optional mesh element vector generation (130), and optionally undergo encoding into a latent form (104) which describes aspects of the pre-restoration tooth's 3D structure in latent form 108.
  • the latent representation 108 may be provided to structure latent denoising diffusion model 112, which may generate a modified structure latent representation 116.
  • Either or both of the modified structure latent representation 116 and the modified latent mesh elements representation 118 may be provided to the decoder 120, which may reconstruct the representations into a reconstructed 3D oral care representation 122 (e.g., a post-restoration tooth design).
  • a reconstructed 3D oral care representation 122 e.g., a post-restoration tooth design
  • that point cloud may subsequently undergo surface reconstruction (124), to generate a 3D mesh 126.
  • Oral care arguments 100 may optionally be provided to either or both of the structure latent denoising diffusion model 112, and the latent mesh elements denoising diffusion model 114, to customize the outputs of those models, and configure those models to generate output which is suitable for use in generation an oral care appliance.
  • Oral care arguments may contain oral care parameters or oral care metrics (e.g., restoration design metrics – such as “Tooth Morphology,” or “Bilateral Symmetry and/or Ratios,” among others described herein).
  • Such latent forms 110 and 108 may reduce the data size of the pre-restoration tooth, leading to improvements in data precision, while still preserving rich information about the shape and/or structure of the pre-restoration tooth.
  • Neural networks may be more easily trained (e.g., require fewer neural network parameters or weights to encode solutions) on input data with a smaller data footprint, as long as those data are information-rich.
  • Reconstruction autoencoders are particularly configured to generate information-rich latent representations, as shown by the reconstruction autoencoder’s ability to reconstruct a close facsimile of the input data.
  • a ground truth post-restoration 3D representation of a tooth may likewise be processed entirely in three dimensions.
  • a Geometry Generation Diffusion Model may be trained, at least in part, by computing a loss which quantifies the difference between a generated post-restoration tooth design (e.g., a design predicted using a denoising diffusion model as described herein) and a ground truth post- restoration tooth (e.g., which may be provided in the training data for a given patient case).
  • a denoising diffusion model (e.g., such as that shown in FIG.2) may involve one or more Markov chains.
  • a Markov chain may, in some instances, be defined as a sequence of stochastic events where each time point depends on the previous time point.
  • the transition distributions on the Markov chain forward process may, in some instances, be conditioned by a low level of Gaussian noise.
  • the forward pass 206 of FIG. 2 generates training data for a denoising diffusion model (e.g., which may, in some implementations, be trained on latent representations of mesh elements).
  • the training data may be used to train the denoising ML model 210.
  • the fully trained denoising ML model 210 shown in FIG.2 may be used by GGDM (e.g., where the model is trained on latent representations of teeth – such as with restoration design generation), by a Diffusion Setups model (e.g., where the model is trained on latent representations of transforms for teeth or other types of 3D oral care representations), or by other types of diffusion models for generating outputs based on training datasets of 3D oral care representations.
  • a DDM may be trained to approximate a data distribution q(x) over a latent variable Y (e.g., where the data distribution may describe 3D oral care representations).
  • a denoising diffusion model a noise scheduler which is used to implement the forward process (e.g., which may introduce progressively more noise of a particular distribution, such as Gaussian noise, to each successive time point – such as isotropic Gaussian noise with mean of zero and variance in all directions), a neural network (e.g., U-Net, VAE or others described herein) which is used to implement the reverse process (e.g., which may, over many time points, recover the original distribution of the data that supplied to the start of the forward process), and/or an optional method of time point encoding (e.g., using positional embeddings).
  • a noise scheduler which is used to implement the forward process (e.g., which may introduce progressively more noise of a particular distribution, such as Gaussian noise, to each successive time point – such as isotropic Gaussian noise with mean of zero and variance in
  • the neural network may be trained, at least in part, using a loss function which is based on an L1 or L2 norm, or a KL-Divergence (among others described herein). KL-Divergence loss may, in some instances, drive minimization of the difference between an observed distribution and a corresponding generated distribution.
  • the forward process may, in some instances, comprise a further encoding of the received data (e.g., latent data), and the backward (or reverse) process may, in some instances, comprise a decoding (or partial decoding) of the data which were processed by the forward process. In some instances, the backward process may be performed starting with random noise, and a sequence of progressively less noisy data structures may be generated.
  • the distribution of latent variables Y may, in some instances, be normally distributed, after the addition of noise (e.g., Gaussian noise) using a forward transition module.
  • the forward process may, in some instances, introduce noise to a 3D oral care representation by, in the example of a 3D representation of oral care data, perturbing one or more mesh elements of the 3D oral care representation. For example, the position of a point cloud point or of a voxel may be perturbed.
  • a mesh element may undergo perturbation which may alter the value of a measured mesh element feature (e.g., length of an edge, area of a face, position of a point or voxel, count of incident edges for a vertex, etc.).
  • y t-1 ) N(y t ; sqrt(1-M t )y t-1 , M t I) [00191]
  • a U-Net may have a descending phase (in which the resolution of the input data is successively decreased using convolutional and down-sampling pooling layers, up to a level of most-coarse resolution, which may reveal increasingly global features of the input data) and an ascending phase (which receives the result of the descending phase and restores the resolution over a succession of complementary stages of unconvolution and unpooling).
  • Skip (or residual) connections may connect corresponding stages of the descending and ascending phases.
  • Attention modules may be integrated into the U-Net.
  • Batch normalization layers may be integrated into the U-Net.
  • FIG. 2 shows an example DDM, for setups prediction.
  • a denoising diffusion model for 3D oral care representations e.g., for denoising latent representations of 3D oral care representations, such as structure latents or latent mesh elements.
  • structure latents or latent mesh elements may be described by one or more latent vectors.
  • the denoising diffusion neural network may generate a target 3D oral care representation (e.g., an appliance component, a tooth restoration design, a fixture model component – such as a trimline, etc.) by iteratively denoising a set of mesh elements (e.g., points of a 3D point cloud) until a 3D oral care representation is generated which meets the specification of the oral care arguments which are supplied to the model.
  • Oral care arguments may include oral care parameters as disclosed herein, or other real-valued, text-based or categorical inputs which specify intended aspects of the one or more target 3D oral care representations which are to be generated.
  • oral care arguments may include oral care metrics, which may describe intended aspects of the one or more 3D oral care representations which are to be generated.
  • a text encoder may encode a set of natural language instructions from the clinician (e.g., generate a text embedding).
  • a text string may comprise tokens.
  • An encoder for generating text embeddings may, in some implementations, apply either mean-pooling or max- pooling between the token vectors.
  • a transformer e.g., BERT or Siamese BERT
  • may be trained to extract embeddings of text for use in digital oral care e.g., by training the transformer on examples of clinical text, such as those given below).
  • such a model for generating text embeddings may be trained using transfer learning (e.g., initially trained on another corpus of text, and then receive further training on text related to digital oral care). Some text embeddings may encode text at the word level. Some text embeddings may encode text at the token level.
  • a transformer for generating a text embedding may, in some implementations, be trained, at least in part, with a loss calculation which compares predicted outputs to ground truth outputs (e.g., softmax loss, multiple negatives ranking loss, MSE margin loss, cross-entropy loss or the like).
  • the non-text oral care arguments such as real values or categorical values, may be converted to text, and subsequently embedded using the techniques described herein.
  • oral care arguments 202 may contain natural language text instructions, such as the following, and may (optionally) be encoded into latent representations by latent encoding module 212 (e.g., which may comprise a text transformer, as described herein, among other architectures described herein).
  • latent encoding module 212 e.g., which may comprise a text transformer, as described herein, among other architectures described herein.
  • the following are examples of natural language instructions that may be provided by a clinician to the generative models described herein, to describe the intended outcome of 3D oral care representation generation using a the denoising diffusion probabilistic models of this disclosure: 1. “Generate a setup to set to Class I molar and canine, 2 mm overbite and add 2mm of expansion 5-5/5-5.” 2.
  • Denoising diffusion probabilistic models are a class of deep generative neural networks which may be trained to generate transformations (e.g., may be used to modify the positions or orientations of 3D representations in 3D space – such as teeth), generate 2D images (e.g., heatmaps or color images), generate 3D representations (e.g., such as point clouds, 3D meshes or voxelized representations), or other 3D oral care representations described herein.
  • Denoising diffusion models may be trained to generate transforms which place the teeth of an arch into poses which are suitable for orthodontic treatment (e.g., intermediate stages or final setups).
  • Denoising diffusion models may be implemented using one or more encoders, one or more MLPs, one or more autoencoders, one or more U- Nets, one or more transformers (e.g., 3D SWIN transformer encoders or 3D SWIN transformer decoders), one or more pyramid encoder-decoders, among other machine learning models.
  • a denoising diffusion model may take as input one or more oral care arguments, and also take as input one or more 3D representations of oral care data, such as teeth (e.g., a full arch of segmented teeth in maloccluded poses), appliance components, or fixture model components.
  • teeth e.g., a full arch of segmented teeth in maloccluded poses
  • appliance components e.g., appliance components
  • fixture model components e.g., fixture model components.
  • the inputted teeth may be in maloccluded poses.
  • the maloccluded transforms for the teeth may be provided to the denoising diffusion model.
  • Non-limitng oral care arguments may include a doctor's treatment plan (comprising at least: a set of zero or more procedure parameters, zero or more doctor preferences and zero or more text samples which describe the nature of the intended oral care treatment – such as a final setup or a restoration design generation).
  • Oral care arguments may include real values, categorical values, natural language instructions, among others.
  • a denoising diffusion model may be trained to generate a setup, such as a final setup, which fulfills the specification described by the procedure parameters and/or text.
  • a denoising diffusion model for generating an orthodontic setup may be termed a Diffusion Setups neural network or Diffusion Setups model.
  • a diffusion model which conditions on text may use a neural network to reformat and/or reduce the dimensionality of that text, for example, to generate a latent encoding or latent embedding of the text (e.g., using a transformer or an encoder to generate a text embedding).
  • a denoising diffusion model may comprise at least one of a forward pass and a reverse pass.
  • the forward pass of the diffusion model may generate training data by iteratively adding noise (e.g., Gaussian noise) to a received 3D oral care representation (e.g., a tooth transform or a point cloud representation of a tooth restoration design).
  • the reverse pass of the diffusion model may further operate through an iterative denoising process (e.g., such as using a U-Net trained for the purpose, or other models described herein), which iteratively removes noise from a received 3D oral care representation (e.g., a tooth transform or a point cloud representation a tooth restoration design).
  • a received 3D oral care representation e.g., a tooth transform or a point cloud representation a tooth restoration design.
  • Such tooth transforms may define the poses of the teeth in one or more arches.
  • a Diffusion Setups model may generate a 3D oral care representation, such as an orthodontic setup (e.g., for final setups or intermediate staging).
  • a DDM may build a training dataset by incrementally adding noise to a latent vector TA that was generated by latent encoding module 214.
  • a DDM e.g., diffusion setups or others described herein
  • the denoising may be customized, at least in part, by oral care arguments 202, which may undergo optional encoding (212) to generate latent vector TC
  • the latent vectors TA and TB contain reduced- dimensionality information about the one or more 3D oral care representations (e.g., tooth mesh information and/or tooth transform information).
  • TA or TB may, in some implementations, be conditioned on (or combined with) latent vectors A which correspond to the one or more 3D representations of teeth.
  • a diffusion model which receives one or more 3D representations of teeth may be trained for generating tooth restoration designs (e.g., such as with GGDM – which may also be trained to generate other types of 3D oral care representations, such as appliance components, fixture model components, transforms, or trimlines).
  • These latent vectors TA or TB may also be conditioned on (or combined with) one or more procedure parameters K and/or one or more doctor preferences L.
  • Such conditioning may be implemented, for example, by concatenating a latent vector TA or TB with K, L, or any of the other oral care arguments described in this disclosure, such as M, N, O, R, S, P, Q, U, V.
  • a latent vector TA (e.g., which may have been generated using an encoder) may undergo many iterations of noise through the course of the forward pass.
  • the series of increasingly noisy versions of TA may be used to train, at least in part, a denoising diffusion neural network (e.g., such as an autoencoder, a U-Net, or others described herein).
  • the latent vector TB may start out as Gaussian noise at the beginning of the reverse pass.
  • TA may evolve into a form that may be reconstructed (e.g., using a decoder) into a 3D oral care representation which may be used in oral care appliance generation (e.g., one or more transforms which may place one or more teeth into a setup configuration – such as a final setup or an intermediate stage).
  • a denoising ML model 210 may use a U-Net architecture with ResNet blocks and self-attention layers as a part of the reverse pass.
  • a ResNet block refers to a residual block, where the activation of one layer in the neural network is forwarded directly to a subsequent deeper layer, with the advantage of enabling deeper networks to be trained.
  • the self-attention layers make use of an attention mechanism which relates different portions of a sequence (i.e., the sequence of intermediate stages, among others) in order to compute a representation of that same sequence.
  • Self- attention is advantageous to intermediate staging prediction, in that each stage of the sequence may be updated or refined with information about other stages in the sequence, as the diffusion model iterates.
  • the denoising ML model 210 (e.g., for setups prediction, or 3D representation generation) may be trained through gradient descent and/or backpropagation.
  • the losses described elsewhere in this disclosure may be used to train, at least in part, the Setups Diffusion model.
  • L1, L2, MSE or other losses described herein may be used to train, at least in part, the denoising ML model 210.
  • FIG.3 shows how the denoising ML model 302 may be trained on a sequence of increasingly noisy copies of patient case data (e.g., orthodontic setups). The noise may, in some instances, be Gaussian.
  • FIG.3 shows a Markov chain 300 comprising 34 time points (though other Markov chain sizes are possible).
  • Orthodontic setups 304 may be provided to Markov chain 300.
  • the orthodontic setups 304 may be encoded (310) into latent form before being provided to the Markov chain 300.
  • the Markov chain 300 has length 34, other lengths are possible.
  • noise may be added to the patient data starting at T0 and proceed for multiple time steps, producing a set of training data with a gradually increasing amount of noise (e.g., setups transforms with more and more noise).
  • Each time point from T0 to T33 may contain transforms (e.g., one transform for each tooth in the arch) and/or 3D representations of teeth.
  • Data e.g., tooth transforms
  • the denoising ML model 302 e.g., an encoder-decoder network, such as a pyramid encoder-decoder, a 3D SWIN transformer, an autoencoder, or a U-Net
  • optional oral care arguments 306 e.g., oral care parameters or oral care metrics
  • the optional oral care arguments may, in some implementations, be encoded (308) into latent form. Training may proceed over many iterations.
  • a transform may comprise a 4x4 affine transform, although other dimensions are also compatible with the diffusion model-based techniques of this disclosure.
  • a transform may include one or more translation vectors, one or more Euler angles and/or one or more quaternions.
  • the “t” input is a time representation, which may control which time point is sampled during a given timestep of diffusion model training (e.g., the training of the denoising neural network).
  • the input “t” may be used for positional encoding.
  • a time point t may correspond to a stage in orthodontic treatment.
  • the use of positional encoding similar to that used in transformers, may enable the model to be aware of the relative positions of the stages in the treatment plan sequence.
  • FIG.4 illustrates an example of a mesh element labelling model which is based on denoising diffusion probabilistic model (DDPM) and which may be used to enable either of the 3D mesh segmentation (e.g., semantic segmentation) and 3D mesh cleanup pipelines.
  • DDPM denoising diffusion probabilistic model
  • a DDPM for 3D representation segmentation may comprise at least one of a forward pass 206 (e.g., which may involve many steps of adding iteratively more noise to an input 3D representation – which may be used in training an ML model to carry out the reverse process) and a reverse pass 208 (e.g., which may use an ML model, such as a neural network, to iteratively denoise a noisy representation, resulting in a segmentation of the input 3D representation).
  • a DDPM for 3D representation segmentation may, in some implementations, approximate aspects of a Markov process.
  • the HNNFEM 406 described in FIG. 4 may be trained to perform the reverse pass 208.
  • Dental arch data 400 may (optionally) be converted (402) to lists of mesh elements.
  • the mesh elements may undergo optional mesh element feature vector calculation (404).
  • Mesh element features described herein may be provided to the HNNFEM 406, such as mesh element features include: edge mid-point, edge curvature, signed dihedral angles, edge length, edge normal, or others described herein.
  • a DDPM may begin operations on a noisy representation (e.g., a noisy representation of mesh element labels) and may iteratively denoise that representation to produce a denoised representation which may comprise one or more mesh element labels or object masks (alternatively the noisy representation may be directly denoised into one or more segmented 3D representations, such as 3D point clouds or 3D meshes). Formulas which formalize the forward and reverse passes are described herein.
  • a DDPM for 3D representation segmentation may comprise the following steps: Iteratively add noise (e.g., Gaussian noise) to an input 3D oral care representation, to produce a succession of increasingly noisy versions of that input 3D oral care representation which may then be used in training a denoising neural network of the reverse process.
  • noise e.g., Gaussian noise
  • the mesh element representation vectors 410 from step #3 may be used to train one or more ML models for mesh element labeling (e.g., to train an ensemble of neural networks 412 for mesh element labeling).
  • the one or more ML models may vote (414), to determine a final output of mesh element labels 416.
  • the one or more ML models may generate labels for one or more mesh element features.
  • the set of mesh element labels which is generated may, in some implementations, comprise one or more object masks.
  • An object mask may label or otherwise indicate which mesh elements belong to which object in the scene (e.g., which mesh elements belong to an upper left central incisor, to the gums or to a hardware element in the arch mesh from intraoral scanner).
  • FIG.4 further shows the input of the 3D oral care representation 400 that is to be segmented (e.g., a dental arch), which may (optionally) be arranged (402) into vectors of mesh elements.
  • One or more mesh element features may be computed (404) for each mesh element.
  • HNNFEM 406 may contain one or more U-Nets 502, one or more pyramid encoder-decoders from FIG.6, one or more 3D SWIN transformer decoders, or other architectures described herein.
  • the mesh elements of 3D representation 500 may be provided to the U-Net 502, which may use a U-shaped structure of convolution 504/unconvolution 512 and/or pooling 506/unpooling operations 510 to extract hierarchical features.
  • the global-most features are extracted at the lowest resolution (508).
  • Vectors 414 or 514 containing hierarchical neural network features may be assembled into mesh element representation vectors 416, and may be provided to an ensemble of ML models 418 (e.g., for resolving mesh element labels).
  • Voting (420) may be performed to resolve ambiguities between the mesh element labels predicted by the ensemble of ML models.
  • the final predicted mesh element labels 422 are sent to the output.
  • training data may include at least a 3D mesh of an arch and the set of mesh element labels (or object masks) which indicates to which tooth each mesh element belongs.
  • This set of mesh element labels (or masks) may undergo the iterative addition of noise, such as in the forward pass shown in FIG.2.
  • the resulting sets of increasingly noisy mesh element labels may be used as training data to train a denoising diffusion ML model (e.g., such as a U-Net).
  • a denoising diffusion ML model e.g., such as a U-Net
  • Such a denoising diffusion ML model may be trained to take as input a noisy (or random) set of mesh element labels (or object masks) for an instant 3D mesh (e.g., of a pre-segmentation dental arch), and gradually, over many iterations, denoise that set of mesh element labels (or object masks).
  • the denoised set of labels (or object masks) may be outputted by the reverse pass in FIG.2 and used to segment the instant mesh (or point cloud or other 3D representation).
  • the set of mesh element labels (or object masks) may undergo encoding using an encoder, which may provide a latent representation to the forward pass. After completion of the reverse pass, in deployment, the denoised representation may undergo reconstruction using a decoder.
  • the mesh element labels may be used to segment the teeth of a pre-segmentation arch.
  • the mesh element labels may, in some instances, be used to label mesh elements as a part of a mesh cleanup operation (e.g., labeling mesh elements for removal, smoothing, scaling or other modifications pursuant to the generation of an artifact for use in digital oral care).
  • Systems of disclosure may train a denoising diffusion model to segment a 3D representation of oral care data, such as a 3D mesh of the patient’s dentition.
  • FIG.4 pertains to 3D mesh segmentation and 3D mesh cleanup. Both of these techniques share an important attribute, the labelling of 3D mesh elements.
  • the various mesh elements of the mesh may be labelled according to which portion of the dental anatomy the mesh elements belong (e.g., gums, upper right central incisor, lower left 2 nd bicuspid). Tooth mesh segmentation may label elements according to membership in the various teeth, as designated by one or more of the dental notation systems mentioned herein (e.g., Palmer). In some implementations, for facial-lingual segmentation, each of the mesh elements may be labelled according to membership in either the facial side of the arch or the lingual side of the arch. Other dental anatomy segmentation implementations are possible.
  • oral care arguments 202 may be provided to the denoising diffusion models for segmentation or mesh cleanup described herein, such as decimation factor (or decimation percentage).
  • the decimation factor may control the extent to which pre-segmentation meshes are decimated before mesh element labeling occurs.
  • the pre-segmentation mesh e.g., a dental arch produced by an intraoral scanner or CT scanner
  • Mesh element feature vectors may be computed, one or more vectors for each mesh element.
  • One or more mesh element features from elsewhere in this disclosure may be used to form a feature vector for a mesh element.
  • Mesh elements may comprise edges, faces, vertices or voxels (or any combination thereof).
  • the lists of mesh element feature vectors may be provided to a neural network that refines the mesh element vectors, to extract local and global features from those feature vectors, such as the U-Net structure shown in FIG.4.
  • a U-Net inside the HNNFEM may comprise 3D mesh convolution and/or subsequent 3D mesh pooling operations, which may serve to reduce the resolution of the mesh and extract neural network features at increasingly global scales. After a succession of such operations, after the global-most neural network features have been extracted, there may be a succession of 3D mesh unpooling and 3D mesh unconvolution operations which return the mesh to the original scale. After the various levels of the U-Net structure have generated outputs, there may be scale-up and concatenation operations.
  • the HNNFEM may generate one or more 3D mesh element-wise feature vectors which may have been scaled-up to the original mesh resolution. These feature vectors may be concatenated and provided as inputs to one or more ML models for mesh element classification or labelling. Any of the supervised ML models disclosed elsewhere in this disclosure may be used for this classification, such as for example SVM or Logistic Regression. In some examples, an ensemble of ML classifiers may take the mesh element representation vectors as inputs and generate mesh element classification labels. In some implementations, an ensemble of fully connected neural networks may be used for such classification.
  • a multi-layer perceptron may be used for this classification, comprising linear layers and associated ReLU activation functions (e.g., which have the advantage in that not all neurons fire at the same time – which enables a more tailored response to the inputs than activation functions which all fire for every evaluation) and batch normalization operations.
  • ReLU activation functions e.g., which have the advantage in that not all neurons fire at the same time – which enables a more tailored response to the inputs than activation functions which all fire for every evaluation
  • Other activation functions are possible, such as the activation functions mentioned elsewhere in this disclosure.
  • Each of the ensemble of ML models may generate a predicted class label for each mesh element.
  • a voting mechanism may then be employed to combine these results and output a final class label prediction for each 3D mesh element.
  • Some implementations may use a transformer to assist in applying labels to mesh elements.
  • the mesh cleanup and mesh segmentation implementations may, in some instances, differ in the arrangement of the ground truth mesh element labels.
  • ground truth data may be given for each arch mesh that is to be segmented (i.e., each of the teeth in the arch has its mesh elements labeled according to which tooth the mesh element belongs).
  • a loss can be computed for facial-lingual segmentation and/or for teeth-gums segmentation (as defined in US Provisional Application No. US63/366490).
  • Various of the loss functions of this disclosure may be used to compare the mesh element labels of a predicted segmentation to the mesh element labels of a corresponding ground truth segmentation.
  • Cross-entropy loss is an example of the several candidate loss functions for this comparison.
  • ground truth data may be given for each arch mesh that is to be cleaned up.
  • each mesh element corresponding to extraneous material in the arch may be labelled as such, and each mesh element that does not comprise extraneous material (i.e., that is to be retained after mesh cleanup) is labelled as such.
  • each mesh element corresponding to a divot in the arch may be labelled as such, and each mesh element that does not comprise a divot (i.e., that is to be retained after mesh cleanup) is labelled as such.
  • a process is executed to copy the mesh element of a particular label into a new mesh (aka mesh cutting), and the mesh is saved (e.g., to an electronic storage medium) for further processing.
  • a process is executed to remove each designated mesh element from the mesh (e.g., using classical mesh processing techniques), and then a process may optionally be executed to fill-in any holes which may have been created by that process (e.g., using the technique described elsewhere in this disclosure which has been trained for mesh in-filling).
  • Mesh element labeling for mesh segmentation may also be accomplished using an autoencoder which has been trained for the purpose, such as a variational autoencoder, as described elsewhere in this disclosure.
  • an encoder-decoder network may be trained for mesh element labeling.
  • the denoising diffusion techniques of this disclosure may be trained to predict archforms.
  • An archform may have a shape which describes aspects of the shape of an arch of teeth.
  • An archform may comprise data structures such as one or more 3D meshes (e.g., such as a mesh which aligns with at least one coordinate axis of at least one tooth’s local coordinate system), 3D point clouds, 3D polylines or sets of control points (e.g., control points which define a spline).
  • Such data structures may be generated, for example, using the denoising diffusion method described in FIG.2.
  • the denoising diffusion techniques of this disclosure may be trained to generate information pertaining to IPR for use in setups prediction.
  • Information pertaining to IPR may include one or more of 1) an IPR cut surface for a particular tooth, 2) a measure of IPR magnitude to be applied to a side of a particular tooth, 3) a designation of whether a particular tooth is to undergo IPR – including an indication of whether IPR is to be applied to the mesial or distal side of the tooth, or 4) a designation of one or more stages in which a particular tooth is to undergo IPR.
  • the training data for such a model may be generated, at least in part, by iteratively adding noise to 1) representations of one or more IPR cut surfaces from the cases in the training dataset, 2) representations of one or more teeth or other aspects of the patient’s detention, or 3) one or more tooth transforms.
  • An ML model e.g., a U-Net
  • IPR cut surfaces may be initialized, at least in part, randomly or are initialized using default values.
  • IPR cut surfaces may be generated by the denoising diffusion techniques described herein.
  • a denoising diffusion model may be trained to render a determination of whether a target tooth is accessible for IPR.
  • Information pertaining to IPR may be generated, for example, using the denoising diffusion method described in FIG.2.
  • the method in FIG.2 may predict an orthodontic setup which has optional associated IPR information.
  • a setup transform may be generated, along with one or more associated items of IPR information (e.g., one or more IPR cut surfaces, a flag to indicate whether IPR is to be performed, a list of states in which to perform IPR, etc.).
  • techniques of this disclosure may be trained to predict trimlines.
  • a trimline may define a cutting path to remove a thermoformed tray from a fixture model (e.g., a tray for orthodontic aligner treatment or an indirect bonding tray to deliver brackets to teeth for orthodontic treatment).
  • a fixture model e.g., a tray for orthodontic aligner treatment or an indirect bonding tray to deliver brackets to teeth for orthodontic treatment.
  • One or more polylines (or meshes or point clouds) may be used to define a trimline. Such polylines (or meshes or point clouds) may be generated, for example, using the denoising diffusion method described in FIG.2.
  • the denoising diffusion techniques of this disclosure may be trained to predict one or more local orthogonal coordinate axes for a tooth (e.g., such as to predict one or more of X, Y and Z orthogonal axes for a tooth).
  • the denoising diffusion techniques of this disclosure may be trained to predict one or more archform coordinate axes.
  • a position along an archform coordinate axis may comprise a tuple [l, d, e] relative to a reference archform spline S which approximates the shape of an arch of teeth.
  • a rotation may comprise a tuple [a, b, g] which stands for alpha, beta and gamma rotations.
  • Alpha describes a rotation around the l-axis.
  • Beta describes a rotation around the d-axis.
  • Gamma describes a rotation around the e-axis.
  • a full tuple to describe position and rotation may comprise [l, d, e, a, b, g].
  • p is a point along S with arch length l.
  • d is the distance between a tooth origin t and the reference archform spline S.
  • the tooth origin t is obtained by translating up along the d axis by a distance 'd', and then translating along the e-axis by a distance 'e'.
  • the e-axis is perpendicular to the d-axis and the l-axis and may be defined to come out of the page or into the page.
  • e stands for eminence.
  • l stands for the length across the archform spline.
  • d stands for the distance away from the archform spline.
  • the coordinate systems described herein may facilitate later automation for oral care treatment of the patient, such as automated setups prediction.
  • One or more transforms, vectors or tuples may be used to define a coordinate system (or to define a pose within a coordinate system). Such transforms (or vectors or tuples) may be generated, for example, using the denoising diffusion method described in FIG. 2.
  • Table 3 describes the input data and generated data for several non-limiting examples of the generative techniques described herein. Denoising diffusion models described herein may, in some implementations, be trained for the generation (or modification) of the input data in Table 3, yielding the generated data in Table 3.
  • Techniques of this disclosure may be trained to generate (or modify) point clouds (e.g., where a point may be described as a 1D vector - such as (x, y, z)), polylines (points connected in order by edges), 3D meshes (points connected via edges to form faces), splines (which may be computed through a set of generated control points), sparse voxelized representations (which may be described as a set of points corresponding to the centroid of each voxel or to some other landmark of the voxel – such as the boundary of the voxel), one or more transforms (which may take the form of one or more 1D vectors or one or more 2D matrices – such as a 4x4 matrix) or the like.
  • point clouds e.g., where a point may be described as a 1D vector - such as (x, y, z)
  • polylines points connected in order by edges
  • 3D meshes points connected
  • a voxelized representation may be computed from a 3D point cloud or a 3D mesh.
  • a 3D point cloud may be computed from a voxelized representation.
  • a 3D mesh may be computed from a 3D point cloud.
  • One or more mesh element labels for one or dentition e.g., a mesh including more aspects of the patient's dentition.
  • a label teeth and gums may flag a mesh element for removal or modification.
  • CTA trimline 3D representation of patient's 3D representation of trimline (e.g., 3D mesh dentition (e.g., a mesh including or 3D polyline) teeth and gums)
  • trimline e.g., 3D mesh dentition (e.g., a mesh including or 3D polyline) teeth and gums)
  • CTA setups Two or more tooth meshes and/or Transforms for one or more teeth (for final setups tooth transforms. Tooth meshes or intermediate may be in their maloccluded poses. staging) Transforms may correspond to maloccluded poses.
  • Hardware e.g., One or more (segmented) teeth Transform for placement of hardware relative bracket/attachm to the one or more teeth ent
  • Archform 3D representation of patient's 3D polyline or a 3D mesh or surface, that generation dentition e.g., a mesh including describes the contours or layout of an arch of teeth and gums.
  • Generation dentition e.g., a mesh including describes the contours or layout of an arch of teeth and gums.
  • Generated oral 3D representation of patient's One or more oral care appliance components care appliance dentition e.g., a mesh including with shape and/or structure that is customized component teeth and gums
  • One or more transforms which place a library appliance dentition e.g., a mesh including component relative to aspects of the patient's component teeth and gums). May comprise dentition (e.g., for dental one or more segmented teeth. restoration) Table 3.
  • an automated setups prediction model e.g., a denoising diffusion probabilistic model
  • Such a model may be trained on cohort patient case data.
  • One or more oral care metrics may be computed on each case to quantify or measure aspects of that case's curve-of-spee.
  • one or more of such metrics may be provided to the setups prediction model, for example, to influence the model regarding the geometry and/or structure of each case's curve-of-spee.
  • that same input pathway to the trained neural network may be configured with one or more values as instructions to the model about an intended curve-of-spee. Such values may automatically generate a setup with a curve-of-spee which meets the aesthetic and/or medical treatment needs of the particular patient case.
  • a curve-of-spee metric may measure the curvature of the occlusal or incisal surfaces of the teeth on either the left or right sides of the arch, with respect to the occlusal plane.
  • the occlusal plane may, in some instances, be computed as a surface which averages the incisal or occlusal surfaces of the teeth (for one or both arches).
  • a curvature metric may be computed along a normal vector, such as a vector which is normal to the occlusal plane.
  • a curvature metric may be computed along the normal vector of another plane.
  • an XY plane may be defined to correspond to the occlusal plane.
  • An orthogonal plane may be defined as the plane that is orthogonal to the occlusal plane, which also passes through a curve- of-spee line segment, where the curve-of-spee line segment is defined by a first endpoint which is a landmarking point on a first tooth (e.g., canine) and a second endpoint which is a landmarking point on the most-posterior tooth of the same side of the arch.
  • a landmarking point can in some implementations be located along the incisal edge of a tooth or on the cusp of a tooth.
  • the landmarking points for the intermediate teeth may form a curved path, such as may be described by a polyline.
  • the following is a non-limiting list of curve-of-spee oral care metrics.
  • a curve-of-spee metric may be computed using the following 4 steps. i) Line: Form a line between the highest cusp on the most posterior tooth and the cusp of the first tooth. ii) Curve_Point_A: Given the set of teeth between the most posterior tooth and the first tooth, find the highest point of the lowest tooth. iii) Curve_Point_B: Project Curve_Point_A onto the Line to find a point (Curve_Point_B) along the line that is closest to Curve_Point_A.
  • Curve-Of-Spee Find the height difference between Curve_Point_B and Curve_Point_A.
  • Compute Curve of Spee in this plane by measuring the distance between the farthest of the projected intermediate points to the projected curve-of-spee line segment. This yields a measure for the curvature of the arch relative to the occlusal plane.
  • 4) Skip the projection and compute the distances and curvatures in the 3D space.
  • Compute Curve of Spee by measuring the distance between the farthest of the intermediate points to the curve-of- spee line segment. This yields a measure for the curvature of the arch in 3D space.
  • 5) Compute the slope of the projected curve-of-spee line segment on the occlusal plane.
  • Curve-of-spee metrics 5 and 6 may help the network to reduce some more degrees of freedom in defining how the patient’s arch is curved in the posterior of the mouth.
  • Techniques described herein e.g., denoising diffusion probabilistic models
  • Optional oral care arguments 202 may be provided to the denoising ML model 210 (e.g., a U-Net) which is trained to perform the reverse pass 220.
  • Oral care arguments may include oral care parameters as disclosed herein, or other real-valued, text-based or categorical inputs which specify intended aspects of the one or more 3D oral care representations which are to be generated.
  • oral care arguments may include oral care metrics, which may describe intended aspects of the one or more 3D oral care representations which are to be generated. Oral care arguments are specifically adapted to the implementations described herein.
  • the oral care arguments may specify the intended the designs (e.g., including shape and/or structure) of 3D oral care representations which may be generated (or modified) according to techniques described herein.
  • implementations using the specific oral care arguments disclosed herein generate more accurate 3D oral care representations than implementations that do not use the specific oral care arguments.
  • a text encoder may encode a set of natural language instructions from the clinician (e.g., generate a text embedding).
  • a text string may comprise tokens.
  • An encoder for generating text embeddings may, in some implementations, apply either mean-pooling or max-pooling between the token vectors.
  • a transformer e.g., BERT or Siamese BERT
  • a transformer may be trained to extract embeddings of text for use in digital oral care (e.g., by training the transformer on examples of clinical text, such as those given below).
  • a model for generating text embeddings may be trained using transfer learning (e.g., initially trained on another corpus of text, and then receive further training on text related to digital oral care).
  • Some text embeddings may encode text at the word level.
  • Some text embeddings may encode text at the token level.
  • a transformer for generating a text embedding may, in some implementations, be trained, at least in part, with a loss calculation which compares predicted outputs to ground truth outputs (e.g., softmax loss, multiple negatives ranking loss, MSE margin loss, cross-entropy loss or the like).
  • a loss calculation which compares predicted outputs to ground truth outputs (e.g., softmax loss, multiple negatives ranking loss, MSE margin loss, cross-entropy loss or the like).
  • the non-text arguments such as real values or categorical values, may be converted to text, and subsequently embedded using the techniques described herein.
  • U-Nets may, in some implementations, use U-Nets to extract local or global neural network features from a 3D point cloud or other 3D representation.
  • 3D oral care representations are described herein as such because 3-dimensional representations are currently state of the art. Nevertheless, 3D oral care representations are intended to be used in a non-limiting fashion to encompass any representations of 3-dimensions or higher orders of dimensionality (e.g., 4D, 5D, etc.), and it should be appreciated that machine learning models can be trained using the techniques disclosed herein to operate on representations of higher orders of dimensionality.
  • input data may comprise 3D mesh data, 3D point cloud data, 3D surface data, 3D polyline data, 3D voxel data, or data pertaining to a spline (e.g., control points).
  • An encoder- decoder structure may comprise one or more encoders, or one or more decoders.
  • the encoder may take as input mesh element feature vectors for one or more of the inputted mesh elements. By processing mesh element feature vectors, the encoder is trained in a manner to generate more accurate representations of the input data.
  • the mesh element feature vectors may provide the encoder with more information about the shape and/or structure of the mesh, and therefore the additional information provided allows the encoder to make better-informed decisions and/or generate more-accurate latent representations of the mesh.
  • encoder-decoder structures include U-Nets, autoencoders or transformers (among others).
  • a representation generation module may comprise one or more encoder-decoder structures (or portions of encoders-decoder structures – such as individual encoders or individual decoders).
  • a representation generation module may generate an information-rich (optionally reduced-dimensionality) representation of the input data, which may be more easily consumed by other generative or discriminative machine learning models.
  • a U-Net may comprise an encoder, followed by a decoder.
  • the architecture of a U-Net may resemble a U shape.
  • the encoder may extract one or more global neural network features from the input 3D representation, zero or more intermediate-level neural network features, or one or more local neural network features (at the most local level as contrasted with the most global level).
  • the output from each level of the encoder may be passed along to the input of corresponding levels of a decoder (e.g., by way of skip connections).
  • the decoder may operate on multiple levels of global-to-local neural network features. For instance, the decoder may output a representation of the input data which may contain global, intermediate or local information about the input data.
  • the U-Net may, in some implementations, generate an information-rich (optionally reduced-dimensionality) representation of the input data, which may be more easily consumed by other generative or discriminative machine learning models.
  • An autoencoder may be configured to encode the input data into a latent form.
  • An autoencoder may train an encoder to reformat the input data into a reduced-dimensionality latent form in between the encoder and the decoder, and then train a decoder to reconstruct the input data from that latent form of the data.
  • a reconstruction error may be computed to quantify the extent to which the reconstructed form of the data differs from the input data.
  • the latent form may, in some implementations, be used as an information-rich reduced-dimensionality representation of the input data which may be more easily consumed by other generative or discriminative machine learning models.
  • an autoencoder may be trained to input a 3D representation, encode that 3D representation into a latent form (e.g., a latent embedding), and then reconstruct a close facsimile of that input 3D representation as the output.
  • a transformer may be trained to use self-attention to generate, at least in part, representations of its input.
  • a transformer may encode long-range dependencies (e.g., encode relationships between a large number of inputs).
  • a transformer may comprise an encoder or a decoder.
  • Such an encoder may, in some implementations, operate in a bi-directional fashion or may operate a self-attention mechanism.
  • a decoder may, in some implementations, may operate a masked self-attention mechanism, may operate a cross-attention mechanism, or may operate in an auto-regressive manner.
  • the self-attention operations of the transformers described herein may, in some implementations, relate different positions or aspects of an individual 3D oral care representation in order to compute a reduced-dimensionality representation of that 3D oral care representation.
  • the cross-attention operations of the transformers described herein may, in some implementations, mix or combine aspects of two (or more) different 3D oral care representations.
  • the auto-regressive operations of the transformers described herein may, in some implementations, consume previously generated aspects of 3D oral care representations (e.g., previously generated points, point clouds, transforms, etc.) as additional input when generating a new or modified 3D oral care representation.
  • the transformer may, in some implementations, generate a latent form of the input data, which may be used as an information-rich reduced-dimensionality representation of the input data, which may be more easily consumed by other generative or discriminative machine learning models.
  • an encoder-decoder structure may first be trained as an autoencoder. In deployment, one or more modifications may be made to the latent form of the input data.
  • This modified latent form may then proceed to be reconstructed by the decoder, yielding a reconstructed form of the input data which differs from the input data in one or more intended aspects.
  • Oral care arguments such as oral care parameters or oral care metrics may be supplied to the encoder, the decoder, or may be used in the modification of the latent form, to influence the encoder-decoder structure in generating a reconstructed form that has desired characteristics (e.g., characteristics which may differ from that of the input data).
  • Techniques of this disclosure may, in some instances, be trained using federated learning.
  • Federated learning may enable multiple remote clinicians to iteratively improve a machine learning model (e.g., validation of 3D oral care representations, mesh segmentation, mesh cleanup, other techniques which involve labeling mesh elements, coordinate system prediction, non-organic object placement on teeth, appliance component generation, tooth restoration design generation, techniques for placing 3D oral care representations, setups prediction, generation or modification of 3D oral care representations using autoencoders, generation or modification of 3D oral care representations using transformers, generation or modification of 3D oral care representations using diffusion models, 3D oral care representation classification, imputation of missing values), while protecting data privacy (e.g., the clinical data may not need to be sent “over the wire” to a third party). Data privacy is particularly important to clinical data, which is protected by applicable laws.
  • a machine learning model e.g., validation of 3D oral care representations, mesh segmentation, mesh cleanup, other techniques which involve labeling mesh elements, coordinate system prediction, non-organic object placement on teeth, appliance component generation, tooth restoration design generation, techniques for
  • a clinician may receive a copy of a machine learning model, use a local machine learning program to further train that ML model using locally available data from the local clinic, and then send the updated ML model back to the central hub or third party.
  • the central hub or third party may integrate the updated ML models from multiple clinicians into a single updated ML model which benefits from the learnings of recently collected patient data at the various clinical sites. In this way, a new ML model may be trained which benefits from additional and updated patient data (possibly from multiple clinical sites), while those patient data are never actually sent to the 3rd party. Training on a local in-clinic device may, in some instances, be performed when the device is idle or otherwise be performed during off-hours (e.g., when patients are not being treated in the clinic).
  • Devices in the clinical environment for the collection of data and/or the training of ML models for techniques described herein may include intra-oral scanners, CT scanners, X- ray machines, laptop computers, servers, desktop computers or handheld devices (such as smart phones with image collection capability).
  • contrastive learning may be used to train, at least in part, the ML models described herein. Contrastive learning may, in some instances, augment samples in a training dataset to accentuate the differences in samples from different classes and/or increase the similarity of samples of the same class.
  • a local coordinate system for a 3D oral care representation such as a tooth
  • a 3D oral care representation such as a tooth
  • transforms e.g., an affine transformation matrix, translation vector or quaternion
  • Systems of this disclosure may be trained for coordinate system prediction using past cohort patient case data.
  • the past patient data may include at least: one or more tooth meshes or one or more ground truth tooth coordinate systems.
  • Machine learning models such as: U-Nets, encoders, autoencoders, pyramid encoder-decoders, transformers, or convolution and/or pooling layers, may be trained for coordinate system prediction.
  • Representation learning may determine a representation of a tooth (e.g., encoding a mesh or point cloud into a latent representation, for example, using a U-Net, encoder, transformer, convolution and/or pooling layers or the like), and then predict a transform for that representation (e.g., using a trained multilayer perceptron, transformer, encoder, transformer, or the like) that defines a local coordinate system for that representation (e.g., comprising one or more coordinate axes).
  • a representation of a tooth e.g., encoding a mesh or point cloud into a latent representation, for example, using a U-Net, encoder, transformer, convolution and/or pooling layers or the like
  • a transform for that representation e.g., using a trained multilayer perceptron, transformer, encoder, transformer, or the like
  • a local coordinate system for that representation e.g., comprising one or more coordinate axes.
  • the mesh convolutional techniques described herein can leverage invariance to rotations, translations, and/or scaling of that tooth mesh to generate predications that techniques that are not invariant to the rotations, translations, and/or scaling of that tooth mesh cannot generate.
  • Pose transfer techniques may be trained for coordinate system prediction, in the form of predicting a transform for a tooth.
  • Reinforcement learning techniques may be trained for coordinate system prediction, in the form of predicting a transform for a tooth.
  • Machine learning models such as: U-Nets, encoders, autoencoders, pyramid encoder- decoders, transformers, or convolution and/or pooling layers, may be trained as a part of a method for hardware (or appliance component) placement.
  • Representation learning may train a first module to determine an embedded representation of a 3D oral care representation (e.g., encoding a mesh or point cloud into a latent form using an autoencoder, or using a U-Net, encoder, transformer, block of convolution and/or pooling layers or the like). That representation may comprise a reduced dimensionality form and/or information-rich version of the inputted 3D oral care representation.
  • a representation may be aided by the calculation of a mesh element feature vector for one or more mesh elements (e.g., each mesh element).
  • a representation may be computed for a hardware element (or appliance component).
  • Such representations are suitable to be provided to a second module, which may perform a generative task, such as transform prediction (e.g., a transform to place a 3D oral care representation relative to another 3D oral care representation, such as to place a hardware element or appliance component relative to one or more teeth) or 3D point cloud generation.
  • transform prediction e.g., a transform to place a 3D oral care representation relative to another 3D oral care representation, such as to place a hardware element or appliance component relative to one or more teeth
  • 3D point cloud generation e.g., a transform to place a 3D oral care representation relative to another 3D oral care representation, such as to place a hardware element or appliance component relative to one or more teeth
  • Such a transform may comprise an affine transformation matrix, translation vector or quatern
  • Machine learning models which may be trained to predict a transform to place a hardware element (or appliance component) relative to elements of patient dentition include: MLP, transformer, encoder, or the like.
  • Systems of this disclosure may be trained for 3D oral care appliance placement using past cohort patient case data.
  • the past patient data may include at least: one or more ground truth transforms and one or more 3D oral care representations (such as tooth meshes, or other elements of patient dentition).
  • the mesh convolution and/or mesh pooling techniques described herein leverage invariance to rotations, translations, and/or scaling of that tooth mesh to generate predications that techniques that are not invariant to the rotations, translations, and/or scaling of that tooth mesh cannot generate.
  • Pose transfer techniques may be trained for hardware or appliance component placement.
  • Reinforcement learning techniques may be trained for hardware or appliance component placement.
  • the denoising diffusion techniques may, in some implementations, trained to fill-in missing aspects of a 3D representation of a patient's dentition (e.g., of a segmented tooth or of a digital fixture model or digital fixture model component).
  • a 3D scan of a tooth crown may be missing some portions of one or more surfaces (e.g., due to occlusion by surrounding teeth during intraoral scanning).
  • a denoising diffusion model may generate training data by iteratively degrading or adding noise to a 3D representation of a tooth (e.g., by removing patches or portions of the surface of the tooth crown or tooth root).
  • a sequence of progressively nosier versions of the 3D representant of a tooth from the training dataset may be generated.
  • Noise may be added to a point cloud/mesh/voxelized representation of the tooth by 1) adding random translation to one or more mesh elements, 2) applying small random scaling or distortion to aspects of the tooth or 3) by removing mesh elements (e.g., creating holes), among other methods.
  • This sequence of progressively noisier versions of the tooth may then be used to train a denoising diffusion neural network (e.g., a U-Net or others encoder- decoder structures – such as pyramid encoders-decoders, transformers or autoencoders) to reconstruct (e.g., in small steps) a noisy tooth.
  • a denoising diffusion neural network e.g., a U-Net or others encoder- decoder structures – such as pyramid encoders-decoders, transformers or autoencoders
  • a noisy set of mesh elements e.g., initialized from a tooth which has missing data
  • the denoising diffusion neural network may be trained to iteratively remove noise and/or repair the surface of the tooth.
  • Fixture model components may be generated or modified using the denoising diffusion techniques of this disclosure (e.g., the method in FIG.2).
  • a digital fixture model may comprise 3D representations of the patient’s dentition, with optional fixture model components attached to that dentition.
  • 3D representation generation techniques of this disclosure e.g., denoising diffusion techniques to generate 3D point clouds, 3D voxelized representations, or other representations
  • Fixture model components may include 3D representations (e.g., 3D point clouds, 3D meshes, or voxelized representations) of one or more of the following non-limiting items: 1) interproximal webbing – which may fill-in space or smooth-out the gaps between teeth to ensure aligner removability. 2) blockout – which may be added to the fixture model to remove overhangs that might interfere with plastic tray thermoforming or to ensure aligner removability. 3) bite blocks - occlusal features on the molars or premolars intended to prop the bite open. 4) bite ramps - lingual features on incisors and cuspids intended to prop the bite open.
  • 3D representations e.g., 3D point clouds, 3D meshes, or voxelized representations
  • interproximal reinforcement - a structure on the exterior of an oral care appliance (e.g., an aligner tray), which may extend from a first gingival edge of the appliance body on a labial side of the appliance body along an interproximal region between the first tooth and the second tooth to a second gingival edge of the appliance body on a lingual side of the appliance body.
  • the effect of the interproximal reinforcement on the appliance body at the interproximal region may be stiffer than a labial face and a lingual face of the first shell. This may allow the aligner to grasp the teeth on either side of the reinforcement more firmly.
  • gingival ridge - a structure which may extend along the gingival edge of a tooth in the mesial-distal direction for the purpose of enhancing engagement between the aligner and a given tooth.
  • torque points - structures which may enhance force delivered to a given tooth at specified locations.
  • power ridges - structures which may enhance force delivered to a given tooth at a specified location.
  • dimples - structures which may enhance force delivered to a given tooth at specified locations.
  • digital pontic tooth – structure which may hold space open or reserve space in an arch for a tooth which is partially erupted, or the like.
  • a physical pontic is a tooth pocket that does not cover a tooth when the aligner is installed on the teeth.
  • the tooth pocket may be filled with tooth-colored wax, silicone, or composite to provide a more aesthetic appearance.
  • power bars - blockout added in an edentulous space to provide strength and support to the tray.
  • a power bar may fill-in voids.
  • Abutments or healing caps may be blocked-out with a power bar.
  • the trimline may define the path along which a clear aligner may be cut or separated from a physical fixture model, after 3D printing.
  • Techniques of this disclosure may also generate (or modify) other geometries that intrude on the tooth pocket or reinforce an oral care appliance (e.g., an orthodontic aligner tray). Techniques of this disclosure may, in some implementations, determine the location, size, and/or shape of a fixture model component to produce a desired treatment outcome.
  • Interproximal webbing may be generated using techniques of this disclosure, for example during the fixture model quality control phase of orthodontic aligner fabrication.
  • Interproximal webbing is material (e.g., which may comprise mesh elements, such as vertices/edges/faces, among others) that smooths-out or creates a positive fill in the interproximal areas of the fixture model.
  • Interproximal webbing is extra material added to the interproximal areas of the teeth in a digital fixture model (e.g., which may be 3D printed and rendering in physical form) to reduce the tendency of the aligner, retainer, attachment template, or bonding tray to lock onto the physical fixture model during the orthodontic aligner thermoforming process.
  • Interproximal webbing may improve the functioning of the aligner tray by improving the ability of the tray to slide off of the fixture model after thermoforming, or by improving the fit of the tray onto the patient’s teeth (e.g., making the tray easier to insert onto the teeth or to remove from the teeth).
  • Blockout e.g., which may comprise mesh elements, such as vertices/edges/faces, among others
  • Blockout is material which may be added to undercut regions of a digital fixture model, so that the aligner, retainer, attachment template, or bonding tray, 3D printed mold, or other oral care appliance, from locking onto the physical fixture model during the orthodontic aligner thermoforming process.
  • Block out may be generated to fill-in a portion of a digital fixture model with an undercut, so that a thermoformed aligner tray does not grab onto that undercut. Blockout may improve the functioning of the aligner tray by improving the ability of the tray to slide off of the fixture model after thermoforming. Blockout may facilitate later thermoforming (e.g., help avoid the aligner tray getting stuck on the physical fixture model).
  • Pontic tooth designs may be generated using techniques of this disclosure.
  • a pontic tooth is a digital 3D representation of a tooth which may act as a placeholder in an arch. In an aligner, a pontic may function as a tooth pocket that may be filled with a tooth-colored material (e.g., wax) to improve aesthetics.
  • a pontic tooth may hold a space in the arch open during orthodontic setups generation.
  • transforms for intermediate stages may be generated.
  • One or more pontic teeth may be defined to hold open space for missing teeth or act as a placeholder as space closes or opens in the arch during setups predicted staging or an unerupted tooth may erupt into a space where a pontic is present. Creating a pocket for the tooth to erupt into.
  • Digital pontic teeth may be placed in (or generated within) an arch (e.g., during setups generation or fixture model generation) to reserve space in an arch for missing or extracted teeth (e.g., so that adjacent teeth do not encroach upon that space over the course of successive intermediate stages of orthodontic treatment).
  • digital pontic tooth may be used when the space (e.g., the space that is to be held open) is at least a threshold dimension (e.g., a width of 4mm, among others).
  • digital pontic teeth for UL4-UR4 or LL4-LR4 may be placed (or generated or modified) when space is available or when there is a partially erupted tooth within the space.
  • a digital pontic tooth may be placed over the erupting tooth to maintain a space for the erupting tooth to erupt into.
  • Pontic teeth may be placed (or generated or modified) to be inside the gingiva, to minimize (or avoid) heavy occlusal contacts (e.g., contacts between the chewing surfaces of the upper or lower arches), or to cover an erupting tooth (when present), among other conditions.
  • occlusal contacts e.g., contacts between the chewing surfaces of the upper or lower arches
  • erupting tooth when present
  • 3D oral care representations which may be generated include: mesh element labels for segmentation or mesh cleanup, transforms for use in setups generation (or fixture model or appliance generation), coordinate systems (e.g., for use as local coordinate systems of teeth), 3D point clouds (or other 3D representations described herein) that describe teeth (or appliance components or fixture model components), or others described herein.
  • computing systems specifically adapted to generate 3D oral care representations for use in oral care appliance generation are improved.
  • aspects of the present disclosure improve the performance of a computing system having a 3D representation of the patient’s dentition by reducing the consumption of computing resources.
  • aspects of the present disclosure reduce computing resource consumption by encoding inputs (e.g., 3D representations of teeth, mesh element labels, tooth transforms, coordinate systems, 3D representations of appliance components, 3D representations of fixture model components, or other 3D oral care representations described herein) into reduced dimensionality latent representations (e.g., using the encoder portion of a reconstruction autoencoder).
  • the methods may potentially reduce thousands of mesh elements (each of which may be described by several real numbers) into a latent vector of potentially hundreds of real numbers, so that computing resources are not unnecessarily wasted by processing excess quantities of data.
  • encoding the input data into latent representations does not reduce the overall predictive accuracy of the computing system (and indeed may actually improve predictions because the input provided to the ML model after the encoding is a more accurate (or better) representation of the patient’s dentition).
  • the latent representations may remove noise or other artifacts which are unimportant (and which may reduce the accuracy of the predictive models) and provide a distilled set of values that accurately describe the 3D oral care representation that is encoded. That is, aspects of the present disclosure provide for more efficient allocation of computing resources, and in a way that improves the accuracy of the underlying system.
  • aspects of the present disclosure may need to be executed in a time-constrained manner, such as when an oral care appliance must be generated for a patient immediately after intraoral scanning (e.g., while the patient waits in the clinician’s office).
  • aspects of the present disclosure are necessarily rooted in the underlying computer technology of 3D oral care representation generation using denoising diffusion models and cannot be performed by a human, even with the aid of pen and paper.
  • implementations of the present disclosure must be capable of: 1) storing thousands or millions of mesh elements of the patient’s dentition (or appliance components or fixture model components) in a manner that can be processed by a computer processor; 2) performing calculation on thousands or millions of mesh elements, e.g., to quantify aspects of the shape and or/structure of a 3D representation of oral care data; 3) for each example in a training dataset of thousands of examples, generating hundreds of iteratively noisier versions of each sample; 4) using those hundreds of thousands of noisy examples to train a denoising ML model; and 5) generating, using the denoising ML model, 3D oral care representations for use in oral care appliance generation (each of which may comprise thousands or millions of mesh elements), and doing so during the course of a short office visit.
  • a denoising ML model 210 can be trained by providing noisy data from the forward pass 206 to the denoising ML model 210, using the denoising ML model 210 to generate a predicted 3D oral care representation (e.g., a denoised 3D oral care representation), and comparing the predicted 3D oral care representation to a ground truth (or reference) 3D oral care representation.
  • a loss may be computed based on the comparing (e.g., MSE loss, or others described herein), and be used to update the weights of the denoising ML model 210 (e.g., via backpropagation).

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Dental Tools And Instruments Or Auxiliary Dental Instruments (AREA)

Abstract

L'invention concerne des systèmes et des techniques pour générer automatiquement une structure de données pour un traitement de soins buccaux à l'aide de modèles d'apprentissage automatique entraînés. Le procédé consiste à recevoir un ou plusieurs attributs qui décrivent la sortie prévue du modèle d'apprentissage automatique entraîné. Une ou plusieurs représentations bruitées de la sortie prévue sont générées par un ou plusieurs processeurs informatiques. Ces représentations bruitées sont ensuite débruitées à l'aide d'un modèle d'apprentissage automatique entraîné, également exécuté par les processeurs informatiques. Les représentations débruitées de la sortie prévue sont générées, fournissant des données plus précises et fiables. Sur la base des représentations débruitées générées, un ou plusieurs aspects de traitements de soins buccaux numériques sont automatiquement définis par les processeurs informatiques. Ces systèmes et techniques permettent la création d'une structure de données complète pour un traitement de soins buccaux, améliorant des processus de planification et de prise de décision de traitement dans le domaine de la santé buccale.
PCT/IB2023/062713 2022-12-14 2023-12-14 Débruitage de modèles de diffusion pour soins buccaux numériques Ceased WO2024127318A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP23832829.8A EP4633528A1 (fr) 2022-12-14 2023-12-14 Débruitage de modèles de diffusion pour soins buccaux numériques
CN202380085976.0A CN120322210A (zh) 2022-12-14 2023-12-14 用于数字口腔护理的去噪扩散模型

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US202263432627P 2022-12-14 2022-12-14
US63/432,627 2022-12-14
US202363522774P 2023-06-23 2023-06-23
US63/522,774 2023-06-23

Publications (1)

Publication Number Publication Date
WO2024127318A1 true WO2024127318A1 (fr) 2024-06-20

Family

ID=89426652

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2023/062713 Ceased WO2024127318A1 (fr) 2022-12-14 2023-12-14 Débruitage de modèles de diffusion pour soins buccaux numériques

Country Status (3)

Country Link
EP (1) EP4633528A1 (fr)
CN (1) CN120322210A (fr)
WO (1) WO2024127318A1 (fr)

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN118675033A (zh) * 2024-08-26 2024-09-20 浙江霖研精密科技有限公司 基于生成对抗网络的形状缺陷样本生成方法、系统及介质
CN118861950A (zh) * 2024-09-25 2024-10-29 国网山东省电力公司济南供电公司 变压器多参数融合状态评估方法、装置、设备及介质
CN118887128A (zh) * 2024-08-08 2024-11-01 西安数合信息科技有限公司 工业缺陷图像生成方法及装置、存储介质和电子设备
CN119066475A (zh) * 2024-11-04 2024-12-03 浙江大学 基于表征引导多模态扩散模型的活动识别方法和装置
CN119337423A (zh) * 2024-10-16 2025-01-21 中电信人工智能科技(北京)有限公司 数据处理方法、系统、电子设备及非易失性存储介质
CN119809970A (zh) * 2024-12-17 2025-04-11 北京百度网讯科技有限公司 点云增强方法和装置
CN119862800A (zh) * 2025-03-25 2025-04-22 西北工业大学 在线预测场景下航空发动机剩余寿命不确定性量化方法
CN120259113A (zh) * 2025-06-05 2025-07-04 福建帝视科技集团有限公司 一种基于扩散模型的文本条件引导的图像外扩方法与终端
CN120279342A (zh) * 2025-06-10 2025-07-08 沈阳久道科技有限公司 基于计算机辅助的农产品品质检测分级方法及系统
CN120356154A (zh) * 2025-04-18 2025-07-22 北京中科金财科技股份有限公司 一种基于动态优化扩散模型的人群密度识别方法及系统
WO2025252257A1 (fr) * 2024-09-29 2025-12-11 重庆邮电大学 Procédé de production d'implant dentaire auxiliaire basé sur un modèle de diffusion

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020026117A1 (fr) 2018-07-31 2020-02-06 3M Innovative Properties Company Procédé de génération automatisée de configurations finales de traitement orthodontique
WO2020240351A1 (fr) 2019-05-31 2020-12-03 3M Innovative Properties Company Création automatisée d'appareils dentaires de restauration des dents
US20210118132A1 (en) * 2019-10-18 2021-04-22 Retrace Labs Artificial Intelligence System For Orthodontic Measurement, Treatment Planning, And Risk Assessment
WO2021233767A1 (fr) * 2020-05-19 2021-11-25 3Shape A/S Procédé de débruitage d'images dentaires par adaptation de domaine
WO2021240290A1 (fr) 2020-05-26 2021-12-02 3M Innovative Properties Company Génération et pose d'appareils dentaires de restauration dentaire basées sur des réseaux neuronaux
WO2021245480A1 (fr) 2020-06-03 2021-12-09 3M Innovative Properties Company Système pour générer un traitement d'aligneur orthodontique par étapes
WO2022057373A1 (fr) 2020-09-18 2022-03-24 苏州浪潮智能科技有限公司 Procédé de gestion de disque à double port, appareil et terminal et support de stockage
WO2022123402A1 (fr) 2020-12-11 2022-06-16 3M Innovative Properties Company Traitement automatisé de balayages dentaires à l'aide d'un apprentissage profond géométrique

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020026117A1 (fr) 2018-07-31 2020-02-06 3M Innovative Properties Company Procédé de génération automatisée de configurations finales de traitement orthodontique
US20210259808A1 (en) 2018-07-31 2021-08-26 3M Innovative Properties Company Method for automated generation of orthodontic treatment final setups
WO2020240351A1 (fr) 2019-05-31 2020-12-03 3M Innovative Properties Company Création automatisée d'appareils dentaires de restauration des dents
US20210118132A1 (en) * 2019-10-18 2021-04-22 Retrace Labs Artificial Intelligence System For Orthodontic Measurement, Treatment Planning, And Risk Assessment
WO2021233767A1 (fr) * 2020-05-19 2021-11-25 3Shape A/S Procédé de débruitage d'images dentaires par adaptation de domaine
WO2021240290A1 (fr) 2020-05-26 2021-12-02 3M Innovative Properties Company Génération et pose d'appareils dentaires de restauration dentaire basées sur des réseaux neuronaux
WO2021245480A1 (fr) 2020-06-03 2021-12-09 3M Innovative Properties Company Système pour générer un traitement d'aligneur orthodontique par étapes
WO2022057373A1 (fr) 2020-09-18 2022-03-24 苏州浪潮智能科技有限公司 Procédé de gestion de disque à double port, appareil et terminal et support de stockage
WO2022123402A1 (fr) 2020-12-11 2022-06-16 3M Innovative Properties Company Traitement automatisé de balayages dentaires à l'aide d'un apprentissage profond géométrique

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
ANWAAR ULHAQ ET AL: "Efficient Diffusion Models for Vision: A Survey", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 7 October 2022 (2022-10-07), XP091346293 *
J. M. KANTERK. VEERAMACHANENI: "Deep feature synthesis: Towards automating data science endeavors", 2015 IEEE INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA, 2015, pages 1 - 10, XP032826310, DOI: 10.1109/DSAA.2015.7344858
P. CIGNONIC. ROCCHINIR. SCOPIGNO: "Computer Graphics Forum", vol. 17, June 1998, BLACKWELL PUBLISHERS, pages: 167 - 174
TONIONI A ET AL.: "Learning to detect good 3D keypoints.", INT J COMPUT. VIS, vol. 126, 2018, pages 1 - 20, XP036405732, DOI: 10.1007/s11263-017-1037-3
XIAOHUI ZENG ET AL: "LION: Latent Point Diffusion Models for 3D Shape Generation", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 12 October 2022 (2022-10-12), XP091342764 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN118887128A (zh) * 2024-08-08 2024-11-01 西安数合信息科技有限公司 工业缺陷图像生成方法及装置、存储介质和电子设备
CN118675033A (zh) * 2024-08-26 2024-09-20 浙江霖研精密科技有限公司 基于生成对抗网络的形状缺陷样本生成方法、系统及介质
CN118861950A (zh) * 2024-09-25 2024-10-29 国网山东省电力公司济南供电公司 变压器多参数融合状态评估方法、装置、设备及介质
WO2025252257A1 (fr) * 2024-09-29 2025-12-11 重庆邮电大学 Procédé de production d'implant dentaire auxiliaire basé sur un modèle de diffusion
CN119337423A (zh) * 2024-10-16 2025-01-21 中电信人工智能科技(北京)有限公司 数据处理方法、系统、电子设备及非易失性存储介质
CN119337423B (zh) * 2024-10-16 2025-08-22 中电信人工智能科技(北京)有限公司 数据处理方法、系统、电子设备及非易失性存储介质
CN119066475A (zh) * 2024-11-04 2024-12-03 浙江大学 基于表征引导多模态扩散模型的活动识别方法和装置
CN119809970A (zh) * 2024-12-17 2025-04-11 北京百度网讯科技有限公司 点云增强方法和装置
CN119862800A (zh) * 2025-03-25 2025-04-22 西北工业大学 在线预测场景下航空发动机剩余寿命不确定性量化方法
CN120356154A (zh) * 2025-04-18 2025-07-22 北京中科金财科技股份有限公司 一种基于动态优化扩散模型的人群密度识别方法及系统
CN120259113A (zh) * 2025-06-05 2025-07-04 福建帝视科技集团有限公司 一种基于扩散模型的文本条件引导的图像外扩方法与终端
CN120279342A (zh) * 2025-06-10 2025-07-08 沈阳久道科技有限公司 基于计算机辅助的农产品品质检测分级方法及系统

Also Published As

Publication number Publication date
EP4633528A1 (fr) 2025-10-22
CN120322210A (zh) 2025-07-15

Similar Documents

Publication Publication Date Title
WO2024127318A1 (fr) Débruitage de modèles de diffusion pour soins buccaux numériques
EP4634798A1 (fr) Techniques de réseau neuronal pour la création d'appareils dans des soins buccodentaires numériques
EP4634936A1 (fr) Apprentissage par renforcement pour configurations finales et organisation intermédiaire dans des aligneurs de plateaux transparents
WO2024127309A1 (fr) Autoencodeurs pour configurations finales et étapes intermédiaires d'aligneurs transparents
WO2024127311A1 (fr) Modèles d'apprentissage automatique pour génération de conception de restauration dentaire
US20250217663A1 (en) Defect Detection, Mesh Cleanup, and Mesh Cleanup Validation in Digital Dentistry
WO2024127316A1 (fr) Autocodeurs pour le traitement de représentations 3d dans des soins buccodentaires numériques
US20250364117A1 (en) Mesh Segmentation and Mesh Segmentation Validation In Digital Dentistry
US20250366959A1 (en) Geometry Generation for Dental Restoration Appliances, and the Validation of That Geometry
EP4634931A1 (fr) Calcul et visualisation de métriques dans des soins buccaux numériques
EP4634934A1 (fr) Apprentissage profond géométrique pour configurations finales et séquençage intermédiaire dans le domaine des aligneurs transparents
WO2024127304A1 (fr) Transformateurs pour configurations finales et stadification intermédiaire dans des aligneurs de plateaux transparents
US20250363269A1 (en) Fixture Model Validation for Aligners in Digital Orthodontics
WO2023242771A1 (fr) Validation de configurations de dents pour des aligneurs en orthodontie numérique
EP4633527A1 (fr) Techniques de transfert de pose pour des représentations de soins bucco-dentaires en 3d
WO2024127314A1 (fr) Imputation de valeurs de paramètres ou de valeurs métriques dans des soins buccaux numériques
WO2024127308A1 (fr) Classification de représentations 3d de soins bucco-dentaires
WO2025126117A1 (fr) Modèles d'apprentissage automatique pour la prédiction de structures de données se rapportant à une réduction interproximale
WO2025074322A1 (fr) Traitements de restauration orthodontique et dentaire combinés
WO2025257746A1 (fr) Formation de modèles d'apprentissage automatique pour des applications de soins bucco-dentaires numériques à l'aide d'une mise en correspondance de flux
WO2025126150A1 (fr) Débruitage de modèles probabilistes de diffusion pour prédiction d'anatomie post-traitement dans des soins buccaux numériques
WO2024127307A1 (fr) Comparaison de montages pour des montages finaux et la stadification intermédiaire de gouttières d'alignement transparentes
WO2024127310A1 (fr) Autocodeurs pour la validation de représentations de soins buccodentaires 3d

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23832829

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2025534723

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 202380085976.0

Country of ref document: CN

Ref document number: 2025534723

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 2023832829

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE

WWP Wipo information: published in national office

Ref document number: 202380085976.0

Country of ref document: CN

ENP Entry into the national phase

Ref document number: 2023832829

Country of ref document: EP

Effective date: 20250714

WWP Wipo information: published in national office

Ref document number: 2023832829

Country of ref document: EP