[go: up one dir, main page]

WO2025245365A1 - Navigation d'instrument médical sur la base d'une imagerie tomodensitométrique peropératoire en temps réel à l'aide d'un dispositif à rayons x classique - Google Patents

Navigation d'instrument médical sur la base d'une imagerie tomodensitométrique peropératoire en temps réel à l'aide d'un dispositif à rayons x classique

Info

Publication number
WO2025245365A1
WO2025245365A1 PCT/US2025/030620 US2025030620W WO2025245365A1 WO 2025245365 A1 WO2025245365 A1 WO 2025245365A1 US 2025030620 W US2025030620 W US 2025030620W WO 2025245365 A1 WO2025245365 A1 WO 2025245365A1
Authority
WO
WIPO (PCT)
Prior art keywords
volume
model
intraoperative
ray
images
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
PCT/US2025/030620
Other languages
English (en)
Inventor
Dorian Averbuch
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of WO2025245365A1 publication Critical patent/WO2025245365A1/fr
Pending legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Definitions

  • the present invention pertains to the field of medical imaging, specifically to a method and system for navigating medical instrument using real-time high-quality three-dimensional (3D) computed tomography (CT) imaging from 2D X-ray images acquired using conventional imaging equipment such as 2D X-Ray or C-Arm.
  • 3D three-dimensional
  • CT computed tomography
  • Three-dimensional (3D) imaging has become an essential tool in a multitude of medical intervention procedures such as spine, dental, orthopedic, and cardiovascular surgeries.
  • the ability to accurately visualize and navigate complex 3D anatomical structures from two- dimensional (2D) X-ray images is a common and critical challenge faced by physicians. Particularly in spinal surgeries, discerning intricate details like the spinal cord and pedicle pathways can be exceedingly difficult from 2D images alone.
  • This invention describes the navigation and tracking of medical instrument on the realtime high-quality 3D CT imaging continuously and iteratively generated from a 2D X-ray images obtained using standard, widely accessible imaging equipment like a simple X-Ray devices or C-Arms.
  • the proposed method stands to revolutionize the current practice by eliminating the need for both navigation systems based on registration of X-Ray with intraoperative or preoperative CT and, also, the use of CT scanners intraoperatively.
  • the benefits of such an innovation include reduced costs, simplified workflow, minimized radiation exposure, and the expansion of interventional procedures to outpatient settings, ultimately democratizing the availability of advanced medical procedures and broadening the scope of minimally invasive procedures.
  • the scope of this invention encompasses a novel method and system for navigation of medical instrument using high-definition 3D images acquired from standard 2D X-ray images with standard X-Ray devices without the need for registration between intraoperative 2D to 3D imaging.
  • This system is designed to overcome the need for 2D to 3D registration increasing the navigation accuracy, deficiencies of blurred or unclear views in Axial and Sagittal planes, which are critical for the real-time navigation and decision-making processes during surgeries.
  • this invention ensures that physicians can spatially localize their instruments during intervention procedures with great accuracy and high visual quality that facilitate better patient outcomes and allow for a broader range of procedures to be performed in various settings, including outpatient facilities.
  • One aspect of the invention provides a solution that transcends the limitations of existing art by enabling instrument navigation using clear spatial visualization on all anatomical planes, including Axial and Sagittal views, while the 3D image is reconstructed from 2D X-ray data.
  • the invention aims to equip physicians with the ability to interactwith a real-time 3D representation of the patient's anatomy using their common instruments during interventional procedures, thereby greatly enhancing the precision and safety of surgical and interventional procedures.
  • Another aspect of the invention introduces a method of Medical Instrument Navigation using real-time 3D CT images generated from a set of 2D X-ray images.
  • CT Computed Tomography
  • C-Arm A medical imaging device named for its C-shaped arm or Fluoroscope, used to connect the X-ray source and X-ray detector to one another, which allows for movement horizontally, vertically, and around the swivel axes, thus providing a variety of imaging angles.
  • Isocenter The fixed point in space around which the X-ray tube and image intensifier (detector) of a C-arm rotate during imaging.
  • Isocentering The technique of aligning the isocenter of the imaging equipment with the region of interest within the patient's body to maintain a consistent focal point throughout different imaging angles.
  • Artificial Intelligence A branch of computer science that aims to create systems capable of performing tasks that usually require human intelligence. This includes the ability to learn, reason, solve problems, perceive, and understand natural language.
  • Machine Learning A subset of artificial intelligence that involves the development of algorithms that can learn and make predictions or decisions based on data. Machine learning enables computers to identify patterns and learn from past experiences without being explicitly programmed.
  • Neural Network A computational model inspired by the way neural networks in the human brain process information. It consists of interconnected nodes, or neurons, that process input data and can adapt and learn by adjusting the connections (weights) between these nodes.
  • Deep Neural Network An advanced type of neural network that contains multiple layers of interconnected nodes. Each layer transforms the input data with the aim to extract increasingly higher-level features of the data fortasks such as image and speech recognition.
  • CNN A class of deep neural networks, most commonly applied to analyzing visual imagery.
  • CNNs are modeled on the organization of the animal visual cortex and are designed to automatically and adaptively learn spatial hierarchies of features from input images. They achieve this through a multilayered architecture composed of convolutional layers that filter inputs for useful information, pooling layers that reduce dimensionality, and fully connected layers that interpret the feature data to make predictions or classifications.
  • This structure enables CNNs to transform raw pixel data into incremental levels of abstract representation, making them highly efficient for tasks such as image recognition, object detection, and more.
  • Neural Radiance Fields A computational framework for rendering highly detailed 3D scenes from a set of 2D images. It uses a deep neural network to model the volumetric scene function, encoding both the color and density of points in 3D space as viewed from any angle. NeRF has been revolutionary in computer graphics and computer vision for its ability to produce photorealistic reconstructions and novel viewpoints of complex scenes with unprecedented detail and realism.
  • MedNeRF Medical Neural Radiance Fields
  • NeRF Neural Radiance Fields
  • Pose of the C-Arm The spatial configuration of the X-ray source in a C-Arm imaging system at a specific point in time, defined by its three-dimensional position and orientation relative to a fixed coordinate system.
  • the pose comprises six degrees of freedom: three translational components (x, y, z) describing the position of the source in space, and three rotational components (0x, 0y, 0z) describing its angular orientation about each principal axis.
  • Accurate estimation of the C-Arm source pose is critical for reconstructing volumetric (3D) images from multiple 2D fluoroscopic projections.
  • FIG. 1 is a flowchart of an embodiment of a method of the invention
  • Fig. 2 is a perspective view of a C-arm being isocentered around the area of interest without preoperative CT according to an embodiment of a method of the invention
  • Fig. 3 is an example of a multilayered system architecture based on residual DNN accordin to an embodiment of the invention.
  • the method 120 begins, at step 122, with placing a region of interest of the patient at the isocenter 110 of the C-arm 102 (isocentering) using application guidance on the intraoperative CT model to ensure the targeted area is at the focal point of the imaging process.
  • isocentering a C-Arm during surgical interventions is a technique aimed at precisely positioning imaging equipment to capture the necessary views of a patient's anatomy with minimal radiation exposure and time.
  • complex procedures such as those involving the spine or in trauma surgery, it is crucial to visualize anatomical structures relative to surgical tools and implants.
  • achieving the correct radiographic views often required a skilled operator to maneuver the C-Arm through various angles, which could be both timeconsuming and expose patients and staff to additional radiation.
  • Methods have been developed by others in attempts to address these challenges that utilize preoperative CT scans to assist in the C-Arm positioning process.
  • the innovative approach of current invention completely obviates the need for a preoperative CT scan. Rather, the method provides the ability to guide the isocenter process using an intraoperative CT model.
  • This approach involves using an adaptable generic anatomic model that can be created using at least one X-ray snapshot of the area of interest with known pose, or positions and orientation of the radiation source. Based on the at least one X-ray image, the anatomic model will be generated and automatically registered with the imaged object. Once established, the known image-based registration technique is used for isocentering.
  • Another innovative step of the current invention is the ability to change the isocenter virtually in the future within the field of view and the area of interest before the 3D tomographic reconstruction to achieve certain level of details and higher image quality of the anatomy reconstructed in the area of interest.
  • Step 124 Acquire At Least One X-ray Image
  • the X-ray 116 of the C-arm 102 is used to acquire at least one X-ray image of a known or roughly estimated pose that preferably includes six degrees of freedom (DOF) defining the angular and positional relationship of the imaging equipment relative to the imaged area or object.
  • the C-Arm captures at least one 2D X- ray image or, alternatively, a series of 2D X-ray images from multiple directions.
  • This task can be completed by taking a single snapshot, distinct snapshots from various, preferably predetermined angles, or through continuous radiation exposure as the C-Arm moves around the target area. For each captured image, the C-Arm's pose is roughly estimated or provided.
  • This information includes the position and orientation of the radiation source relative to the patient's body and, depending on reconstruction method, maybe pivotal for the accurate reconstruction of CT images. This data influences the transformation of two-dimensional images for integration into the 3D model generation process.
  • a standard C-Arm device does not inherently provide pose information; therefore, supplementary methods are needed. These may include external tracking systems, mechanical boards equipped with metal markers, algorithms for image registration, or the use of positional encoders attached to the C-Arm. External tracking may use optical, electromagnetic, or electromechanical sensors. Mechanical boards are designed to use 2D or 3D patterns of markers that can be automatically detected on X-ray images for further processing. These methods for pose calculation are well-documented in existing literature and prior art. The advantage of the present invention is that continuous instrument navigation and tracking becomes an integral part of the process.
  • Analog C-Arms which typically use image intensifiers, necessitate distortion calibration to address inherent pincushion or S-distortions.
  • Distortion calibration involves capturing images of a known grid or pattern to map distortions and correct new images, a critical step for accurately depicting the imaged area.
  • the calibration grid serves a dual purpose in medical imagingwith C-Arm fluoroscopy. It is used for distortion calibration and as an aid for C-Arm pose estimation.
  • the grid's known geometry and pattern provide reference points in fluoroscopic images that software algorithms can use to deduce the C-Arm's pose during imaging, streamlining the workflow by combining calibration and pose estimation into a single step.
  • the imaging system can align each 2D image with its spatial orientation, foundational for the Al-driven reconstruction process. These calculated poses allow for the precise merging of images to form a cohesive 3D visualization critical for both diagnostic and interventional use.
  • An innovation of this invention lies in determining the optimal number and orientation of 2D X-ray images required for Al-driven CT reconstruction, depending on the precision needs of each medical application when reconstructed CT image and 3D models will be used. This innovation is pivotal in enhancingthe efficiency of the imaging process by reducing the number of images and thus exposure to radiation, while ensuring the adequate quality of the reconstructed CT per specific medical application requirements.
  • Another alternative non-limiting solution involves generating limited angle tomography from a series of 2D X-ray images using a traditional approach and such is used as an input for the ML model in the following step for certain or all layers of the ML model.
  • the method 120 involves performing a 2D to 3D transformation on the image of step 124 using an ML Pipeline, which is and end-to-end system that handles the full flow of data from input to final output.
  • Figs. 3 and 4 each depict an ML Pipeline - 3000 and 4000, respectfully.
  • the process of transforming 2D X-ray data into 3D CT images in this invention involves a hierarchical ML model specially designed for CT imaging generated from a single 2D X-ray snapshot or from a small or limited number of 2D X-ray images or projection views.
  • This deep learning approach operates within an encoder-decoder framework, a representation-generation model that learns and applies the complex relationship between the 2D X-ray and 3D CT data.
  • One innovative aspect of the present invention lies in the multiple processing layers within the system as shown in Fig. 3 and Fig. 4.
  • the advantage of utilizing multiple layers is the ability to recognize different types of instances from the input image.
  • the non-limiting examples of such instances are static patient anatomy such as vertebrae or rib, blood vessels that are temporarily highlighted by the injected contrast and moving or static medical instruments within the body.
  • Each layer is methodically tailored to recognize, enhance, and reconstruct these different instances. This specialization is achieved through separate configuration and architecture of each layer, ensuring that they are finely tuned to the unique characteristics of the respective instances including their static or dynamic nature.
  • Some medical applications may use just a single layer of representation network while others may use multiple layers.
  • the architecture of each layer is detailed below.
  • Fig. 3 shows the progression of a multilayer system architecture based on Residual DNN.
  • Fig. 4 shows the progression of a multilayer system architecture based on MedNeRF.
  • a 2D X-ray image 300 or 400 is pre-processed at 310 or 410 respectively, to emphasize the instances of the input image and/or reduce noise.
  • the processed images then pass through a tailored ML model 320 or 420, respectively that allows an accurate 3D reconstruction of the instance. The details of these models are discussed in more detail below.
  • the instance representations 330 and 430 from Figs. 3 and 4, respectively, of each layer, specifically the 3D Models 332, 432 and the CTVolumes 334, 434, are merged using a novel model-weighing approach 350, 450 that allows the reconstruction of a high-resolution CT-like volume as well as a combined 3D view of the underlying instances.
  • the innovative aspects of the ML model 320, 420 include its ability to decipher higherdimensional information from multiple 2D projections, whereas a single X-ray image does not necessarily have the information required to reconstruct a 3D volume because it may be obscured to vital features.
  • the ML model 320, 420 overcomes this limitation by employing an iterative approach, where the output generated at each instance layer is refined with the inclusion of additional X-ray images.
  • the nonlimiti ng examples of the instance layer model may entail an encoder-decoder architecture, where the representation network serves as the encoder, compressing the input into a dense feature representation, where each layer of the generation network acts as the decoder for the correspondent layer of the representation network (Fig.
  • the layer-specific X-ray image preprocessing aims to highlight the relevant instances in the acquired images per the layer requirements at 310.
  • processing may include thresholding, low or high-pass filtering, histogram equalization, etc.
  • the representation network (CNN) 322 of the ML model receives processed images as inputs. It then undergoes multiple convolutional layers that extract the semantic features necessary for the subsequent reconstruction of the instance at 324.
  • the features identified in step 324 remain in 2D space.
  • the transformation module at 326, reshapes the features into a representation that is conducive to the instance generation phase.
  • the generation network 328 takes as input the transformed images from 326 and uses them to reconstruct a 3D model 332 of the instance at 330.
  • the non-limiting examples of 3D model representation could be voxel-based, solid, polynomial, NURBS, and others.
  • the generation network may use a series of 3D deconvolutional layers to upscale the feature maps and output a volume 334. It may also include elements of residual learning or skip connections to retain fine-grained details necessary for high-quality reconstruction.
  • Other examples of such generative models can be Generative Adversarial Networks (GANs) or Variational Autoencoders (VAEs).
  • the output of the generation network is a 3D model 332 and a CT volume 334 that progressively improves with each iteration.
  • the layer-specific X-ray image preprocessing aims to highlight the relevant instances in the acquired images per the layer requirements 410.
  • processing may include thresholding, low or high-pass filtering, histogram equalization, etc.
  • Positional encoding 422 is applied to enable the neural network to better capture high- frequency details by transforming the input coordinates (e.g., spatial positions) into a higherdimensional space. This is achieved by applying a set of sinusoidal functions with different frequencies to each input coordinate, thereby allowing the model to differentiate between positions more effectively with subsequent input of this information into MedNeRF DNN 424.
  • MedNeRF DNN 424 represents a continuous volumetric scene. This neural network predicts the color and density of points in 3D space, given their coordinates and viewing direction. This representation is particularly suited for medical images, where capturing the intricate details of anatomical structures is crucial.
  • a Structure Activation Function 426 is applied to emphasize anatomically relevant regions during the volumetric representation process. This function acts as a spatially aware gating mechanism that selectively enhances features within the neural network based on their anatomical salience, such as bone, soft tissue, or implants. By weighting internal feature activations using learned structural relevance cues, the network more effectively captures the fine-grained variations criticalfor medical reconstruction.
  • MedNeRF employs volume rendering techniques 428 to synthesize 2D projections (views) from the 3D neural representation. This involves casting rays through the volume and accumulating color and opacity values along each ray based on the neural network's predictions. This process is key for generating views from any desired perspective, not present in the original dataset.
  • the output 430 of the volume rendering 428 is a 3D model 432 and a CT volume 434 that progressively improve with each iteration.
  • the ML model 420 is trained using a large number of paired X-ray images and instance 3D models, as well as simulated data. Since the training of such a network entails a large number of X-ray images some may be simulated computationally.
  • each instance layer may have a different architecture and the generation of a Combined 3D Model 360, 460 and Combined CT volume 370, 470 from multiple 3D Models 332, 432 and CT volumes 334, 434 is achieved through a sophisticated process of Weighted 3D Model and CT Volume Composition 350/450 that involves a weighted averaging technique at the voxel level.
  • a conceptual explanation of the process applied for CT Volume is
  • a separate CT volume 334, 434 is generated for each layer using the ML model's multilayer structure.
  • Each volume provides a three-dimensional representation of specific instances.
  • the CT volumes 334, 434 are composed of numerous tiny units called voxels, analogous to pixels in a 2D image but extending into three dimensions. Each voxel in the volume contains data about the structure at that point in space.
  • a weight is assigned to each voxel in each volume.
  • the weights are determined based on the relevance and accuracy of the information that a particular voxel represents, which could be influenced by factors such as tissue density, the clarity of the feature, the confidence of the ML model 320, 420 in its reconstruction, movement, and the clinical significance of the anatomical structure.
  • the combined volume may be further refined to enhance the visibility of critical structures, and parts of medical instruments or to suppress artifacts. This involves additional post-processing steps such as smoothing, thresholding, or applying advanced imaging filters.
  • Combining separately generated 3D models within the same coordinate system for display in a single scene involves importing each model into a shared environment and positioning them relative to one another before rendering the composite scene. This process is managed by tools provided in 3D graphics software and standard graphic engines.
  • Steps 124 and 126 are repeated several times, each iteration being either a refinement of the previous iteration or a new transformation depending on the application needs.
  • Step 128 Display Instrument on CT Volume and 3D Model
  • Step 128 involves displaying a medical instrument being used to perform a medical procedure on the patient 104 on the CT Volume and 3D Model on the c-arm monitor 112 to facilitate user interaction with the 3D anatomy.
  • the ML Model will continue to update the 3D model progressively and continuously.
  • This visual format can be easily analyzed by medical professionals.
  • the software presents Axial, Sagittal, and Coronal views, Multiple Intensive projection images or a three-dimensional image that can be manipulated on-screen, allowing for rotation, zooming, and slicing through different planes to examine various angles and layers of the anatomy. This visualization depends on the specific medical application and aids clinicians in diagnosis, surgical planning, and patient education.
  • the tip of the instrument may be used by doctor as an interaction mean with CT Volume and 3D Model.

Landscapes

  • Apparatus For Radiation Diagnosis (AREA)
  • Image Processing (AREA)

Abstract

L'invention concerne un procédé de production de modèles d'imagerie médicale par tomodensitométrie 3D de haute qualité, en temps réel, à partir d'images radiographiques 2D acquises en provenance d'un équipement d'imagerie classique, tel que des dispositifs à rayons X ou à bras en C, déjà présents dans une salle opératoire, ce qui évite de devoir amener et stériliser un équipement supplémentaire actuellement utilisé pour la navigation d'image 3D.
PCT/US2025/030620 2024-05-22 2025-05-22 Navigation d'instrument médical sur la base d'une imagerie tomodensitométrique peropératoire en temps réel à l'aide d'un dispositif à rayons x classique Pending WO2025245365A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US202463650726P 2024-05-22 2024-05-22
US63/650,726 2024-05-22

Publications (1)

Publication Number Publication Date
WO2025245365A1 true WO2025245365A1 (fr) 2025-11-27

Family

ID=97796109

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2025/030620 Pending WO2025245365A1 (fr) 2024-05-22 2025-05-22 Navigation d'instrument médical sur la base d'une imagerie tomodensitométrique peropératoire en temps réel à l'aide d'un dispositif à rayons x classique

Country Status (1)

Country Link
WO (1) WO2025245365A1 (fr)

Similar Documents

Publication Publication Date Title
JP5348868B2 (ja) 医用システムの作動方法、医用システムおよびコンピュータ読み取り可能な媒体
US10650513B2 (en) Method and system for tomosynthesis imaging
RU2479038C2 (ru) Трехмерная реконструкция тела и контура тела
US20250308043A1 (en) Guidance during medical procedures
US8897514B2 (en) Imaging method for motion analysis
US8942455B2 (en) 2D/3D image registration method
US7302286B2 (en) Method and apparatus for the three-dimensional presentation of an examination region of a patient in the form of a 3D reconstruction image
JP2022517581A (ja) 冠動脈の動的ロードマップを提供するための方法およびシステム
US11900620B2 (en) Method and system for registering images containing anatomical structures
US20030181809A1 (en) 3D imaging for catheter interventions by use of 2D/3D image fusion
US20070270689A1 (en) Respiratory gated image fusion of computed tomography 3D images and live fluoroscopy images
US20080107312A1 (en) Device and Process for Multimodal Registration of Images
JP6967983B2 (ja) 画像処理装置、画像処理方法、及びプログラム
JP6349278B2 (ja) 放射線撮像装置、画像処理方法及びプログラム
CN111093505B (zh) 放射线拍摄装置以及图像处理方法
WO2025245365A1 (fr) Navigation d'instrument médical sur la base d'une imagerie tomodensitométrique peropératoire en temps réel à l'aide d'un dispositif à rayons x classique
US20230083134A1 (en) Generating a temporary image
Al-Shayea et al. An efficient approach to 3d image reconstruction
Schwerter et al. A novel approach for a 2D/3D image registration routine for medical tool navigation in minimally invasive vascular interventions
CN115089294B (zh) 介入手术导航的方法
WO2025231206A1 (fr) Systèmes et procédés d'imagerie médicale et de navigation
WO2025250065A1 (fr) Procédé mis en oeuvre par ordinateur pour assistance chirurgicale
JP2025508764A (ja) 患者の特定の身体部位に対する器具の位置決めを支援するための方法、コンピューティングデバイス、システム、およびコンピュータプログラム製品
CN119770168A (zh) 一种x射线透射检查的人体定位方法
JP2023173813A (ja) 画像処理装置、方法およびプログラム