WO2025000915A1 - Holographic three-dimensional medical data visualization method and system capable of achieving mid-air gesture interaction - Google Patents
Holographic three-dimensional medical data visualization method and system capable of achieving mid-air gesture interaction Download PDFInfo
- Publication number
- WO2025000915A1 WO2025000915A1 PCT/CN2023/137206 CN2023137206W WO2025000915A1 WO 2025000915 A1 WO2025000915 A1 WO 2025000915A1 CN 2023137206 W CN2023137206 W CN 2023137206W WO 2025000915 A1 WO2025000915 A1 WO 2025000915A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user
- mixed reality
- dimensional
- virtual image
- holographic
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
Definitions
- the embodiments of the present application relate to the field of human-computer interaction technology, and in particular to a holographic three-dimensional medical data visualization method and system capable of air gesture interaction.
- the medical interactive visualization system based on XR and equipped with HoloLen2 head-mounted display has demonstrated its potential to present high-quality 3D patient data.
- surgical simulation provides doctors with an immersive surgical training environment, allowing doctors to operate 3D objects in the environment, while enhancing doctors' perception and helping them make intuitive diagnoses, and effectively avoiding the risks that real surgery may bring.
- efficient preoperative volume rendering and real-time and accurate intraoperative interactive technology are the key to assisting doctors in surgical diagnosis.
- the embodiments of the present application provide a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air.
- a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air.
- users can move, rotate, scale and cut virtual tissues from any angle, manipulate virtual surgical instruments in real time and observe the patient's lesions, thereby realizing gesture interaction between users and virtual scenes in the air.
- an embodiment of the present application provides a holographic three-dimensional medical data visualization method that can interact with gestures in the air, comprising the following steps: first, obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the user wears a mixed reality display device; then, based on the user environment information and user motion information, a stereo interpolation volume rendering method is used to generate a virtual image; next, an optical model is used to project the virtual image into the user's field of view, and the virtual image and the real scene are fused to obtain a three-dimensional model rendered by mixed reality; finally, based on the three-dimensional model rendered by mixed reality, the user's gestures are captured by a gesture recognition method and gesture interaction is performed in the air, so as to realize the interaction between the user and the virtual scene.
- the mixed reality display device is a head-mounted mixed reality display device HoloLens2; and the optical model is an optical model in the head-mounted mixed reality display device HoloLens2.
- the mixed reality display device has a built-in depth camera, an inertial measurement unit, and multiple sensors to obtain user environment information and user motion information.
- the mixed reality display device is further provided with a recognition camera and a tracking camera;
- the recognition camera is used to detect and recognize specific images, objects or scenes in the real scene;
- the tracking camera is used to track the position and direction of the camera, and place the virtual image in the real world according to the output of the recognition camera to achieve virtual-reality registration.
- a virtual image and a real scene are fused to obtain a three-dimensional model of mixed reality rendering, including: generating the virtual image into a three-dimensional model of the virtual image through 3D modeling and rendering technology, and adjusting the properties of the virtual image in real time, and fusing the three-dimensional model of the virtual image with the real scene to obtain a three-dimensional model of mixed reality rendering; wherein the properties of the virtual image include the position, size and transparency of the virtual image.
- the three-dimensional model rendered by the mixed reality includes three-dimensional models at multiple different angles; wherein the three-dimensional models at different angles correspond to the user's observation angle.
- the user when performing air gesture interaction, uses gestures to move, rotate, scale and cut the virtual tissue in the three-dimensional model rendered by the mixed reality from any angle, manipulate virtual surgical instruments in real time and observe the patient's lesion condition.
- the method further includes: evaluating the three-dimensional model of mixed reality rendering.
- the three-dimensional model rendered by the mixed reality is evaluated, including: using medical images of three-dimensional CT original data, medical images of three-dimensional CT reconstructed data, and medical images of three-dimensional ultrasound original data as comparison data, and evaluating the efficiency and real-time performance of the three-dimensional model rendered by the mixed reality.
- the embodiments of the present application also provide a holographic three-dimensional medical data visualization system that can interact with gestures through the air, including a mixed reality display device and a gesture recognition system connected to the mixed reality display device for communication;
- the mixed reality display device is worn on the user's head;
- the mixed reality display device has a built-in depth camera, an inertial measurement unit and multiple sensors for obtaining user environment information and user motion information;
- the user environment information is image information of the real scene around the user;
- the mixed reality display device also generates a virtual image based on the user environment information and user motion information using a stereo interpolation volume rendering method; and uses an optical model to project the virtual image into the user's field of view, and fuses the virtual image and the real scene to obtain a three-dimensional model rendered by mixed reality;
- the gesture recognition system is used to capture the user's gestures through a gesture recognition method based on the three-dimensional model rendered by mixed reality and perform gesture interaction through the air, so as to realize interaction between the user and the virtual
- the embodiment of the present application provides a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air, and the method includes the following steps: first, obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the user wears a mixed reality display device; then, based on the user environment information and the user motion information, a stereo interpolation volume rendering method is used to generate a virtual image; next, an optical model is used to project the virtual image into the user's field of view, and the virtual image and the real scene are fused to obtain a three-dimensional model rendered by mixed reality; finally, based on the three-dimensional model rendered by mixed reality, the user's gestures are captured by a gesture recognition method and gesture interaction is performed in the air, so as to realize the interaction between the user and the virtual scene.
- this application proposes a holographic three-dimensional medical data intuitive visualization system under air gesture interaction suitable for surgical disciplines, covering preoperative planning and holographic three-dimensional visualization of tissue structure, using augmented reality technology to help surgeons predict tissue structure information and cut tissues and organs under the guidance of virtual information, combined with registration tracking technology and air gesture interaction, to assist surgeons in performing surgical operations naturally, thereby transforming traditional surgery into predictable and accurately guided surgery.
- FIG1 is a schematic diagram of a flow chart of a holographic three-dimensional medical data visualization method capable of air gesture interaction provided by an embodiment of the present application;
- FIG2 is a schematic diagram of image recognition and virtual-real registration based on Vuforia provided in one embodiment of the present application;
- FIG3 is a schematic diagram of the holographic gesture interaction principle provided in an embodiment of the present application.
- the volume rendering method in the prior art uses direct volume rendering technology to enhance visual and spatial perception from a three-dimensional data set.
- the existing methods have never fully achieved the same quality and performance as in medical augmented reality. Therefore, the methods proposed in the prior art have the technical problem of being limited to rendering polygonal surface data and unable to meet the high efficiency requirements of XR auxiliary diagnosis systems for volume rendering.
- XR can present three-dimensional anatomical structures more realistically and accurately during intraoperative diagnosis and preoperative surgical planning.
- volume rendering technology has made great progress in recent years.
- Direct volume rendering technology is used to enhance visual and spatial perception from three-dimensional data sets.
- a related technology proposes ClearView for non-segmented, real-time, focus and background visualization of three-dimensional data. ClearView draws the viewer's attention to the focus area, where important features of the three-dimensional data are embedded. The transparency of the surrounding environment layer is adjusted according to the curvature characteristics of the environment layer or the distance from the focus layer.
- immersive interactive interface can use people's stereoscopic vision to display data, shifting the data expression space from the two-dimensional plane to the three-dimensional space around the user; providing mobility, so that the user's physical working environment is no longer limited to a fixed office desk; providing natural interaction methods such as gestures, making the interaction more intuitive, allowing users to use multiple channels and interact with data in parallel.
- the embodiment of the present application provides a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air, and the method includes the following steps: first, obtaining user environment information and user motion information; the user environment information is the image information of the real scene around the user; the user wears a mixed reality display device; then, based on the user environment information and user motion information, a stereo interpolation volume rendering method is used to generate a virtual image; next, an optical model is used to project the virtual image into the user's field of view, and the virtual image and the real scene are merged to obtain a three-dimensional model rendered by mixed reality; finally, based on the three-dimensional model rendered by mixed reality, the user's gesture is captured by a gesture recognition method and gestures are interacted in the air to realize the interaction between the user and the virtual scene.
- the embodiment of the present application provides a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air, and through efficient three-dimensional object body rendering and real-time and accurate interaction method design, the user can move, rotate, scale and cut virtual tissues from any angle, manipulate virtual surgical instruments in real time and observe the patient's lesions, so as to realize the interaction between the user and the virtual scene through gestures in the air.
- an embodiment of the present application provides a holographic three-dimensional medical data visualization method capable of air gesture interaction, comprising:
- Step S1 obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the user wears a mixed reality display device.
- Step S2 Based on the user environment information and the user motion information, a stereo interpolation volume rendering method is used to generate a virtual image.
- Step S3 Use an optical model to project the virtual image into the user's field of view, and fuse the virtual image with the real scene to obtain a three-dimensional model rendered by mixed reality.
- Step S4 Based on the three-dimensional model rendered by mixed reality, the user's gestures are captured through a gesture recognition method and gesture interaction is performed in the air to enable interaction between the user and the virtual scene.
- the mixed reality display device is a head-mounted mixed reality display device HoloLens2; and the optical model is an optical model in the head-mounted mixed reality display device HoloLens2.
- This application proposes an interactive three-dimensional medical data holographic visualization system adapted for use with the HoloLens2 head-mounted display device.
- the system enables surgeons to move, rotate, scale and shear virtual tissues from any angle, manipulate virtual surgical instruments in real time and observe the patient's lesions, and diagnose the area of interest intuitively and accurately.
- the mixed reality display device has a built-in depth camera, an inertial measurement unit, and multiple sensors to obtain user environment information and user motion information.
- this application uses Microsoft's new generation of mixed reality head-mounted display device HoloLens2 to realize the mixed reality display of three-dimensional models.
- HoloLens2 Based on the technology of HoloLens2 to integrate virtual images with real scenes, its principle is mainly to use multiple sensors such as depth cameras, inertial measurement units and optical sensors to obtain image information of real scenes and user motion information, and then transmit this information to the computer for processing and analysis, and finally generate virtual images and project them into the user's field of view through the optical device of HoloLens2 to achieve the display effect of mixed reality.
- realizing the integration of virtual images and real scenes is the key to realizing mixed reality.
- HoloLens2 adopts a variety of technologies, including optical technology, sensor technology and computer graphics technology.
- optical technology is the basis of mixed reality display. It uses transparent waveguides to project virtual images into the user's field of view, and realizes the integration of virtual images and real scenes through optical principles such as reflection and refraction.
- HoloLens2 uses a variety of sensors, including depth cameras, inertial measurement units, and optical sensors. These sensors can obtain the user's position, posture, gesture and other information in real time, so as to achieve accurate tracking and positioning of the user, and then adjust the position and size of the virtual image to make it perfectly integrated with the real scene.
- Computer graphics are used to generate and display virtual images in the user's field of view.
- virtual images are generated into realistic three-dimensional models through 3D modeling and rendering technology, and their position, size, transparency and other properties are adjusted in real time to achieve integration with the real scene.
- the mixed reality display device is also provided with a recognition camera and a tracking camera;
- the recognition camera is used to detect and recognize specific images, objects or scenes in the real scene;
- the tracking camera is used to track the position and direction of the camera, and place the virtual image in the real world according to the output of the recognition camera to achieve virtual-reality registration.
- the Vuforia engine is used in the holographic three-dimensional medical data visualization method designed by this application that can interact with gestures in the air, as shown in Figure 2.
- Vuforia can use the built-in cameras (recognition camera and tracking camera) of the mixed reality device to locate specific patterns using image recognition and tracking functions.
- the basic architecture of Vuforia includes a recognition engine and a tracking engine.
- the recognition engine is used to detect and identify specific images, objects or scenes in the real world.
- the tracking engine is responsible for tracking the position and direction of the camera and accurately placing virtual content in the real world based on the output of the recognition engine. Through this method, this application can achieve accurate virtual-real registration.
- the virtual image and the real scene are fused to obtain a three-dimensional model of mixed reality rendering, including: generating the virtual image into a three-dimensional model of the virtual image through 3D modeling and rendering technology, and adjusting the properties of the virtual image in real time, and fusing the three-dimensional model of the virtual image with the real scene to obtain the three-dimensional model of mixed reality rendering; wherein the properties of the virtual image include the position, size and transparency of the virtual image.
- the three-dimensional model rendered by the mixed reality includes three-dimensional models at multiple different angles; wherein the three-dimensional models at different angles correspond to the user's observation angle.
- the holographic three-dimensional medical data visualization method capable of air gesture interaction provided in the embodiment of the present application, according to the above-mentioned illumination model, it is rendered by HoloLens 2 for mixed reality. After the model rendering is completed, the model display at different angles is provided based on the viewing angle and position of the wearer of HoloLens 2.
- the user when performing air gesture interaction, uses gestures to move, rotate, scale, and cut the virtual tissue in the three-dimensional model rendered by the mixed reality from any angle, manipulate virtual surgical instruments in real time, and observe the patient's lesion condition.
- the holographic three-dimensional medical data visualization method capable of air gesture interaction adopts a portable holographic visualization interaction based on gestures for human-computer interaction.
- the portable holographic visualization interaction technology based on gestures is a human-computer interaction method based on holographic projection and gesture recognition. It mainly realizes a natural and intuitive interaction between the user and the virtual scene by projecting the virtual three-dimensional scene into the real scene, and then uses gesture recognition technology to capture the user's gestures and realize interaction, thereby realizing more sophisticated and flexible operations.
- Interaction requires capturing the user's hand movements to identify gestures and converting them into corresponding commands to achieve control of the virtual scene.
- the accuracy, real-time and stability of gesture recognition can be improved by training and identifying a large amount of gesture data.
- the gesture recognition system uses devices such as cameras or depth cameras to capture the user's hand movements and convert them into digital signals. These signals are input into the computer and processed by algorithms and models to recognize different gestures. These gestures can represent different instructions, such as sliding, zooming, rotating, etc. The gesture recognition system undergoes a lot of training to recognize different gestures.
- Computer vision technology helps gesture recognition systems identify specific patterns and shapes in images, thereby accurately identifying different gestures. For example, it detects information such as the position of fingers, the direction, speed, and size of gestures, thereby helping the system determine the user's gestures.
- the interactive method of this application supports the movement and rotation of the entire model, and can also scale local key parts.
- users can cut using hand gestures, which enables surgeons to view various cross-sections of the target area, allowing them to observe and analyze the patient's lesions without blind spots.
- the method further includes: evaluating the three-dimensional model of mixed reality rendering.
- the three-dimensional model rendered by mixed reality is evaluated, including: using medical images of three-dimensional CT original data, medical images of three-dimensional CT reconstructed data, and medical images of three-dimensional ultrasound original data as comparison data, and evaluating the efficiency and real-time performance of the three-dimensional model rendered by mixed reality.
- This application uses two types of medical data, CT images and ultrasound images, to conduct experiments to test and evaluate the efficiency and real-time performance of the three-dimensional model rendered by mixed reality provided by this application.
- the experiment is conducted according to the following steps:
- this application designed an experiment to compare three typical medical images: 3D CT raw data, 3D CT reconstructed data and 3D ultrasound raw data, and the rendering results were obtained using six volume rendering algorithm models.
- the basic model no lighting
- illumination model back-to-front direct volume rendering model
- back-to-front direct volume rendering + cubic convolution interpolation model early ray termination model + cubic convolution interpolation model and early ray termination were used.
- the algorithm of this application has the highest contrast resolution, which can more clearly see the tiny blood vessels and structures on the heart, and can perform appropriate texture analysis in the same tissue with homogeneous features.
- the algorithm model of this application is more suitable for the evaluation of the myocardium.
- the sphere centered on the focus can focus the focus area, and the three-dimensional composite display of various tissue compositions highlights their boundaries and contours.
- this application conducted 20 interactive visualization experiments, recorded and calculated the average number of frames per second, to evaluate the real-time performance of the medical interactive visualization system of this application.
- the maximum displayable frame rate of HoloLens2 is 60FPS
- the basic model obtains the highest rendering speed of 59.95 ⁇ 2.25FPS, which is close to the highest resolution that HoloLens 2 can display, followed by the early ray termination model (method of this application) 40.00 ⁇ 5.80FPS, but from the perspective of visual authenticity restoration, the basic model cannot provide an effective spatial reference, and this application is significantly better than the basic model.
- the model of this application obtains the highest rendering speed, which is 60.00 ⁇ 1.80FPS and 42.05 ⁇ 5.55FPS, respectively, and in ultrasound images, the rendering speed reaches the highest display frame rate of HoloLens2. Therefore, compared with some other five representative methods, the method of the present application can achieve the best rendering speed and has better real-time performance.
- the medical interactive visualization system provided by the present application can quickly respond to changes in the posture of three-dimensional data.
- doctors can better perceive the spatial structure of the patient's target area, and develop a holographic 3D medical data visualization system that is interactive with gestures in the air and suitable for surgical medicine. It covers preoperative planning and holographic 3D visualization of tissue structure.
- Augmented reality technology is used to help surgeons perceive tissue structure information and cut tissue organs under the guidance of virtual information. Combined with registration tracking technology and gesture interaction in the air, it assists surgeons to perform surgical operations naturally and intuitively, thereby transforming traditional surgery into predictable and accurately guided surgery.
- a suitable medical visualization prototype system was developed, and related interaction methods were formulated. The feasibility of the system was verified using 3D CT raw data, 3D CT reconstruction data, and 3D ultrasound raw data. On this basis, in-depth research and exploration were carried out to determine its practical application value in clinical practice, and continuous improvement and optimization were carried out, laying a solid foundation for its future.
- the present invention uses air gesture interaction and three-dimensional medical data visualization technology to assist doctors in achieving intuitive diagnosis.
- the method includes two strategies: stereo interpolation volume rendering based on the illumination model and holographic mid-air gesture interaction.
- our method can achieve the fastest rendering speed while ensuring the visualization effect.
- the interaction algorithm can meet the real-time and accuracy requirements of assisting doctors in diagnosis. Therefore, our method can provide users with a more intuitive three-dimensional medical data visualization effect in a holographic XR environment.
- the embodiment of the present application also provides a holographic three-dimensional medical data visualization system that can interact with gestures through the air, including a mixed reality display device and a gesture recognition system connected to the mixed reality display device for communication; the mixed reality display device is worn on the user's head; the mixed reality display device has a built-in depth camera, an inertial measurement unit and multiple sensors for obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the mixed reality display device also generates a virtual image based on the user environment information and user motion information using a stereo interpolation volume rendering method; and uses an optical model to project the virtual image into the user's field of view, and fuses the virtual image and the real scene to obtain a three-dimensional model rendered by mixed reality; the gesture recognition system is used to capture the user's gestures through a gesture recognition method based on the three-dimensional model rendered by mixed reality and perform gesture interaction through the air, so as to realize interaction between the user and the virtual scene.
- a mixed reality display device is
- the advantage of the holographic three-dimensional medical data visualization system with air gesture interaction is that: although there are many medical visualization systems at this stage, most of them are based on two-dimensional screen displays, and the holographic display of medical images is still very limited, and multi-perspective interactive surgical operations cannot be achieved. Surgeons use mice to operate on the screen, which greatly limits the doctor's field of vision compared to the mid-air gesture interaction used in our system, and this interaction method is far from the natural method in the real surgical scene, limiting the surgeon's workspace for medical image analysis, and there is a risk of potential diagnostic errors caused by the surgeon's visual impairment of two-dimensional images.
- the embodiment of the present application provides a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air, and the method includes the following steps: first, obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the user wears a mixed reality display device; then, based on the user environment information and user motion information, a stereo interpolation volume rendering method is used to generate a virtual image; next, an optical model is used to project the virtual image into the user's field of view, and the virtual image and the real scene are fused to obtain a three-dimensional model rendered by mixed reality; finally, based on the three-dimensional model rendered by mixed reality, the user's gestures are captured by a gesture recognition method and gesture interaction is performed in the air, so as to realize the interaction between the user and the virtual scene.
- this application proposes a holographic three-dimensional medical data intuitive visualization system under air gesture interaction suitable for surgical disciplines, covering preoperative planning and holographic three-dimensional visualization of tissue structure, using augmented reality technology to help surgeons predict tissue structure information and cut tissues and organs under the guidance of virtual information, combined with registration tracking technology and air gesture interaction, to assist surgeons in performing surgical operations naturally, thereby transforming traditional surgery into predictable and accurately guided surgery.
- the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), disk or optical disk and other media that can store program codes.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computer Graphics (AREA)
- Processing Or Creating Images (AREA)
- Image Generation (AREA)
Abstract
Description
本申请实施例涉及人机交互技术领域,特别涉及一种可隔空手势交互的全息三维医学数据可视化方法及系统。The embodiments of the present application relate to the field of human-computer interaction technology, and in particular to a holographic three-dimensional medical data visualization method and system capable of air gesture interaction.
随着数字医疗的不断发展与革新,扩展现实逐渐成为辅助外科医生进行临床诊疗的关键技术。得益于医学和计算机科学领域间的学科交融,医学影像如计算机断层扫描(Computed Tomography,CT)、磁共振成像(Magnetic Resonance Imaging,MRI)和超声检查(Ultrasound,US)得到的多模态医学数据,经过计算机渲染和重建得到的模型,使得医学数据的三维可视化成为可能,为扩展现实(Extended Reality,XR)辅助医疗的发展提供了广阔的空间。With the continuous development and innovation of digital medicine, extended reality has gradually become a key technology to assist surgeons in clinical diagnosis and treatment. Thanks to the fusion of disciplines between medicine and computer science, multimodal medical data obtained from medical images such as computed tomography (CT), magnetic resonance imaging (MRI) and ultrasound (US), and models obtained through computer rendering and reconstruction, make three-dimensional visualization of medical data possible, providing a broad space for the development of extended reality (XR) assisted medicine.
基于XR并配备HoloLen2头戴式显示器的医疗交互可视化系统已经展现了它呈现高质量3D患者数据的潜力,在三维可视化的基础上,手术模拟是为医生提供了一种沉浸式的手术培训环境,允许医生在环境中操作3D物体,在增强医生观感的同时能够帮助医生进行直观的诊断,并有效避免真实手术可能带来的风险。对于全息XR医疗可视化系统而言,高效的术前体绘制以及术中实时精准的交互技术是实现辅助医生进行手术诊断的关键。The medical interactive visualization system based on XR and equipped with HoloLen2 head-mounted display has demonstrated its potential to present high-quality 3D patient data. Based on 3D visualization, surgical simulation provides doctors with an immersive surgical training environment, allowing doctors to operate 3D objects in the environment, while enhancing doctors' perception and helping them make intuitive diagnoses, and effectively avoiding the risks that real surgery may bring. For holographic XR medical visualization systems, efficient preoperative volume rendering and real-time and accurate intraoperative interactive technology are the key to assisting doctors in surgical diagnosis.
与传统医学数据的二维显示技术不同的是,XR 可以在术中诊断过程和术前手术规划中更真实、更准确的展现三维解剖结构。随着GPU技术的发展,体绘制技术近些年取得了巨大的进步。体绘制技术是利用直接体绘制技术从三维数据集中增强视觉和空间感知。在使用GPU进行体绘制的过程中,现有方法从来没有完全实现过与在医疗增强现实中相同的质量和性能。提出的方法仅限于渲染多边形表面数据,虽然相关的解剖和结构可以被描绘出来,但是这些方法过于耗时,无法满足XR辅助诊断系统对于体绘制的高效性需求。Unlike traditional two-dimensional display technology for medical data, XR can present three-dimensional anatomical structures more realistically and accurately during intraoperative diagnosis and preoperative surgical planning. With the development of GPU technology, volume rendering technology has made great progress in recent years. Volume rendering technology uses direct volume rendering technology to enhance visual and spatial perception from three-dimensional data sets. In the process of volume rendering using GPU, existing methods have never fully achieved the same quality and performance as in medical augmented reality. The proposed methods are limited to rendering polygonal surface data. Although the relevant anatomy and structures can be depicted, these methods are too time-consuming and cannot meet the efficiency requirements of XR-assisted diagnosis systems for volume rendering.
本申请实施例提供一种可隔空手势交互的全息三维医学数据可视化方法及系统,通过高效的三维物体体绘制以及实时精准的交互方法设计,使用户能够从任何角度移动、旋转、放缩和剪切虚拟组织,实时操纵虚拟手术器械并观察患者的病变情况,实现用户与虚拟场景的隔空手势交互。The embodiments of the present application provide a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air. Through efficient three-dimensional object rendering and real-time and accurate interaction method design, users can move, rotate, scale and cut virtual tissues from any angle, manipulate virtual surgical instruments in real time and observe the patient's lesions, thereby realizing gesture interaction between users and virtual scenes in the air.
为解决上述技术问题,第一方面,本申请实施例提供一种可隔空手势交互的全息三维医学数据可视化方法,包括以下步骤:首先,获取用户环境信息和用户运动信息;用户环境信息为用户周围的现实场景的图像信息;用户穿戴有混合现实显示设备;然后,基于用户环境信息和用户运动信息,采用立体插值体绘制方法,生成虚拟图像;接下来,采用光学模型将虚拟图像投射至用户的视野中,并对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型;最后,基于混合现实渲染的三维模型,通过手势识别方法捕捉用户的手势并进行隔空手势交互,实现用户与虚拟场景的交互。To solve the above technical problems, in the first aspect, an embodiment of the present application provides a holographic three-dimensional medical data visualization method that can interact with gestures in the air, comprising the following steps: first, obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the user wears a mixed reality display device; then, based on the user environment information and user motion information, a stereo interpolation volume rendering method is used to generate a virtual image; next, an optical model is used to project the virtual image into the user's field of view, and the virtual image and the real scene are fused to obtain a three-dimensional model rendered by mixed reality; finally, based on the three-dimensional model rendered by mixed reality, the user's gestures are captured by a gesture recognition method and gesture interaction is performed in the air, so as to realize the interaction between the user and the virtual scene.
在一些示例性实施例中,混合现实显示设备为头戴式混合现实显示设备HoloLens2;光学模型为头戴式混合现实显示设备HoloLens2中的光学模型。In some exemplary embodiments, the mixed reality display device is a head-mounted mixed reality display device HoloLens2; and the optical model is an optical model in the head-mounted mixed reality display device HoloLens2.
在一些示例性实施例中,混合现实显示设备内置深度摄像头、惯性测量单元和多个传感器,以获取用户环境信息和用户运动信息。In some exemplary embodiments, the mixed reality display device has a built-in depth camera, an inertial measurement unit, and multiple sensors to obtain user environment information and user motion information.
在一些示例性实施例中,混合现实显示设备内还设有识别摄像头和跟踪摄像头;识别摄像头用于检测和识别现实场景中特定的图像、物体或场景;跟踪摄像头用于跟踪相机的位置和方向,并根据识别摄像头的输出将虚拟图像放置在现实世界中,实现虚实配准。In some exemplary embodiments, the mixed reality display device is further provided with a recognition camera and a tracking camera; the recognition camera is used to detect and recognize specific images, objects or scenes in the real scene; the tracking camera is used to track the position and direction of the camera, and place the virtual image in the real world according to the output of the recognition camera to achieve virtual-reality registration.
在一些示例性实施例中,对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型,包括:通过3D 建模和渲染技术将所述虚拟图像生成为虚拟图像的三维模型,并实时调整所述虚拟图像的属性,将所述虚拟图像的三维模型与现实场景进行融合,得到混合现实渲染的三维模型;其中,所述虚拟图像的属性包括所述虚拟图像的位置、大小以及透明度。In some exemplary embodiments, a virtual image and a real scene are fused to obtain a three-dimensional model of mixed reality rendering, including: generating the virtual image into a three-dimensional model of the virtual image through 3D modeling and rendering technology, and adjusting the properties of the virtual image in real time, and fusing the three-dimensional model of the virtual image with the real scene to obtain a three-dimensional model of mixed reality rendering; wherein the properties of the virtual image include the position, size and transparency of the virtual image.
在一些示例性实施例中,混合现实渲染的三维模型包括多种不同角度的三维模型;其中,不同角度的三维模型与用户的观察角度相对应。In some exemplary embodiments, the three-dimensional model rendered by the mixed reality includes three-dimensional models at multiple different angles; wherein the three-dimensional models at different angles correspond to the user's observation angle.
在一些示例性实施例中,在进行隔空手势交互时,用户通过手势从任意角度移动、旋转、放缩和剪切所述混合现实渲染的三维模型中的虚拟组织,实时操纵虚拟手术器械并观察患者的病变情况。In some exemplary embodiments, when performing air gesture interaction, the user uses gestures to move, rotate, scale and cut the virtual tissue in the three-dimensional model rendered by the mixed reality from any angle, manipulate virtual surgical instruments in real time and observe the patient's lesion condition.
在一些示例性实施例中,在得到混合现实渲染的三维模型之后,还包括:对混合现实渲染的三维模型进行评估。In some exemplary embodiments, after obtaining the three-dimensional model of mixed reality rendering, the method further includes: evaluating the three-dimensional model of mixed reality rendering.
在一些示例性实施例中,对所述混合现实渲染的三维模型进行评估,包括:以三维CT原始数据的医学图像、三维CT重建数据的医学图像、三维超声原始数据的医学图像为对比数据,对所述混合现实渲染的三维模型的高效性和实时性分别进行评估。In some exemplary embodiments, the three-dimensional model rendered by the mixed reality is evaluated, including: using medical images of three-dimensional CT original data, medical images of three-dimensional CT reconstructed data, and medical images of three-dimensional ultrasound original data as comparison data, and evaluating the efficiency and real-time performance of the three-dimensional model rendered by the mixed reality.
第二方面,本申请实施例还提供了一种可隔空手势交互的全息三维医学数据可视化系统,包括混合现实显示设备以及与其通信连接的手势识别系统;混合现实显示设备穿戴于用户的头部;混合现实显示设备内置深度摄像头、惯性测量单元和多个传感器,用于获取用户环境信息和用户运动信息;用户环境信息为用户周围的现实场景的图像信息;混合现实显示设备还根据用户环境信息和用户运动信息,采用立体插值体绘制方法,生成虚拟图像;以及采用光学模型将虚拟图像投射至用户的视野中,并对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型;手势识别系统用于根据混合现实渲染的三维模型,通过手势识别方法捕捉用户的手势并进行隔空手势交互,实现用户与虚拟场景的交互。On the second aspect, the embodiments of the present application also provide a holographic three-dimensional medical data visualization system that can interact with gestures through the air, including a mixed reality display device and a gesture recognition system connected to the mixed reality display device for communication; the mixed reality display device is worn on the user's head; the mixed reality display device has a built-in depth camera, an inertial measurement unit and multiple sensors for obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the mixed reality display device also generates a virtual image based on the user environment information and user motion information using a stereo interpolation volume rendering method; and uses an optical model to project the virtual image into the user's field of view, and fuses the virtual image and the real scene to obtain a three-dimensional model rendered by mixed reality; the gesture recognition system is used to capture the user's gestures through a gesture recognition method based on the three-dimensional model rendered by mixed reality and perform gesture interaction through the air, so as to realize interaction between the user and the virtual scene.
本申请实施例提供的技术方案至少具有以下优点:The technical solution provided by the embodiments of the present application has at least the following advantages:
本申请实施例提供一种可隔空手势交互的全息三维医学数据可视化方法及系统,该方法包括以下步骤:首先,获取用户环境信息和用户运动信息;用户环境信息为用户周围的现实场景的图像信息;用户穿戴有混合现实显示设备;然后,基于用户环境信息和用户运动信息,采用立体插值体绘制方法,生成虚拟图像;接下来,采用光学模型将虚拟图像投射至用户的视野中,并对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型;最后,基于混合现实渲染的三维模型,通过手势识别方法捕捉用户的手势并进行隔空手势交互,实现用户与虚拟场景的交互。The embodiment of the present application provides a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air, and the method includes the following steps: first, obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the user wears a mixed reality display device; then, based on the user environment information and the user motion information, a stereo interpolation volume rendering method is used to generate a virtual image; next, an optical model is used to project the virtual image into the user's field of view, and the virtual image and the real scene are fused to obtain a three-dimensional model rendered by mixed reality; finally, based on the three-dimensional model rendered by mixed reality, the user's gestures are captured by a gesture recognition method and gesture interaction is performed in the air, so as to realize the interaction between the user and the virtual scene.
为了让医生更好的感知患者靶区空间结构,本申请提出一种适用于外科学科的隔空手势交互下全息三维医学数据直观可视化系统,覆盖术前规划、组织结构全息三维可视化,利用增强现实技术帮助外科医生预知组织结构信息,并在虚拟信息的引导下剪切组织器官,结合配准追踪技术和隔空手势交互,协助术者自然地进行手术操作,从而将传统手术转化成为可预期和准确指引的手术。In order to enable doctors to better perceive the spatial structure of the patient's target area, this application proposes a holographic three-dimensional medical data intuitive visualization system under air gesture interaction suitable for surgical disciplines, covering preoperative planning and holographic three-dimensional visualization of tissue structure, using augmented reality technology to help surgeons predict tissue structure information and cut tissues and organs under the guidance of virtual information, combined with registration tracking technology and air gesture interaction, to assist surgeons in performing surgical operations naturally, thereby transforming traditional surgery into predictable and accurately guided surgery.
一个或多个实施例通过与之对应的附图中的图片进行示例性说明,这些示例性说明并不构成对实施例的限定,除非有特别申明,附图中的图不构成比例限制。One or more embodiments are exemplarily described by pictures in the corresponding drawings, and these exemplifications do not constitute limitations on the embodiments. Unless otherwise stated, the pictures in the drawings do not constitute proportional limitations.
图1为本申请一实施例提供的一种可隔空手势交互的全息三维医学数据可视化方法的流程示意图;FIG1 is a schematic diagram of a flow chart of a holographic three-dimensional medical data visualization method capable of air gesture interaction provided by an embodiment of the present application;
图2为本申请一实施例提供的基于Vuforia的图像识别和虚实配准示意图;FIG2 is a schematic diagram of image recognition and virtual-real registration based on Vuforia provided in one embodiment of the present application;
图3为本申请一实施例提供的全息手势交互原理示意图。FIG3 is a schematic diagram of the holographic gesture interaction principle provided in an embodiment of the present application.
由背景技术可知,现有技术中体绘制方法是利用直接体绘制技术从三维数据集中增强视觉和空间感知。在使用GPU进行体绘制的过程中,现有方法从来没有完全实现过与在医疗增强现实中相同的质量和性能。因此,现有技术提出的方法存在着仅限于渲染多边形表面数据,无法满足XR辅助诊断系统对于体绘制的高效性需求的技术问题。As can be seen from the background technology, the volume rendering method in the prior art uses direct volume rendering technology to enhance visual and spatial perception from a three-dimensional data set. In the process of using GPU for volume rendering, the existing methods have never fully achieved the same quality and performance as in medical augmented reality. Therefore, the methods proposed in the prior art have the technical problem of being limited to rendering polygonal surface data and unable to meet the high efficiency requirements of XR auxiliary diagnosis systems for volume rendering.
与传统医学数据的二维显示技术不同的是,XR 可以在术中诊断过程和术前手术规划中更真实、更准确的展现三维解剖结构。随着GPU技术的发展,体绘制技术近些年取得了巨大的进步。利用直接体绘制技术从三维数据集中增强视觉和空间感知。一相关技术提出了ClearView,用于三维数据的无分割、实时、焦点和背景可视化。ClearView将观看者的注意力吸引到焦点区域,三维数据的重要特征被嵌入到其中。周围环境层的透明度根据环境层的曲率特性或与焦点层的距离进行调节。Unlike traditional two-dimensional display technology for medical data, XR can present three-dimensional anatomical structures more realistically and accurately during intraoperative diagnosis and preoperative surgical planning. With the development of GPU technology, volume rendering technology has made great progress in recent years. Direct volume rendering technology is used to enhance visual and spatial perception from three-dimensional data sets. A related technology proposes ClearView for non-segmented, real-time, focus and background visualization of three-dimensional data. ClearView draws the viewer's attention to the focus area, where important features of the three-dimensional data are embedded. The transparency of the surrounding environment layer is adjusted according to the curvature characteristics of the environment layer or the distance from the focus layer.
大多数早期的交互技术都是从设备为中心的角度来设计的,比如使用二维触摸屏输入以及触控笔、键盘和设备传感器来与虚拟物体进行交互。还有一相关技术对沉浸式的可视化进行了全面的分析后提出:沉浸式交互界面能够利用人们的立体视觉展现数据,将数据表达空间从二维平面转向用户周围的三维空间;提供可移动性,从而让用户的物理工作环境不再限于固定的办公桌面;提供如手势等自然的交互方式,使交互更加直观,让用户利用多通道然而并行地与数据交互。Most early interactive technologies were designed from a device-centric perspective, such as using two-dimensional touch screen input and stylus, keyboard and device sensors to interact with virtual objects. Another related technology made a comprehensive analysis of immersive visualization and proposed: immersive interactive interface can use people's stereoscopic vision to display data, shifting the data expression space from the two-dimensional plane to the three-dimensional space around the user; providing mobility, so that the user's physical working environment is no longer limited to a fixed office desk; providing natural interaction methods such as gestures, making the interaction more intuitive, allowing users to use multiple channels and interact with data in parallel.
由于现有方法从来没有完全实现过与在医疗增强现实中相同的质量和性能,提出的方法仅限于渲染多边形表面数据,虽然相关的解剖和结构可以被描绘出来,但是这些方法过于耗时,无法满足XR辅助诊断系统对于体绘制的高效性需求。通过对相关学者和案例的分析研究,医疗辅助诊断系统应具有高效性,并尽可能提高渲染速度。Since existing methods have never fully achieved the same quality and performance as in medical augmented reality, the proposed methods are limited to rendering polygonal surface data. Although the relevant anatomy and structures can be depicted, these methods are too time-consuming and cannot meet the high efficiency requirements of XR-assisted diagnosis systems for volume rendering. Through the analysis and research of relevant scholars and cases, the medical-assisted diagnosis system should be efficient and the rendering speed should be as fast as possible.
为解决上述技术问题,本申请实施例提供一种可隔空手势交互的全息三维医学数据可视化方法及系统,该方法包括以下步骤:首先,获取用户环境信息和用户运动信息;用户环境信息为用户周围的现实场景的图像信息;用户穿戴有混合现实显示设备;然后,基于用户环境信息和用户运动信息,采用立体插值体绘制方法,生成虚拟图像;接下来,采用光学模型将虚拟图像投射至用户的视野中,并对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型;最后,基于混合现实渲染的三维模型,通过手势识别方法捕捉用户的手势并进行隔空手势交互,实现用户与虚拟场景的交互。本申请实施例通过提供一种可隔空手势交互的全息三维医学数据可视化方法及系统,通过高效的三维物体体绘制以及实时精准的交互方法设计,使用户能够从任何角度移动、旋转、放缩和剪切虚拟组织,实时操纵虚拟手术器械并观察患者的病变情况,实现用户与虚拟场景的隔空手势交互。In order to solve the above technical problems, the embodiment of the present application provides a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air, and the method includes the following steps: first, obtaining user environment information and user motion information; the user environment information is the image information of the real scene around the user; the user wears a mixed reality display device; then, based on the user environment information and user motion information, a stereo interpolation volume rendering method is used to generate a virtual image; next, an optical model is used to project the virtual image into the user's field of view, and the virtual image and the real scene are merged to obtain a three-dimensional model rendered by mixed reality; finally, based on the three-dimensional model rendered by mixed reality, the user's gesture is captured by a gesture recognition method and gestures are interacted in the air to realize the interaction between the user and the virtual scene. The embodiment of the present application provides a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air, and through efficient three-dimensional object body rendering and real-time and accurate interaction method design, the user can move, rotate, scale and cut virtual tissues from any angle, manipulate virtual surgical instruments in real time and observe the patient's lesions, so as to realize the interaction between the user and the virtual scene through gestures in the air.
下面将结合附图对本申请的各实施例进行详细的阐述。然而,本领域的普通技术人员可以理解,在本申请各实施例中,为了使读者更好地理解本申请而提出了许多技术细节。但是,即使没有这些技术细节和基于以下各实施例的种种变化和修改,也可以实现本申请所要求保护的技术方案。The following will describe the various embodiments of the present application in detail with reference to the accompanying drawings. However, it will be appreciated by those skilled in the art that in the various embodiments of the present application, many technical details are provided in order to enable the reader to better understand the present application. However, even without these technical details and various changes and modifications based on the following embodiments, the technical solution claimed in the present application can be implemented.
参看图1,本申请实施例提供了一种可隔空手势交互的全息三维医学数据可视化方法,包括:Referring to FIG. 1 , an embodiment of the present application provides a holographic three-dimensional medical data visualization method capable of air gesture interaction, comprising:
步骤S1、获取用户环境信息和用户运动信息;用户环境信息为用户周围的现实场景的图像信息;用户穿戴有混合现实显示设备。Step S1, obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the user wears a mixed reality display device.
步骤S2、基于用户环境信息和用户运动信息,采用立体插值体绘制方法,生成虚拟图像。Step S2: Based on the user environment information and the user motion information, a stereo interpolation volume rendering method is used to generate a virtual image.
步骤S3、采用光学模型将虚拟图像投射至用户的视野中,并对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型。Step S3: Use an optical model to project the virtual image into the user's field of view, and fuse the virtual image with the real scene to obtain a three-dimensional model rendered by mixed reality.
步骤S4、基于混合现实渲染的三维模型,通过手势识别方法捕捉用户的手势并进行隔空手势交互,实现用户与虚拟场景的交互。Step S4: Based on the three-dimensional model rendered by mixed reality, the user's gestures are captured through a gesture recognition method and gesture interaction is performed in the air to enable interaction between the user and the virtual scene.
在一些实施例中,混合现实显示设备为头戴式混合现实显示设备HoloLens2;光学模型为头戴式混合现实显示设备HoloLens2中的光学模型。In some embodiments, the mixed reality display device is a head-mounted mixed reality display device HoloLens2; and the optical model is an optical model in the head-mounted mixed reality display device HoloLens2.
本申请提出了一种适配HoloLens2头戴式显示设备使用的交互式三维医学数据全息可视化系统。通过高效的三维物体体绘制以及实时精准的交互方法设计,该系统使外科医生能够从任何角度移动、旋转、放缩和剪切虚拟组织,实时操纵虚拟手术器械并观察患者的病变情况,直观准确地对感兴趣区域进行诊断。通过在HoloLens2上进行了基于三维CT原始数据、三维CT重建数据和三维超声原始数据的实验,与先进的体绘制算法模型相比,本申请的方法渲染效果与速度均具有一定的优越性,在HoloLens2分别达到42.00±5.80 FPS、60.00±1.80 FPS和42.05±5.55 FPS的渲染速度,验证了本申请的系统的可行性。This application proposes an interactive three-dimensional medical data holographic visualization system adapted for use with the HoloLens2 head-mounted display device. Through efficient three-dimensional object volume rendering and real-time and accurate interactive method design, the system enables surgeons to move, rotate, scale and shear virtual tissues from any angle, manipulate virtual surgical instruments in real time and observe the patient's lesions, and diagnose the area of interest intuitively and accurately. Through experiments based on three-dimensional CT raw data, three-dimensional CT reconstruction data and three-dimensional ultrasound raw data on HoloLens2, compared with advanced volume rendering algorithm models, the rendering effect and speed of the method of this application have certain advantages, reaching rendering speeds of 42.00±5.80 FPS, 60.00±1.80 FPS and 42.05±5.55 FPS on HoloLens2, respectively, verifying the feasibility of the system of this application.
在一些实施例中,混合现实显示设备内置深度摄像头、惯性测量单元和多个传感器,以获取用户环境信息和用户运动信息。In some embodiments, the mixed reality display device has a built-in depth camera, an inertial measurement unit, and multiple sensors to obtain user environment information and user motion information.
具体的,本申请采用微软公司的新一代混合现实头戴式显示设备HoloLens2来实现三维模型的混合现实显示。基于HoloLens2将虚拟图像与现实场景相融合的技术,其原理主要是利用深度摄像头、惯性测量单元和光学传感器等多个传感器获取现实场景的图像信息和用户的运动信息,然后将这些信息传输到计算机中进行处理和分析,最终生成虚拟图像并将其通过HoloLens2的光学设备投射到用户的视野中,实现混合现实的显示效果。在HoloLens2的混合现实显示中,实现虚拟图像与现实场景相融合是实现混合现实的关键。为了实现这一目标,HoloLens2采用了多种技术,包括光学技术、传感器技术和计算机图形学技术等。其中,光学技术是混合现实显示的基础,它利用透明的波导将虚拟图像投射到用户的视野中,并通过反射和折射等光学原理来实现虚拟图像与现实场景的融合。Specifically, this application uses Microsoft's new generation of mixed reality head-mounted display device HoloLens2 to realize the mixed reality display of three-dimensional models. Based on the technology of HoloLens2 to integrate virtual images with real scenes, its principle is mainly to use multiple sensors such as depth cameras, inertial measurement units and optical sensors to obtain image information of real scenes and user motion information, and then transmit this information to the computer for processing and analysis, and finally generate virtual images and project them into the user's field of view through the optical device of HoloLens2 to achieve the display effect of mixed reality. In the mixed reality display of HoloLens2, realizing the integration of virtual images and real scenes is the key to realizing mixed reality. In order to achieve this goal, HoloLens2 adopts a variety of technologies, including optical technology, sensor technology and computer graphics technology. Among them, optical technology is the basis of mixed reality display. It uses transparent waveguides to project virtual images into the user's field of view, and realizes the integration of virtual images and real scenes through optical principles such as reflection and refraction.
传感器技术感知用户周围的环境和用户的运动信息,并将这些信息传输到计算机中进行处理和分析。HoloLens2采用了多种传感器,包括深度摄像头、惯性测量单元和光学传感器等,这些传感器可以实时获取用户的位置、姿态、手势等信息,从而实现对用户的精确跟踪和定位,进而调整虚拟图像的位置和大小,使其与现实场景完美融合。利用计算机图形学将虚拟图像生成并显示到用户的视野中。在HoloLens2中,通过3D 建模和渲染技术将虚拟图像生成为逼真的三维模型,并实时调整其位置、大小、透明度等属性,从而实现与现实场景的融合。Sensor technology senses the environment around the user and the user's motion information, and transmits this information to the computer for processing and analysis. HoloLens2 uses a variety of sensors, including depth cameras, inertial measurement units, and optical sensors. These sensors can obtain the user's position, posture, gesture and other information in real time, so as to achieve accurate tracking and positioning of the user, and then adjust the position and size of the virtual image to make it perfectly integrated with the real scene. Computer graphics are used to generate and display virtual images in the user's field of view. In HoloLens2, virtual images are generated into realistic three-dimensional models through 3D modeling and rendering technology, and their position, size, transparency and other properties are adjusted in real time to achieve integration with the real scene.
在一些实施例中,混合现实显示设备内还设有识别摄像头和跟踪摄像头;识别摄像头用于检测和识别现实场景中特定的图像、物体或场景;跟踪摄像头用于跟踪相机的位置和方向,并根据识别摄像头的输出将虚拟图像放置在现实世界中,实现虚实配准。In some embodiments, the mixed reality display device is also provided with a recognition camera and a tracking camera; the recognition camera is used to detect and recognize specific images, objects or scenes in the real scene; the tracking camera is used to track the position and direction of the camera, and place the virtual image in the real world according to the output of the recognition camera to achieve virtual-reality registration.
本申请设计的可隔空手势交互的全息三维医学数据可视化方法中,采用了Vuforia引擎,如图2所示。Vuforia可以通过混合现实设备的内置摄像头(识别摄像头和跟踪摄像头),使用图像识别和跟踪功能,对特定的图案进行定位。Vuforia 的基本架构包括一个识别引擎和一个跟踪引擎。识别引擎用于在现实世界中检测和识别特定的图像、物体或场景。跟踪引擎则负责跟踪相机的位置和方向,并根据识别引擎的输出将虚拟内容准确地放置在现实世界中。通过这个方法,本申请可以实现精确的虚实配准。The Vuforia engine is used in the holographic three-dimensional medical data visualization method designed by this application that can interact with gestures in the air, as shown in Figure 2. Vuforia can use the built-in cameras (recognition camera and tracking camera) of the mixed reality device to locate specific patterns using image recognition and tracking functions. The basic architecture of Vuforia includes a recognition engine and a tracking engine. The recognition engine is used to detect and identify specific images, objects or scenes in the real world. The tracking engine is responsible for tracking the position and direction of the camera and accurately placing virtual content in the real world based on the output of the recognition engine. Through this method, this application can achieve accurate virtual-real registration.
在一些实施例中,对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型,包括:通过3D 建模和渲染技术将所述虚拟图像生成为虚拟图像的三维模型,并实时调整所述虚拟图像的属性,将所述虚拟图像的三维模型与现实场景进行融合,得到混合现实渲染的三维模型;其中,所述虚拟图像的属性包括所述虚拟图像的位置、大小以及透明度。In some embodiments, the virtual image and the real scene are fused to obtain a three-dimensional model of mixed reality rendering, including: generating the virtual image into a three-dimensional model of the virtual image through 3D modeling and rendering technology, and adjusting the properties of the virtual image in real time, and fusing the three-dimensional model of the virtual image with the real scene to obtain the three-dimensional model of mixed reality rendering; wherein the properties of the virtual image include the position, size and transparency of the virtual image.
在一些实施例中,混合现实渲染的三维模型包括多种不同角度的三维模型;其中,不同角度的三维模型与用户的观察角度相对应。In some embodiments, the three-dimensional model rendered by the mixed reality includes three-dimensional models at multiple different angles; wherein the three-dimensional models at different angles correspond to the user's observation angle.
具体的,在本申请实施例提供的可隔空手势交互的全息三维医学数据可视化方法中,根据上述的光照模型,将其用HoloLens2进行混合现实渲染。在模型渲染完毕后,将基于HoloLens2的佩戴者的观察角度和位置,提供不同角度的模型显示。Specifically, in the holographic three-dimensional medical data visualization method capable of air gesture interaction provided in the embodiment of the present application, according to the above-mentioned illumination model, it is rendered by HoloLens 2 for mixed reality. After the model rendering is completed, the model display at different angles is provided based on the viewing angle and position of the wearer of HoloLens 2.
在一些实施例中,在进行隔空手势交互时,用户通过手势从任意角度移动、旋转、放缩和剪切所述混合现实渲染的三维模型中的虚拟组织,实时操纵虚拟手术器械并观察患者的病变情况。In some embodiments, when performing air gesture interaction, the user uses gestures to move, rotate, scale, and cut the virtual tissue in the three-dimensional model rendered by the mixed reality from any angle, manipulate virtual surgical instruments in real time, and observe the patient's lesion condition.
具体的,本申请提供的可隔空手势交互的全息三维医学数据可视化方法,人机交互采用基于手势的便携全息可视化交互,基于手势的便携全息可视化交互技术是一种基于全息投影和手势识别的人机交互方式。它主要通过将虚拟三维场景投影到现实场景中,再通过手势识别技术来捕捉用户的手势并实现交互,从而实现用户与虚拟场景的自然、直观的交互方式,可以实现更加精细和灵活的操作。Specifically, the holographic three-dimensional medical data visualization method capable of air gesture interaction provided by the present application adopts a portable holographic visualization interaction based on gestures for human-computer interaction. The portable holographic visualization interaction technology based on gestures is a human-computer interaction method based on holographic projection and gesture recognition. It mainly realizes a natural and intuitive interaction between the user and the virtual scene by projecting the virtual three-dimensional scene into the real scene, and then uses gesture recognition technology to capture the user's gestures and realize interaction, thereby realizing more sophisticated and flexible operations.
交互需要通过捕捉用户的手部动作来识别手势,并将其转化为相应的指令,从而实现对虚拟场景的控制。通过对大量手势数据的训练和识别,来提高手势识别的准确性、实时性和稳定性。Interaction requires capturing the user's hand movements to identify gestures and converting them into corresponding commands to achieve control of the virtual scene. The accuracy, real-time and stability of gesture recognition can be improved by training and identifying a large amount of gesture data.
手势交互方法的原理如图2所示。手势识别系统会使用摄像头或深度相机等设备来捕捉用户的手部动作,并将其转化为数字信号。这些信号被输入到计算机中,通过算法和模型进行处理,从而识别出不同的手势。这些手势可以代表不同的指令,例如滑动、缩放、旋转等。手势识别系统进行大量的训练,以对不同的手势进行识别。The principle of gesture interaction is shown in Figure 2. The gesture recognition system uses devices such as cameras or depth cameras to capture the user's hand movements and convert them into digital signals. These signals are input into the computer and processed by algorithms and models to recognize different gestures. These gestures can represent different instructions, such as sliding, zooming, rotating, etc. The gesture recognition system undergoes a lot of training to recognize different gestures.
计算机视觉技术帮助手势识别系统识别图像中的特定模式和形状,从而准确地识别不同的手势。例如,检测手指的位置、手势的方向、速度和大小等信息,从而帮助系统判断用户的手势。本申请的交互方法支持整个模型的移动和旋转,同时也可以对局部的关键部位进行放缩。Computer vision technology helps gesture recognition systems identify specific patterns and shapes in images, thereby accurately identifying different gestures. For example, it detects information such as the position of fingers, the direction, speed, and size of gestures, thereby helping the system determine the user's gestures. The interactive method of this application supports the movement and rotation of the entire model, and can also scale local key parts.
此外,用户可以通过手势进行剪切,这使得外科医生能够查看目标区域的各种横截面,从而无死角地观察并分析病人的病变。Additionally, users can cut using hand gestures, which enables surgeons to view various cross-sections of the target area, allowing them to observe and analyze the patient's lesions without blind spots.
在一些实施例中,在得到混合现实渲染的三维模型之后,还包括:对混合现实渲染的三维模型进行评估。In some embodiments, after obtaining the three-dimensional model of mixed reality rendering, the method further includes: evaluating the three-dimensional model of mixed reality rendering.
在一些实施例中,对混合现实渲染的三维模型进行评估,包括:以三维CT原始数据的医学图像、三维CT重建数据的医学图像、三维超声原始数据的医学图像为对比数据,对混合现实渲染的三维模型的高效性和实时性分别进行评估。In some embodiments, the three-dimensional model rendered by mixed reality is evaluated, including: using medical images of three-dimensional CT original data, medical images of three-dimensional CT reconstructed data, and medical images of three-dimensional ultrasound original data as comparison data, and evaluating the efficiency and real-time performance of the three-dimensional model rendered by mixed reality.
本申请使用CT图像和超声图像两种类型的医学数据进行实验,以对本申请提供的混合现实渲染的三维模型的高效性和实时性分别进行测试和评估。对于每一个医学数据的按照如下步骤进行实验:This application uses two types of medical data, CT images and ultrasound images, to conduct experiments to test and evaluate the efficiency and real-time performance of the three-dimensional model rendered by mixed reality provided by this application. For each type of medical data, the experiment is conducted according to the following steps:
(1)获得尺寸为(270,512,512)的三维CT原始数据、512×512切片分辨率和 0.7-0.8mm 的切片厚度的三维CT重建数据或分辨率为183×115×126 FPS 的三维超声原始数据。(1) Obtain 3D CT raw data with a size of (270, 512, 512), 3D CT reconstructed data with a slice resolution of 512×512 and a slice thickness of 0.7-0.8 mm, or 3D ultrasound raw data with a resolution of 183×115×126 FPS.
(2)调整转移函数,强调和分类数据中感兴趣的特征,以真实还原骨、血管或软组织等区域的材质。(2) Adjust the transfer function to emphasize and classify the features of interest in the data to truly restore the material of areas such as bones, blood vessels, or soft tissues.
(3)外科医生佩戴HoloLens2在进行医学三维数据的交互。(3) Surgeons wearing HoloLens2 interact with medical 3D data.
下面分别对医学交互式可视化方法的高效性和实时性进行测试评估。The efficiency and real-time performance of the medical interactive visualization method are tested and evaluated below.
为了验证医学交互式可视化方法的高效性,本申请设计实验比较三种典型的医学图像:三维CT原始数据、三维CT重建数据和三维超声原始数据,使用六种体绘制算法模型的渲染得到结果。其中使用了基础模型(无光照)、照明模型、从后到前的直接体绘制模型、从后到前的直接体绘制+三次卷积插值模型、早期射线终止模型+三次卷积插值模型和早期射线终止。本申请的算法具有最高的对比度分辨率,可以更加清晰地看到心脏上细小的血管与结构,可在具有同质特征的相同组织中进行适当的纹理分析。本申请的算法模型更加适合于心肌的评估,以焦点为中心的球面可以聚焦焦点区域,三维复合显示各种组织构成,突出其边界和轮廓。In order to verify the efficiency of the medical interactive visualization method, this application designed an experiment to compare three typical medical images: 3D CT raw data, 3D CT reconstructed data and 3D ultrasound raw data, and the rendering results were obtained using six volume rendering algorithm models. The basic model (no lighting), illumination model, back-to-front direct volume rendering model, back-to-front direct volume rendering + cubic convolution interpolation model, early ray termination model + cubic convolution interpolation model and early ray termination were used. The algorithm of this application has the highest contrast resolution, which can more clearly see the tiny blood vessels and structures on the heart, and can perform appropriate texture analysis in the same tissue with homogeneous features. The algorithm model of this application is more suitable for the evaluation of the myocardium. The sphere centered on the focus can focus the focus area, and the three-dimensional composite display of various tissue compositions highlights their boundaries and contours.
为了验证医学交互式可视化方法的实时性,性本申请进行了20次交互式可视化交互实验,记录并计算了平均的每秒帧数,来评估本申请医学交互式可视化系统的实时性。在HoloLens2的最高可显示帧率为60FPS的前提条件下,对于三维CT原始数据,基础模型得到最高的渲染速度59.95±2.25FPS,且近乎HoloLens 2可显示的最高分辨率,其次是早期射线终止模型(本申请的方法)40.00±5.80FPS,但从可视化真实性还原的角度,基础模型无法提供有效的空间参考,本申请方面要明显优于基础模型。对于三维超声原始数据和三维CT重建数据,本申请的模型得到最高的渲染速度,分别为60.00±1.80FPS和42.05±5.55FPS,并且在超声图像中,渲染速度达到了HoloLens2的最高显示帧率。因此,与一些其他五种代表性的方法相比,本申请的方法可以达到最好的渲染的速度,具有更好的实时性,本申请提供的医学交互式可视化系统能够快速响应三维数据的位姿变化。In order to verify the real-time performance of the medical interactive visualization method, this application conducted 20 interactive visualization experiments, recorded and calculated the average number of frames per second, to evaluate the real-time performance of the medical interactive visualization system of this application. Under the premise that the maximum displayable frame rate of HoloLens2 is 60FPS, for the three-dimensional CT raw data, the basic model obtains the highest rendering speed of 59.95±2.25FPS, which is close to the highest resolution that HoloLens 2 can display, followed by the early ray termination model (method of this application) 40.00±5.80FPS, but from the perspective of visual authenticity restoration, the basic model cannot provide an effective spatial reference, and this application is significantly better than the basic model. For three-dimensional ultrasound raw data and three-dimensional CT reconstruction data, the model of this application obtains the highest rendering speed, which is 60.00±1.80FPS and 42.05±5.55FPS, respectively, and in ultrasound images, the rendering speed reaches the highest display frame rate of HoloLens2. Therefore, compared with some other five representative methods, the method of the present application can achieve the best rendering speed and has better real-time performance. The medical interactive visualization system provided by the present application can quickly respond to changes in the posture of three-dimensional data.
综上所述,让医生更好的感知患者靶区空间结构,研发出适用于外科医学的隔空手势交互式的全息三维医学数据可视化系统,覆盖术前规划、组织结构全息三维可视化,利用增强现实技术帮助外科医生感知组织结构信息,并在虚拟信息的引导下剪切组织器官,结合配准追踪技术和隔空手势交互,协助术者自然直观地进行手术操作,从而将传统手术转化成为可预期和准确指引的手术。根据外科医学的特点研发出适用的医学可视化原型系统,并制定了与之相关的交互方式,利用三维CT原始数据、三维CT重建数据和三维超声原始数据验证了系统的可行性。在该基础上,进行了深入的研究和探索,以确定其在临床上的实际应用价值,并不断进行改进和优化,为其未来奠定了坚实的基础。In summary, doctors can better perceive the spatial structure of the patient's target area, and develop a holographic 3D medical data visualization system that is interactive with gestures in the air and suitable for surgical medicine. It covers preoperative planning and holographic 3D visualization of tissue structure. Augmented reality technology is used to help surgeons perceive tissue structure information and cut tissue organs under the guidance of virtual information. Combined with registration tracking technology and gesture interaction in the air, it assists surgeons to perform surgical operations naturally and intuitively, thereby transforming traditional surgery into predictable and accurately guided surgery. According to the characteristics of surgical medicine, a suitable medical visualization prototype system was developed, and related interaction methods were formulated. The feasibility of the system was verified using 3D CT raw data, 3D CT reconstruction data, and 3D ultrasound raw data. On this basis, in-depth research and exploration were carried out to determine its practical application value in clinical practice, and continuous improvement and optimization were carried out, laying a solid foundation for its future.
本发明利用隔空手势交互与三维医学数据可视化技术辅助医生实现直观诊断,该方法包括两个策略:基于光照模型的立体插值体绘制和全息半空手势交互。对于每个策略我们分别进行了多次实验,以分别验证它们的有效性。我们使用多种光照模型和渲染方式对三种类型的医学图像进行体绘制,与一些有代表性的方法相比,我们的方法在保证可视化效果的同时可以达到最快的渲染速度,同时交互算法能满足辅助医生诊断的实时性和精准度要求。因此,我们的方法可以给用户提供全息XR环境下更直观的三维医学数据可视化效果。The present invention uses air gesture interaction and three-dimensional medical data visualization technology to assist doctors in achieving intuitive diagnosis. The method includes two strategies: stereo interpolation volume rendering based on the illumination model and holographic mid-air gesture interaction. We conducted multiple experiments for each strategy to verify their effectiveness. We used a variety of illumination models and rendering methods to perform volume rendering on three types of medical images. Compared with some representative methods, our method can achieve the fastest rendering speed while ensuring the visualization effect. At the same time, the interaction algorithm can meet the real-time and accuracy requirements of assisting doctors in diagnosis. Therefore, our method can provide users with a more intuitive three-dimensional medical data visualization effect in a holographic XR environment.
本申请实施例还提供了一种可隔空手势交互的全息三维医学数据可视化系统,包括混合现实显示设备以及与其通信连接的手势识别系统;混合现实显示设备穿戴于用户的头部;混合现实显示设备内置深度摄像头、惯性测量单元和多个传感器,用于获取用户环境信息和用户运动信息;用户环境信息为用户周围的现实场景的图像信息;混合现实显示设备还根据用户环境信息和用户运动信息,采用立体插值体绘制方法,生成虚拟图像;以及采用光学模型将虚拟图像投射至用户的视野中,并对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型;手势识别系统用于根据混合现实渲染的三维模型,通过手势识别方法捕捉用户的手势并进行隔空手势交互,实现用户与虚拟场景的交互。The embodiment of the present application also provides a holographic three-dimensional medical data visualization system that can interact with gestures through the air, including a mixed reality display device and a gesture recognition system connected to the mixed reality display device for communication; the mixed reality display device is worn on the user's head; the mixed reality display device has a built-in depth camera, an inertial measurement unit and multiple sensors for obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the mixed reality display device also generates a virtual image based on the user environment information and user motion information using a stereo interpolation volume rendering method; and uses an optical model to project the virtual image into the user's field of view, and fuses the virtual image and the real scene to obtain a three-dimensional model rendered by mixed reality; the gesture recognition system is used to capture the user's gestures through a gesture recognition method based on the three-dimensional model rendered by mixed reality and perform gesture interaction through the air, so as to realize interaction between the user and the virtual scene.
与现有技术相比,本申请提供的可隔空手势交互的全息三维医学数据可视化系统的优势在于:虽然现阶段有很多医学可视化的系统,由于大多是基于二维的屏幕显示,对医学图像的全息展示仍有很大有局限,无法做到多视角的交互式手术操作。外科医生使用鼠标在屏幕上进行操作,与我们系统中使用的半空手势交互相比,极大地限制了医生的视野范围,并且这种交互方式与现实手术场景中的自然方式相去甚远,局限了外科医生进行医学图像分析的工作空间,以及有外科医生因二维图像的视觉障碍引起的潜在诊断失误的风险。与此相反,在我们的系统中,佩戴了HoloLens2的外科医生可以自由移动,同时查看虚拟-物理融合场景并执行手术任务。我们的系统能够提高医生诊断时的舒适度进而优化患者治疗效果。外科医生可以将医学知识直观地呈现给患者,帮助患者更好地理解和接受治疗,从而减轻患者的心理压力和焦虑情绪。Compared with the prior art, the advantage of the holographic three-dimensional medical data visualization system with air gesture interaction provided by the present application is that: although there are many medical visualization systems at this stage, most of them are based on two-dimensional screen displays, and the holographic display of medical images is still very limited, and multi-perspective interactive surgical operations cannot be achieved. Surgeons use mice to operate on the screen, which greatly limits the doctor's field of vision compared to the mid-air gesture interaction used in our system, and this interaction method is far from the natural method in the real surgical scene, limiting the surgeon's workspace for medical image analysis, and there is a risk of potential diagnostic errors caused by the surgeon's visual impairment of two-dimensional images. In contrast, in our system, surgeons wearing HoloLens2 can move freely, view virtual-physical fusion scenes and perform surgical tasks at the same time. Our system can improve the doctor's comfort during diagnosis and optimize the patient's treatment effect. Surgeons can present medical knowledge to patients intuitively, helping patients better understand and accept treatment, thereby reducing patients' psychological pressure and anxiety.
由以上技术方案,本申请实施例提供一种可隔空手势交互的全息三维医学数据可视化方法及系统,该方法包括以下步骤:首先,获取用户环境信息和用户运动信息;用户环境信息为用户周围的现实场景的图像信息;用户穿戴有混合现实显示设备;然后,基于用户环境信息和用户运动信息,采用立体插值体绘制方法,生成虚拟图像;接下来,采用光学模型将虚拟图像投射至用户的视野中,并对虚拟图像和现实场景进行融合,得到混合现实渲染的三维模型;最后,基于混合现实渲染的三维模型,通过手势识别方法捕捉用户的手势并进行隔空手势交互,实现用户与虚拟场景的交互。Based on the above technical scheme, the embodiment of the present application provides a holographic three-dimensional medical data visualization method and system that can interact with gestures in the air, and the method includes the following steps: first, obtaining user environment information and user motion information; the user environment information is image information of the real scene around the user; the user wears a mixed reality display device; then, based on the user environment information and user motion information, a stereo interpolation volume rendering method is used to generate a virtual image; next, an optical model is used to project the virtual image into the user's field of view, and the virtual image and the real scene are fused to obtain a three-dimensional model rendered by mixed reality; finally, based on the three-dimensional model rendered by mixed reality, the user's gestures are captured by a gesture recognition method and gesture interaction is performed in the air, so as to realize the interaction between the user and the virtual scene.
为了让医生更好的感知患者靶区空间结构,本申请提出一种适用于外科学科的隔空手势交互下全息三维医学数据直观可视化系统,覆盖术前规划、组织结构全息三维可视化,利用增强现实技术帮助外科医生预知组织结构信息,并在虚拟信息的引导下剪切组织器官,结合配准追踪技术和隔空手势交互,协助术者自然地进行手术操作,从而将传统手术转化成为可预期和准确指引的手术。In order to enable doctors to better perceive the spatial structure of the patient's target area, this application proposes a holographic three-dimensional medical data intuitive visualization system under air gesture interaction suitable for surgical disciplines, covering preoperative planning and holographic three-dimensional visualization of tissue structure, using augmented reality technology to help surgeons predict tissue structure information and cut tissues and organs under the guidance of virtual information, combined with registration tracking technology and air gesture interaction, to assist surgeons in performing surgical operations naturally, thereby transforming traditional surgery into predictable and accurately guided surgery.
即,本领域技术人员可以理解,实现上述实施例方法中的全部或部分步骤是可以通过程序来指令相关的硬件来完成,该程序存储在一个存储介质中,包括若干指令用以使得一个设备(可以是单片机,芯片等)或处理器(processor)执行本申请各个实施例上述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。That is, those skilled in the art can understand that all or part of the steps in the above-mentioned embodiment method can be completed by instructing the relevant hardware through a program, and the program is stored in a storage medium, including a number of instructions to enable a device (which can be a single-chip microcomputer, chip, etc.) or a processor to execute all or part of the steps of the above-mentioned method of each embodiment of the present application. The aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), disk or optical disk and other media that can store program codes.
本领域的普通技术人员可以理解,上述各实施方式是实现本申请的具体实施例,而在实际应用中,可以在形式上和细节上对其作各种改变,而不偏离本申请的精神和范围。任何本领域技术人员,在不脱离本申请的精神和范围内,均可作各自更动与修改,因此本申请的保护范围应当以权利要求限定的范围为准。Those skilled in the art can understand that the above-mentioned embodiments are specific examples for implementing the present application, and in practical applications, various changes can be made to them in form and detail without departing from the spirit and scope of the present application. Any person skilled in the art can make their own changes and modifications without departing from the spirit and scope of the present application, so the scope of protection of the present application shall be based on the scope defined in the claims.
Claims (10)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310794023.3A CN116909442A (en) | 2023-06-29 | 2023-06-29 | Holographic three-dimensional medical data visualization method and system capable of isolating gesture interaction |
| CN202310794023.3 | 2023-06-29 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2025000915A1 true WO2025000915A1 (en) | 2025-01-02 |
Family
ID=88355567
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2023/137206 Pending WO2025000915A1 (en) | 2023-06-29 | 2023-12-07 | Holographic three-dimensional medical data visualization method and system capable of achieving mid-air gesture interaction |
Country Status (2)
| Country | Link |
|---|---|
| CN (1) | CN116909442A (en) |
| WO (1) | WO2025000915A1 (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119414970A (en) * | 2025-01-06 | 2025-02-11 | 湖南晟医智能科技有限公司 | A MR virtual image visualization method and system based on AI intelligence |
| CN119533396A (en) * | 2025-01-20 | 2025-02-28 | 济南大学 | Space distance measurement method and device based on mixed reality |
Families Citing this family (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN116909442A (en) * | 2023-06-29 | 2023-10-20 | 深圳先进技术研究院 | Holographic three-dimensional medical data visualization method and system capable of isolating gesture interaction |
| CN119229705B (en) * | 2024-11-28 | 2025-04-04 | 湖南晟医智能科技有限公司 | Simulation medical picture visualization system and method based on MR technology |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN114612640A (en) * | 2022-03-24 | 2022-06-10 | 航天宏图信息技术股份有限公司 | Space-based situation simulation system based on mixed reality technology |
| CN115157261A (en) * | 2022-07-27 | 2022-10-11 | 清华大学深圳国际研究生院 | Human-computer interaction device and method for teleoperation of flexible manipulator based on mixed reality |
| CN115185368A (en) * | 2022-06-28 | 2022-10-14 | 西安电子科技大学广州研究院 | Mobile robot interactive operation system based on Hololens |
| CN116909442A (en) * | 2023-06-29 | 2023-10-20 | 深圳先进技术研究院 | Holographic three-dimensional medical data visualization method and system capable of isolating gesture interaction |
-
2023
- 2023-06-29 CN CN202310794023.3A patent/CN116909442A/en active Pending
- 2023-12-07 WO PCT/CN2023/137206 patent/WO2025000915A1/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN114612640A (en) * | 2022-03-24 | 2022-06-10 | 航天宏图信息技术股份有限公司 | Space-based situation simulation system based on mixed reality technology |
| CN115185368A (en) * | 2022-06-28 | 2022-10-14 | 西安电子科技大学广州研究院 | Mobile robot interactive operation system based on Hololens |
| CN115157261A (en) * | 2022-07-27 | 2022-10-11 | 清华大学深圳国际研究生院 | Human-computer interaction device and method for teleoperation of flexible manipulator based on mixed reality |
| CN116909442A (en) * | 2023-06-29 | 2023-10-20 | 深圳先进技术研究院 | Holographic three-dimensional medical data visualization method and system capable of isolating gesture interaction |
Non-Patent Citations (1)
| Title |
|---|
| "Master's Thesis", 1 April 2020, HARBIN UNIVERSITY OF SCIENCE AND TECHNOLOGY, China, article LIANG DEXIAN: "Human-machine Interactive Control of Seed Implantation Robot for Prostate Based on Mixed Reality", pages: 1 - 116, XP093254162 * |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119414970A (en) * | 2025-01-06 | 2025-02-11 | 湖南晟医智能科技有限公司 | A MR virtual image visualization method and system based on AI intelligence |
| CN119533396A (en) * | 2025-01-20 | 2025-02-28 | 济南大学 | Space distance measurement method and device based on mixed reality |
Also Published As
| Publication number | Publication date |
|---|---|
| CN116909442A (en) | 2023-10-20 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| WO2025000915A1 (en) | Holographic three-dimensional medical data visualization method and system capable of achieving mid-air gesture interaction | |
| JP5992448B2 (en) | Image system and method | |
| CN113302660B (en) | Methods for visualizing dynamic anatomical structures | |
| Sielhorst et al. | Advanced medical displays: A literature review of augmented reality | |
| JP7624975B2 (en) | Method, computer program, user interface, and system for analyzing medical image data in virtual multi-user collaboration | |
| JP2022507622A (en) | Use of optical cords in augmented reality displays | |
| Jiang et al. | Registration technology of augmented reality in oral medicine: A review | |
| CN107296650A (en) | Intelligent operation accessory system based on virtual reality and augmented reality | |
| JP2014512550A6 (en) | Image system and method | |
| CN101797182A (en) | Nasal endoscope minimally invasive operation navigating system based on augmented reality technique | |
| CN112618026A (en) | Remote operation data fusion interactive display system and method | |
| TWI697317B (en) | Digital image reality alignment kit and method applied to mixed reality system for surgical navigation | |
| EP3803540B1 (en) | Gesture control of medical displays | |
| Kutter et al. | Real-time volume rendering for high quality visualization in augmented reality | |
| Vogt | Real-Time Augmented Reality for Image-Guided Interventions | |
| CN104274247A (en) | Medical surgical navigation method | |
| Krapichler et al. | VR interaction techniques for medical imaging applications | |
| Abou El-Seoud et al. | An interactive mixed reality ray tracing rendering mobile application of medical data in minimally invasive surgeries | |
| Huang et al. | Augmented reality-based autostereoscopic surgical visualization system for telesurgery | |
| JP7504942B2 (en) | Representation device for displaying a graphical representation of an augmented reality - Patent Application 20070123633 | |
| US10854005B2 (en) | Visualization of ultrasound images in physical space | |
| Zhang et al. | From AR to AI: augmentation technology for intelligent surgery and medical treatments | |
| Li et al. | Real three-dimensional medical interactive display system based on integral imaging and gesture recognition | |
| Stoyanov et al. | Current issues of photorealistic rendering for virtual and augmented reality in minimally invasive surgery | |
| US11869216B2 (en) | Registration of an anatomical body part by detecting a finger pose |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 23943390 Country of ref document: EP Kind code of ref document: A1 |