[go: up one dir, main page]

CN119055358A - A surgical operation force feedback guidance method based on virtual marker tracking and instrument posture - Google Patents

A surgical operation force feedback guidance method based on virtual marker tracking and instrument posture Download PDF

Info

Publication number
CN119055358A
CN119055358A CN202411580088.9A CN202411580088A CN119055358A CN 119055358 A CN119055358 A CN 119055358A CN 202411580088 A CN202411580088 A CN 202411580088A CN 119055358 A CN119055358 A CN 119055358A
Authority
CN
China
Prior art keywords
instrument
image
surgical
dimensional
tracking
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202411580088.9A
Other languages
Chinese (zh)
Other versions
CN119055358B (en
Inventor
莫杭杰
方进
杨善林
李霄剑
李玲
肖夕林
王昕�
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hefei University of Technology
Original Assignee
Hefei University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hefei University of Technology filed Critical Hefei University of Technology
Priority to CN202411580088.9A priority Critical patent/CN119055358B/en
Publication of CN119055358A publication Critical patent/CN119055358A/en
Application granted granted Critical
Publication of CN119055358B publication Critical patent/CN119055358B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/30Surgical robots
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/70Manipulators specially adapted for use in surgery
    • A61B34/76Manipulators having means for providing feel, e.g. force or tactile feedback
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/016Input arrangements with force or tactile feedback as computer generated output to the user
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G16INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
    • G16HHEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
    • G16H40/00ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
    • G16H40/60ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
    • G16H40/63ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/101Computer-aided simulation of surgical operations
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/101Computer-aided simulation of surgical operations
    • A61B2034/105Modelling of the patient, e.g. for ligaments or bones
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/10Computer-aided planning, simulation or modelling of surgical operations
    • A61B2034/107Visualisation of planned trajectories or target regions
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/20Surgical navigation systems; Devices for tracking or guiding surgical instruments, e.g. for frameless stereotaxis
    • A61B2034/2046Tracking techniques
    • A61B2034/2065Tracking using image or pattern recognition
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/30Surgical robots
    • A61B2034/301Surgical robots for introducing or steering flexible instruments inserted into the body, e.g. catheters or endoscopes
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B34/00Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
    • A61B34/30Surgical robots
    • A61B2034/302Surgical robots specifically adapted for manipulations within body cavities, e.g. within abdominal or thoracic cavities
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/04Indexing scheme for image data processing or generation, in general involving 3D image data
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Surgery (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Public Health (AREA)
  • Veterinary Medicine (AREA)
  • Animal Behavior & Ethology (AREA)
  • Molecular Biology (AREA)
  • Heart & Thoracic Surgery (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Robotics (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Primary Health Care (AREA)
  • Human Computer Interaction (AREA)
  • Computer Graphics (AREA)
  • Epidemiology (AREA)
  • General Business, Economics & Management (AREA)
  • Business, Economics & Management (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a surgical operation force feedback guiding method, a system, a storage medium and electronic equipment based on virtual mark tracking and instrument pose, and relates to the field of medical instruments. According to the invention, a real-time three-dimensional point cloud reconstruction is carried out on an in-vivo scene, a virtual mark dynamic tracking method based on key frame guidance is used for tracking a surgical path marked on a first image frame by a doctor, a three-dimensional surgical path is obtained based on the real-time three-dimensional point cloud, a 6D pose method of a surgical instrument based on 3D Gaussian tracking is combined, and finally a guiding type virtual clamp along the dynamic three-dimensional surgical path is established, so that the doctor can be effectively assisted in completing the operation along a preset surgical path through a force feedback mechanism, and man-machine co-fusion operation conforming to the operation intention of the doctor is realized.

Description

Surgical operation force feedback guiding method based on virtual mark tracking and instrument pose
Technical Field
The invention relates to the field of medical instruments, in particular to a surgical operation force feedback guiding method, a system, a storage medium and electronic equipment based on virtual mark tracking and instrument pose.
Background
Robotic surgery has gained widespread acceptance by surgeons as a minimally invasive surgical technique due to its high degree of precision. Among other things, force haptic feedback is a major challenge facing current surgical robots.
In the related art, patent CN202211328226.5 discloses a force feedback method and device for a surgical robot main manipulator, which comprises the steps of obtaining first force feedback data and scene data of the surgical robot main manipulator, generating a virtual protection domain corresponding to the surgical robot main manipulator based on the scene data, generating target force feedback data by using an impedance control algorithm based on the first force feedback data and the virtual protection domain, and sending a first instruction to a force feedback device controller, wherein the first instruction is used for instructing the force feedback device controller to control the surgical robot main manipulator based on the target force feedback data.
However, in the original scheme, aiming at the establishment of a tension field in a virtual protection domain, a marking path needs to be planned in advance, and the marking path is difficult to be accurately and real-time confirmed due to large intra-operative environmental change.
Disclosure of Invention
(One) solving the technical problems
Aiming at the defects of the prior art, the invention provides a surgical operation force feedback guiding method, a system, a storage medium and electronic equipment based on virtual mark tracking and instrument pose, which solve the technical problems that a mark path needs to be planned in advance for the establishment of a pull field in a virtual protection domain, and the mark path is difficult to be accurately and real-timely confirmed due to large intra-operative environmental change.
(II) technical scheme
In order to achieve the above purpose, the invention is realized by the following technical scheme:
a surgical operation force feedback guiding method based on virtual mark tracking and instrument pose, comprising:
acquiring a three-dimensional point cloud of an in-vivo scene based on depth estimation;
The method for estimating the 6D pose of the instrument based on the 3D Gaussian sputtering obtains the optimal 6D pose of the instrument of the current frame image so as to estimate the position of the tail end of the surgical instrument;
Tracking a surgical path marked on a first image frame by a doctor based on a virtual mark dynamic tracking method guided by a key frame to acquire a tracking path of a current frame image, wherein the tracking path is represented by a two-dimensional target point set;
Acquiring a three-dimensional operation path based on the three-dimensional point cloud and the two-dimensional target point set;
Searching a point cloud position closest to the tail end position of the surgical instrument on the three-dimensional surgical path, introducing an additional investigation radius, acquiring a manual vector of the tail end position of the surgical instrument, and establishing a guiding type virtual force field along the three-dimensional surgical path based on the manual vector so as to guide a doctor to operate the surgical robot system through a force feedback mechanism.
Preferably, the method for estimating the 6D pose of the instrument based on 3D Gaussian sputtering acquires the optimal 6D pose of the instrument of the current frame image so as to estimate the position of the tail end of the surgical instrument, and comprises the following steps:
The method comprises the steps of reading a current frame image, obtaining an instrument mask through a semantic segmentation algorithm, carrying out depth estimation on the current frame image by utilizing a binocular depth estimation algorithm, and respectively obtaining a tissue depth map and an instrument depth map by combining the instrument mask;
initializing the pose of the instrument, and performing 3D Gaussian sputtering by using a pre-reconstructed instrument Gaussian model to obtain a view, a projection mask and a depth map under the corresponding pose;
The method comprises the steps of constructing a total loss function according to a rendering result, a current frame image, an instrument mask and an instrument depth map thereof, updating and learning the instrument pose through a back propagation algorithm, minimizing the total loss function, and obtaining the optimal 6D pose of the instrument of the current frame image;
And determining the position of the instrument tail end corresponding to the Gaussian kernel based on the optimal 6D pose of the instrument of the current frame image, and taking the position as the position of the surgical instrument tail end.
Preferably, the initializing the instrument pose comprises:
For the first frame image, combining the camera model to obtain a rendering view of the instrument Gaussian model on an image plane, enabling the image to coincide with the area where the rendering view is located by moving the cavity mirror, and taking the instrument pose at the moment as an initial instrument pose of the first frame image;
for images other than the first frame, the instrument pose of the current frame image is initialized based on the optimal instrument pose of the previous frame image and optical flow information of two continuous frames.
Preferably, the reconstruction process of the Gaussian model of the instrument comprises the following steps:
Acquiring images of a plurality of known camera poses and including all view angles of the instrument so as to reconstruct sparse point clouds of the instrument;
initializing each point of the sparse point cloud into a Gaussian, and acquiring an initialized instrument Gaussian model;
And carrying out iterative optimization on the initialized instrument Gaussian model by combining the image sequences which are formed by the images with the known camera poses and comprise all view angles of the instrument and the corresponding camera poses, and obtaining the reconstructed instrument Gaussian model.
Preferably, the iterative optimization of the initialized instrument gaussian model to obtain the reconstructed instrument gaussian model by combining the image sequence composed of the images of the known camera poses and including all view angles of the instrument and the corresponding camera poses comprises:
The color consistency loss function between constructing the 3D gaussian rendering view and the original image retaining only the instrument portion is as follows:
Wherein, Respectively representing the center position, rotation and size of any Gaussian; L is the number of images included in the image sequence, L 1 represents a norm loss; Representing the process of 3D Gaussian micro-rendering an image when the pose of the camera is R l, G i represents the ith Gaussian point, I is the number of gauss; representing an original image derived from the sequence of images leaving only the instrument portion;
Introducing a multi-region alignment loss function, wherein the projection shape of each hinged part on the reconstructed instrument Gaussian model on the image plane and an instrument mask M lc on the original image are respectively aligned, and the multi-region alignment loss function is as follows:
Wherein, C represents the number of articulating components of the instrument; a process of performing 3D gaussian micro-renderable projection map when the camera pose is R l;
the total loss function of the instrument Gaussian model reconstruction process is as follows:
wherein alpha represents a corresponding weight coefficient;
Obtaining the reconstructed instrument Gaussian model by minimizing the total loss function L recons The following is shown:
Wherein, Representing the i-th reconstructed gaussian point; The value of the variable representing the smallest function value.
Preferably, the method comprises the steps of constructing a total loss function according to a rendering result, a current frame image, an instrument mask and an instrument depth map thereof, updating and learning the instrument pose through a back propagation algorithm, minimizing the total loss function, and obtaining an optimal 6D pose of the instrument of the current frame image, wherein the method comprises the following steps:
A color consistency loss function between the 3D gaussian rendered view and the current frame image is constructed as follows:
wherein T represents the current time, T t represents the pose of the current frame instrument; Representing a rendered view; , Representing a process of performing a 3D Gaussian micro-renderable image, M t representing an instrument mask for current image segmentation;
Constructing a region alignment loss function between the 3D gaussian micro-renderable projection map and an instrument mask of the current frame image, as follows:
Wherein, An instrument mask respectively representing the projection shape of the c-th hinge part on the instrument Gaussian model on the image plane and the current frame image;, a process for performing 3D gaussian micro-renderable projection maps; ;
a loss of depth consistency between the instrument depth map constructing the rendered depth map and the prior frame image is as follows:
Wherein, An instrument depth map representing the rendered depth map and the previous frame image, respectively;, representing a process of rendering a depth map of the instrument gaussian model;
The total loss function is constructed as:
Wherein, alpha 1、α2、α3 respectively represents corresponding weight coefficients;
Then, the instrument pose T t is iteratively updated through a back propagation algorithm to minimize L pose, and an instrument optimal 6D pose estimation result of the current frame image is obtained The following is shown:
preferably, the method for dynamically tracking virtual marks based on key frame guidance tracks a surgical path marked on a first image frame by a doctor to acquire a tracking path of a current frame image, and comprises the following steps:
Receiving an operation path marked on a first image frame by a doctor, wherein the operation path is represented by adopting a two-dimensional target point set at a pixel level;
acquiring a target area where a two-dimensional target point set on a first image frame is located, predicting the track of each point by utilizing an optical flow algorithm, and acquiring the two-dimensional target point set on a plurality of later image frames and the target area where the two-dimensional target point set is located so as to initialize a key frame pool;
Traversing the key frame pool aiming at the current frame image, inquiring similar key frames by using a sparse matching method, and acquiring an optimal template based on the sparse matching result between the inquired similar key frames and the current frame image;
when the optimal template is not an empty set, acquiring a tracking area on the current frame image based on a target area on each similar key frame in the optimal template;
predicting the motion trail of each point by utilizing an optical flow algorithm based on the two-dimensional target point set in each target area to obtain the two-dimensional target point set on the corresponding tracking area;
And when the current frame image is judged to be a new key frame based on preset conditions, the current frame image is put into the key frame pool, wherein the preset conditions comprise:
(1) A complete tracking path is obtained compared to the surgical path;
(2) Acquiring a tracking area where the final two-dimensional target point set is located, wherein the center of the tracking area is always within a preset range of the current frame image;
(3) If the similar key frame in the optimal template is greater than one, placing the last new key frame into the key frame pool at a distance exceeding a first preset number of image frames;
If the similar key frame in the optimal template is equal to one, placing the last new key frame into the key frame pool, wherein the distance exceeds a second preset number of image frames, and the ratio of the length of the tracking path to the length of the operation path is consistent with the scale change of the image frames where the tracking path and the operation path are positioned;
(4) The maximum capacity of the keyframe pool is not exceeded.
Preferably, the initialization key frame pool specifically comprises:
By using Representing a key frame, wherein j is a key frame index in a key frame pool; Representing RGB values, W, H representing width and height, respectively; representing a target area; respectively represent extraction by using descriptor-based sparse matching method Feature points and feature descriptions in (3); Representing a two-dimensional set of target points;
Designating the 0 th image frame F 0 as a key frame K 0 E K, defining a two-dimensional target point set thereon as ,Represents the coordinates of the s-th pixel point, n represents the number of the pixel points,Representing a real set;
To be used for A region being the centerGenerated by the following formula:
wherein the center point ,Is thatAn arithmetic average of the maximum and minimum values of the abscissa of all pixel points,Is thatThe arithmetic average value of the maximum value and the minimum value of the ordinate of all pixel points in the image, L is the side length of the target area, and L is not less thanDifferences between maximum and minimum values of the abscissa and the ordinate of all pixel points in (a);
Defining a target area Storing four vertices and center pointsExpressed asWhereinFrom the following componentsIt can be seen that the representationUpper left, upper right, lower left and lower right;
Predicting motion V 0→1 E of two later image frames F 1 and F 2 by optical flow algorithm ,V0→2And calculates each pixel point on the operation path according to the optical flow resultTrajectories on F 1 and F 2:
wherein Interp (∙) represents a sampling process to obtain a corresponding set of target points AndAnd referring to formula (1), acquiring a target regionAnd;
The F 1、F2 correspondence is noted as key frame K 1、K2, together with K 0, constitutes an initial key frame pool p= { K 0,K1,K2 }.
Preferably, the step of traversing the key frame pool for the current frame image, performing similar key frame query by using a sparse matching method, and acquiring an optimal template based on a sparse matching result between the queried similar key frame and the current frame image comprises the following steps:
acquiring characteristic points and characteristic descriptions of the current frame image F i;
Traversing each key frame K j in the key frame pool, and acquiring a matching point between F i and K j by using a sparse matching method;
counting the number n ji of the matching points of the target area on the K j, selecting three key frames with the maximum n ji as similar key frames K simi={Ka,Kb,Kc } ⊆ P,
The following formula is adopted to obtain the optimal template:
Wherein K j represents the j-th key frame in the key frame pool, j has the value of a, b or c, and n ji represents the target area on K j The number of matching points between the current frame image F i and the current frame image F i, wherein sigma is a first threshold value, and K simi represents all queried similar key frames;
And/or
The method for acquiring the tracking area on the current frame image based on the target area on each similar key frame in the optimal template specifically comprises the following steps:
Based on the matching point between K j⊆Kbest and F i, the accuracy of estimating homography matrix H ji,Hji by using RANSAC is determined by N in is the number of inner points;
When (when) Not less than lambda, 0< lambda <1, lambda being a second threshold, for the target regionFour vertices ofHomography conversion is carried out to obtainCorresponding position on F i :
Otherwise, carrying out vertex coordinate conversion by using a neighborhood similarity transformation method:
Extracting a set of matches of F i and K j from sparse matching results ,ximAnd y jm EIs a two-dimensional column vector representing the feature point locations;
Searching for target area vertices N neighbors of (a)Obtaining corresponding matching pointsTo obtain a dotCorresponding position on F i :
Wherein the method comprises the steps ofW t andAndThe distance between them is inversely proportional;
based on vertex coordinates An initial region is determined, expressed as:
Wherein, Respectively representCorresponding position on F i U min、umax represents the maximum and minimum values of the abscissa of the four vertices, respectively, v min、vmax represents the maximum and minimum values of the ordinate of the four vertices, respectively;
determining a tracking area based on coordinates of four vertices of the initial area Is the center point of (2);
Wherein, Respectively represent tracking areasIs the center point of (2)Is the abscissa and ordinate of (2);
Referring to formula (1), a tracking area on the current frame image is acquired
Preferably, the method comprises predicting motion trajectories of points by optical flow algorithm based on two-dimensional target point sets in each target area to obtain two-dimensional target point sets on corresponding tracking areas, merging the two-dimensional target point sets on each tracking area to obtain final two-dimensional target point sets, and specifically comprises:
Two-dimensional target point set Coordinates of each point p j,s in the whole frame K j and relative target areaCoordinates of (c)Has the following mapping relation:
Optical flow in target area And tracking areaIn the above, the optical flow prediction result is V j→i, and the sparse matching result and the optical flow prediction result should be consistent, and the difference between the two is calculated by the following formula:
if d is less than or equal to gamma, and gamma is a third threshold value, the tracking area Target point set onP ji,s is obtained by the following method:
Otherwise the first set of parameters is selected, Not participate inIs calculated;
Merging two-dimensional target point sets on each tracking area Obtaining a final two-dimensional target point setExpressed as:
wherein lambda ji is the weight, and =1, Λ ji is proportional to n ji.
Preferably, the searching the point cloud position closest to the end position of the surgical instrument on the three-dimensional surgical path, introducing an additional investigation radius, and obtaining the artificial vector of the end position of the surgical instrument, comprises:
searching the end position of the surgical instrument on the three-dimensional surgical path by using a K-nearest neighbor algorithm Nearest point cloud locationObtaining the relative distance x d between the two points;
Providing that all points on the three-dimensional surgical path are contained in a set of virtual marker points V r for the surgical instrument tip position Distance on three-dimensional surgical pathPoint-closest point cloud locationFind the set of sub-points of V r WhereinTo additionally investigate the radius;
The surgical instrument end position The corresponding artificial vector is expressed in the form:
The artificial vector length function is calculated as follows:
Wherein, Is the position of the tail end of the surgical instrumentThe distance between the virtual mark point v and the virtual mark point v, R and R are upper and lower boundaries of the distance respectively, when the distance is smaller than or equal to R, the maximum weight q max is allocated to the virtual mark point v, when the distance is larger than R and smaller than R, the weight of the virtual mark point v is between q max and 0, and when the distance is larger than or equal to R, the weight of the virtual mark point v is 0; The representation is composed of Vector pointing to point v, and card (∙) represents the cardinality or number of elements of the collection.
Preferably, the guided virtual force field is expressed as:
Wherein K vf represents the proportionality coefficient of attraction force and artificial vector, D vf(xd) represents the damping coefficient Representation ofIs a function of the speed of the machine.
A surgical operative force feedback guidance system based on virtual marker tracking and instrument pose, comprising:
the depth estimation module is used for acquiring a three-dimensional point cloud of the in-vivo scene based on the depth estimation;
the pose estimation module is used for acquiring the optimal 6D pose of the instrument of the current frame image based on a 3D Gaussian sputtering instrument 6D pose estimation method so as to estimate the tail end position of the surgical instrument;
The mark tracking module is used for tracking the operation path marked on the first image frame by a doctor based on a virtual mark dynamic tracking method guided by the key frame to acquire a tracking path of the current frame image, wherein the tracking path is represented by a two-dimensional target point set;
The path conversion module is used for acquiring a three-dimensional operation path based on the three-dimensional point cloud and the two-dimensional target point set;
The force field establishment module is used for searching a point cloud position closest to the tail end position of the surgical instrument on the three-dimensional surgical path, introducing an additional investigation radius, acquiring an artificial vector of the tail end position of the surgical instrument, and establishing a guiding type virtual force field along the three-dimensional surgical path based on the artificial vector so as to guide a doctor to operate the surgical robot system through a force feedback mechanism.
A storage medium storing a computer program for surgical operation force feedback guidance based on virtual marker tracking and instrument pose, wherein the computer program causes a computer to execute the surgical operation force feedback guidance method as described above.
An electronic device, comprising:
The surgical force feedback guidance system comprises one or more processors, memory, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the programs comprising instructions for performing the surgical force feedback guidance method as described above.
(III) beneficial effects
The invention provides a surgical operation force feedback guiding method, a system, a storage medium and electronic equipment based on virtual mark tracking and instrument pose. Compared with the prior art, the method has the following beneficial effects:
According to the invention, a real-time three-dimensional point cloud reconstruction is carried out on an in-vivo scene, a virtual mark dynamic tracking method based on key frame guidance is used for tracking a surgical path marked on a first image frame by a doctor, a three-dimensional surgical path is obtained based on the real-time three-dimensional point cloud, a 6D pose method of a surgical instrument based on 3D Gaussian tracking is combined, and finally a guiding type virtual clamp along the dynamic three-dimensional surgical path is established, so that the doctor can be effectively assisted in completing the operation along a preset surgical path through a force feedback mechanism, and man-machine co-fusion operation conforming to the operation intention of the doctor is realized.
Drawings
In order to more clearly illustrate the embodiments of the invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, it being obvious that the drawings in the following description are only some embodiments of the invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a block diagram of a surgical force feedback guidance method based on virtual marker tracking and instrument pose according to an embodiment of the present invention;
fig. 2 is an overall flowchart of an instrument 6D pose estimation method based on 3D gaussian sputtering according to an embodiment of the present invention;
FIG. 3 is a schematic view of a coordinate system for articulated motion according to an embodiment of the present invention;
fig. 4 is an overall flowchart of a dynamic tracking method of virtual marks based on key frame guidance according to an embodiment of the present invention.
Detailed Description
For the purpose of making the objects, technical solutions and advantages of the embodiments of the present invention more clear, the technical solutions in the embodiments of the present invention are clearly and completely described, and it is obvious that the described embodiments are some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The embodiment of the application solves the technical problems that a marking path needs to be planned in advance for the establishment of a pull field in a virtual protection domain, and the marking path is difficult to be accurately and real-timely confirmed due to large in-vivo environmental change in an operation by providing a surgical operation force feedback guiding method, a system, a storage medium and electronic equipment based on virtual marking tracking and instrument pose.
The technical scheme in the embodiment of the application aims to solve the technical problems, and the overall thought is as follows:
In the embodiment of the invention, the real-time three-dimensional point cloud reconstruction is carried out on the in-vivo scene, the virtual mark dynamic tracking method based on key frame guidance is used for dynamically tracking the operation path marked on the first image frame by a doctor, the three-dimensional operation path is obtained based on the real-time three-dimensional point cloud, and the guiding virtual clamp along the dynamic three-dimensional operation path is finally established by combining the 6D pose method of the operation instrument based on the 3D Gaussian tracking, so that the doctor can be effectively helped to complete the operation along the preset operation path by a force feedback mechanism, and the man-machine co-fusion operation conforming to the operation intention of the doctor is realized.
In particular, a binocular depth estimation-based instrument 6D pose estimation method is provided. Aiming at the difficulties of 6D pose estimation and tracking of the instrument caused by the characteristics of fast motion, low texture, serious reflection and the like of the surgical instrument and the problem of reduced tissue depth estimation precision caused by the existence of the instrument, the three-dimensional position of the instrument is accurately estimated, and the complete three-dimensional reconstruction of a surgical scene and the three-dimensional position estimation and tracking of the surgical instrument can be realized.
And a key frame guided virtual mark dynamic tracking method is provided, so that the two-dimensional virtual mark on the laparoscopic image is tracked with pixel-level precision, and the intuitiveness and the accuracy of information transmission are ensured. The two-stage tracking method adopted reduces error tracking. In addition, the diversity of the key frame pool is enriched by dynamically updating the key frame pool, so that the robustness and accuracy of dynamic tracking of the virtual mark are further improved.
In order to better understand the above technical solutions, the following detailed description will refer to the accompanying drawings and specific embodiments.
Example 1:
as shown in fig. 1, an embodiment of the present invention provides a surgical operation force feedback guiding method based on virtual mark tracking and instrument pose, including:
s1, acquiring a three-dimensional point cloud of an in-vivo scene based on depth estimation;
S2, acquiring the optimal 6D pose of the instrument of the current frame image by using a 3D Gaussian sputtering-based instrument 6D pose estimation method so as to estimate the tail end position of the surgical instrument;
S3, tracking a surgical path marked on a first image frame by a doctor based on a virtual mark dynamic tracking method guided by a key frame to acquire a tracking path of a current frame image, wherein the tracking path is represented by a two-dimensional target point set;
s4, acquiring a three-dimensional operation path based on the three-dimensional point cloud and the two-dimensional target point set;
S5, searching a point cloud position closest to the tail end position of the surgical instrument on the three-dimensional surgical path, introducing an additional investigation radius, obtaining an artificial vector of the tail end position of the surgical instrument, and establishing a guiding type virtual force field along the three-dimensional surgical path based on the artificial vector so as to guide a doctor to operate the surgical robot system through a force feedback mechanism.
In the embodiment of the invention, the real-time three-dimensional point cloud reconstruction is carried out on the in-vivo scene, the virtual mark dynamic tracking method based on key frame guidance is used for tracking the operation path marked on the first image frame by a doctor, the three-dimensional operation path is obtained based on the real-time three-dimensional point cloud, the 6D pose method of the operation instrument based on 3D Gaussian tracking is combined, and finally, the guiding type virtual clamp along the dynamic three-dimensional operation path is established, so that the doctor can be effectively helped to complete the operation along the preset operation path through a force feedback mechanism, and the man-machine co-fusion operation conforming to the operation intention of the doctor is realized.
The steps of the above scheme will be described in detail as follows:
in step S1, a three-dimensional point cloud of an in-vivo scene is acquired based on the depth estimation.
In this step, a depth estimation algorithm may be used to obtain a three-dimensional point cloud V d of the in-vivo scene based on a sequence of image frames acquired by laparoscopic surgery that show the in-vivo scene.
By way of example, a neural network may be used herein for depth estimation to present a three-dimensional field of view to a physician, creating a surgical experience that is intuitive to the physician.
In step S2, an instrument 6D pose estimation method based on 3D gaussian sputtering obtains an instrument optimal 6D pose of the current frame image to estimate a surgical instrument end position.
It can be understood that the estimation result of the instrument pose, the relative position of the instrument and the tissue and the interaction state can provide high-precision positioning and navigation information in real time, thereby being beneficial to improving the accuracy and safety of the minimally invasive surgery. Based on the above, the embodiment of the invention provides an instrument-tissue interaction distance estimation method based on binocular depth estimation, instrument 6D pose estimation and depth complementation, which specifically comprises the following steps:
S21, reading a current frame image, and acquiring an instrument mask through a semantic segmentation algorithm, carrying out depth estimation on the current frame image by utilizing a binocular depth estimation algorithm, and respectively acquiring a tissue depth map and an instrument depth map by combining the instrument mask;
S22, initializing the pose of the instrument, and performing 3D Gaussian sputtering by using a pre-reconstructed instrument Gaussian model to obtain a view, a projection mask and a depth map under the corresponding pose;
S23, constructing a total loss function according to the rendering result, the current frame image, an instrument mask and an instrument depth map thereof, updating and learning the instrument pose through a back propagation algorithm, minimizing the total loss function, and obtaining the optimal 6D pose of the instrument of the current frame image;
s24, determining the position of the instrument end corresponding to the Gaussian kernel based on the optimal 6D pose of the instrument of the current frame image, and taking the position as the position of the surgical instrument end.
According to the scheme, the instrument pose estimation is estimated by aligning the instrument Gaussian model and the image characteristics, so that the 6D pose of the instrument can be accurately and robustly recovered in a complex operation scene, and the instrument pose estimation is realized on the basis of a 3D Gaussian sputtering frame, so that the instrument pose can be rapidly recovered in a shorter iteration time and a small number of iteration times.
Fig. 2 shows an overall flowchart of the above scheme, and the method for estimating the pose and depth of the instrument 6D according to the embodiment of the present invention includes, but is not limited to, the pose estimation and the relative distance estimation of the instrument-tissue of the surgical instrument 6D used in laparoscopic surgery or robot-assisted surgery. The method mainly comprises depth estimation and instrument segmentation based on binocular laparoscopic images, 3D Gaussian model reconstruction of surgical instruments, 6D pose estimation of the surgical instruments based on 3D Gaussian optimization and the like.
The details of step S2 will be described in detail below in conjunction with fig. 2:
in step S21, reading the current frame image, obtaining an instrument mask through a semantic segmentation algorithm, carrying out depth estimation on the current frame image by utilizing a binocular depth estimation algorithm, and respectively obtaining a tissue depth map and an instrument depth map by combining the instrument mask;
in the step, the image is subjected to multi-value segmentation by utilizing a semantic segmentation algorithm (random forest, SAM, perSAM and the like) to obtain an instrument mask on the image Where K represents the number of articulations according to the instrument. And performing depth estimation on the endoscope image by using a binocular depth estimation method (MIOL algorithm, CREStereo algorithm and the like) to obtain a depth map D t, and combining the instrument mask M t to obtain a tissue depth map D t(1-Mt of a non-instrument shielding region and an instrument depth map D tMt of an instrument visible part. Where t represents the current time.
In step S22, the pose of the instrument is initialized, and 3D gaussian sputtering is performed by using the pre-reconstructed instrument gaussian model, so as to obtain a view, a projection mask and a depth map under the corresponding pose.
Note that 3D gaussian sputter rendering is mainly used for new view synthesis, where the 3D position of an object represented by a 3D gaussian is represented by the center position of each gaussianRotatingSize of the productColor ofThe density sigma is determined jointly, and is not suitable for expressing the 6D pose transformation of a dynamic object.
Accordingly, the embodiment of the invention establishes a 3D Gaussian model of the surgical instrument (i.e. a reconstructed instrument Gaussian model) in advance, so that the 3D Gaussian optimization is converted into Gaussian pose optimization, i.e. the pose optimization of the surgical instrument.
Specifically, the reconstruction process of the instrument Gaussian model comprises the following steps:
Firstly, acquiring images of a plurality of known camera poses and including all view angles of the instrument so as to reconstruct sparse point clouds of the instrument, wherein the method specifically comprises the following steps:
a) Fixing the position of the surgical instrument, aligning the camera with the instrument and shooting the camera around the instrument to obtain a plurality of instrument images I l containing all view angles, L is the number of images. However, the characteristics of the surgical instrument such as no texture, specular reflection and the like result in that the relative camera pose cannot be easily solved through the image features between adjacent frames. The feature points are provided by attaching marker points to the instrument or by fixing a checkerboard in the field of view.
B) Reconstructing a sparse point cloud of the photographed scene through sfm, and enabling the image to correspond to 2D-3D points of the sfm point cloud.
C) Scaling the sfm point cloud according to the real scale of the instrument, recovering the absolute scale of the sparse point cloud, and solving the absolute camera pose R l of each image by using the 2D-3D corresponding points.
D) The segmentation algorithm is utilized to segment the instrument to obtain a segmentation mask of the instrument, which is used for removing the background outside the instrument in the image to obtain the image with the instrument only
E) Image known from camera pose R l The sparse point cloud M sfm of the instrument was reconstructed using sfm algorithm for initialization of the 3D gaussian model reconstruction.
Each point of the sparse point cloud is then initialized to a gaussian G i,Each Gaussian has the property that the center position of the Gaussian isRotatingSize of the productColor ofDensity σ, probability f of belonging to the foreground.
It should be noted that the parameters to be optimized in the reconstruction process of the Gaussian model of the surgical instrument are only the center position of each GaussianRotatingSize of the product
In addition, the instrument Gaussian model { G i } at this stage is initialized by initializing each point of the instrument point cloud M sfm to a Gaussian G i,I is the number of gauss, the position and the color of the point cloud are initialized to be the central position u and the color c of the gauss, the rotation q of the gauss is initialized to be {1,0 }, the distance from each point to the nearest neighbor is initialized to be the size of the gauss, the density sigma of the gauss is initialized to be 1, and the probability f of belonging to the foreground is initialized to be 1.
On the basis, the image sequence formed by combining the images of the known camera poses and including all view angles of the instrument and the corresponding camera poses thereof iteratively optimizes the initialized instrument Gaussian model to obtain the reconstructed instrument Gaussian model, and the method comprises the following steps:
The color consistency loss function between constructing the 3D gaussian rendering view and the original image retaining only the instrument portion is as follows:
Wherein L 1 represents a norm loss; representing an original image derived from the sequence of images leaving only the instrument portion; The camera pose is represented as R l.
In addition, in order to ensure that the reconstructed instrument Gaussian model is of a hinged structure, the embodiment of the invention introduces a multi-region alignment loss function, wherein the projection shape of each hinged part on the reconstructed instrument Gaussian model on an image plane and an instrument mask M lc on an original image are constrained to be aligned respectively, and the multi-region alignment loss function is as follows:
Wherein, C represents the number of articulating components of the instrument; and (3) representing a process of performing 3D Gaussian micro-renderable projection graph when the pose of the camera is R l.
The total loss function of the instrument gaussian model reconstruction process is:
wherein alpha represents a corresponding weight coefficient;
Obtaining the reconstructed instrument Gaussian model by minimizing the total loss function L recons The following is shown:
Wherein, Representing the i-th reconstructed gaussian point; The value of the variable representing the smallest function value.
Optimized instrument Gaussian modelIs marked asI represents the index of the gaussian kernel.
In the process of acquiring instrument Gaussian modelOn the basis of the above, for the rigid movement T between the hinge parts of the surgical instrument, the embodiment of the invention defines the hinge movement of the instrument by DH parameter method, regards the rod part and the left clamp (the right clamp is symmetrical with the left clamp) of the instrument as a two-link structure, and defines the local coordinate system F 1、F2 for the rod part and the left clamp (the right clamp is symmetrical with the left clamp) of the instrument respectively as shown in fig. 3, and establishes the camera coordinate system F cam. The instrument pose estimation solves the problem that the instrument is transformed into a matrix from a coordinate system of the instrument to a camera coordinate system, at the moment, a point X under a local coordinate F 1、F2 of the instrument is transformed into a formula under the camera coordinate system as follows:
Wherein, Representing the points under the local coordinate system F 1、F2, respectively, the point of the left (right) clip in fig. 3 is in the local coordinate system F 2, the point of the instrument shaft is in the local coordinate system F 1,Is a transformation matrix of the instrument bar coordinate system F 1 to the camera coordinate system F cam,Is a transformation matrix of the instrument left clamp coordinate system F 2 to the instrument rod coordinate system F 1, and the instrument left clamp coordinate system F 2 must rotate around the z axis of the rod coordinate system F 1, so that the Euler angle θ of one degree of freedom can represent the rotation between the local coordinate systems F 1、F2.
In summary, the stiffness transformation T of the articulating instrument can be expressed asWhereinR represents a 3D rotation matrix,Representing the 3D translation vector.
After defining the reconstruction process of the Gaussian model of the instrument and the rigid motion parameters corresponding to the Gaussian model, as shown in FIG. 2 and FIG. 3, the pose parameters of the instrument are optimized based on the defined Gaussian model to obtain the 6D pose of the articulated instrument. And giving a binocular image of the current frame, acquiring a depth image of the current frame through a binocular depth estimation algorithm, and obtaining a segmentation mask of the instrument through an instrument segmentation algorithm. The input image, the depth map and the instrument mask are used as monitoring information for subsequent pose learning. On the basis of the instrument initialization position, a view, a depth map and a projection mask (represented as rendered images, segmentation and depth information in fig. 2) under the corresponding pose are rendered through 3D Gaussian sputtering, and a loss function is respectively constructed by the view, the depth map, the projection mask, an original image, a semantic segmentation mask and a depth estimation result (represented as images, segmentation and depth information in fig. 2), so that a total loss function is constructed according to the subsequent steps, and the instrument pose is updated and learned through a back propagation algorithm, so that the optimal 6D pose of the instrument of the current frame image is obtained.
In the instrument pose optimization stage, the update of the instrument Gaussian model follows the following formula:
Where i denotes the index of the gaussian kernel, t denotes the current frame, Is an updated analogy formula (5) of the reconstructed instrument gaussian model, u i,t, as follows:
when u i,0 belongs to the point of the instrument shaft, the following formula is updated:
Wherein, Representing a transformation matrix converting u i,0 into a local coordinate system F 1; and determining when the Gaussian model of the instrument is reconstructed and corresponding pose parameters are defined.
When u i,0 belongs to the point of the instrument left (right) clamp, the following formula is updated:
Wherein, Representing a transformation matrix converting u i,0 into a local coordinate system F 2; and determining when the Gaussian model of the instrument is reconstructed and corresponding pose parameters are defined.
Initialization of the instrument pose is mainly divided into two cases.
Firstly, for a first frame of image, combining a camera model to obtain a rendering view of an instrument Gaussian model on an image plane, enabling an image to coincide with an area where the rendering view is located by moving a cavity mirror, and taking the instrument pose at the moment as an initial instrument pose of the first frame of image;
second, for images other than the first frame, the instrument pose of the current frame image is initialized based on the optimal instrument pose of the previous frame image and optical flow information of two continuous frames.
Initial instrument pose of current frame image I t 3D translation from previous frame instrument pose T t-1 and coarse estimateObtaining:
Wherein, Is calculated as follows:
Wherein, Representing the mean of the predicted optical flow in the instrument region from the previous frame to the current frame,Representing the average depth value of the instrument area of the previous frame, f x,fy represents the focal length in the x, y direction of the camera, respectively.
In step S23, a total loss function is constructed according to the rendering result, the current frame image, the instrument mask and the instrument depth map thereof, and the instrument pose is updated and learned through a back propagation algorithm, so that the total loss function is minimized, and the optimal 6D pose of the instrument of the current frame image is obtained.
Furthermore, the invention implements the optimization of the pose of the setting instrument mainly by rendering the viewColor consistency loss L im between original image I t, rendered instrument projection maskRegion alignment loss L region between the instrument segmentation mask M t, rendered instrument depth mapAnd the depth consistency loss L depth between the instrument depth estimation D t.
Under the action of the iteratively updatable instrument pose T t, a color consistency loss function between the 3D Gaussian rendering view and the current frame image is constructed as follows:
wherein T represents the current time, T t represents the pose of the current frame instrument; Representing a rendered view; , Representing a process of performing a 3D Gaussian micro-renderable image, M t representing an instrument mask for current image segmentation;
Constructing a region alignment loss function between the 3D gaussian micro-renderable projection map and an instrument mask of the current frame image, as follows:
Wherein, An instrument mask respectively representing the projection shape of the c-th hinge part on the instrument Gaussian model on the image plane and the current frame image;, a process for performing 3D gaussian micro-renderable projection maps; ;
a loss of depth consistency between the instrument depth map constructing the rendered depth map and the prior frame image is as follows:
Wherein, An instrument depth map representing the rendered depth map and the previous frame image, respectively;, representing a process of rendering a depth map of the instrument gaussian model;
Thus, the total loss function is constructed as:
Wherein, alpha 1、α2、α3 represents corresponding weight coefficients respectively.
Then, the instrument pose T t is iteratively updated through a back propagation algorithm to minimize L pose, and an instrument optimal 6D pose estimation result of the current frame image is obtainedThe following is shown:
In step S24, based on the optimal 6D pose of the instrument in the current frame image, the position of the end of the instrument corresponding to the gaussian kernel is determined and used as the position of the end of the surgical instrument.
Thus, given a binocular endoscopic image, embodiments of the present invention successfully provide the surgical instrument end position of the articulating instrument
In step S3, a virtual mark dynamic tracking method based on key frame guidance tracks a surgical path marked on a first image frame by a doctor to obtain a tracking path of a current frame image, wherein the tracking path is represented by a two-dimensional target point set, and the method comprises the following steps:
S31, receiving a surgical path marked on a first image frame by a doctor, wherein the surgical path is represented by adopting a two-dimensional target point set at a pixel level;
S32, acquiring a target area where a two-dimensional target point set on a first image frame is located, predicting the track of each point by utilizing an optical flow algorithm, and acquiring the two-dimensional target point set on a plurality of later image frames and the target area where the two-dimensional target point set is located so as to initialize a key frame pool;
S33, traversing the key frame pool aiming at the current frame image, inquiring similar key frames by using a sparse matching method, and acquiring an optimal template based on the sparse matching result between the inquired similar key frames and the current frame image;
S34, when the optimal template is not an empty set, acquiring a tracking area on the current frame image based on a target area on each similar key frame in the optimal template;
S35, predicting the motion trail of each point by utilizing an optical flow algorithm based on the two-dimensional target point set in each target area to obtain the two-dimensional target point set on the corresponding tracking area;
S36, based on preset conditions, when the current frame image is judged to be a new key frame, the current frame image is put into the key frame pool, wherein the preset conditions comprise:
(1) A complete tracking path is obtained compared to the surgical path;
(2) Acquiring a tracking area where the final two-dimensional target point set is located, wherein the center of the tracking area is always within a preset range of the current frame image;
(3) If the similar key frame in the optimal template is greater than one, placing the last new key frame into the key frame pool at a distance exceeding a first preset number of image frames;
If the similar key frame in the optimal template is equal to one, placing the last new key frame into the key frame pool, wherein the distance exceeds a second preset number of image frames, and the ratio of the length of the tracking path to the length of the operation path is consistent with the scale change of the image frames where the tracking path and the operation path are positioned;
(4) The maximum capacity of the keyframe pool is not exceeded.
The scheme realizes the pixel-level precision tracking of the two-dimensional virtual mark on the laparoscopic image, and ensures the intuitiveness and accuracy of information transmission.
As shown in fig. 4, fig. 4 shows an overall flowchart of the virtual marker dynamic tracking method. The details of step S30 will be described in detail below with reference to fig. 4:
In step S31, receiving a surgical path marked by a doctor on a first image frame, wherein the surgical path is represented by adopting a two-dimensional target point set at a pixel level;
It will be appreciated that in minimally invasive surgery, virtual markers track the position of a dynamically positioned target, providing real-time feedback to the physician regarding the surgical location. The doctor selects the template frame as the 0 th image frame F 0 epsilon Virtual marks such as planned operation tracks are marked on a display by using a tool such as a mouse or a touch screen pen, and W, H respectively represent the width and the height of F 0.
The step receives the operation path and adopts a two-dimensional target point set at a pixel levelRepresenting, whereinRepresenting the virtual marker on frame 0, p 0,s epsilonRepresents the coordinates of the s-th pixel, n represents the number of pixels,Representing a set of real numbers.
Through the setting, the following movement along with the laparoscope is facilitated, and the frame F i E is aimed at the ith frameThe tracking path on the frame is obtained using the proposed tracking methodWill track the resultsLabeling the current frame image F i can provide visual guidance for doctors.
In step S32, a target area where a two-dimensional target point set on the first image frame is located is obtained, and the track of each point is predicted by using an optical flow algorithm, and the two-dimensional target point set on the next plurality of image frames and the target area where the two-dimensional target point set is located are obtained, so as to initialize a keyframe pool.
The key frame strategy has a core principle that it can identify and record key time instants as reference points for error correction. The key frames are utilized to reduce the computational complexity, improve the tracking efficiency and ensure the stability of long-time tracking.
Illustratively, this step requires determining the initial three key frames as templates for later tracking. Correspondingly, the specific steps of initializing the key frame pool are as follows:
It should be noted that in the embodiments of the present invention Representing a key frame, wherein j is a key frame index in a key frame pool; Representing RGB values, W, H representing width and height, respectively; representing a target area; respectively represent extraction by using descriptor-based sparse matching method Feature points and feature descriptions in (3); representing a two-dimensional set of target points.
First, the 0 th color image frame F 0 is designated as a key frame K 0 E K, and a two-dimensional target point set is defined as,Represents the coordinates of the s-th pixel point, n represents the number of the pixel points,Representing a set of real numbers.
Second, byA region being the centerGenerated by the following formula:
wherein the center point ,Is thatAn arithmetic average of the maximum and minimum values of the abscissa of all pixel points,Is thatThe arithmetic average value of the maximum value and the minimum value of the ordinate of all pixel points in the image, L is the side length of the target area, and L is not less thanThe difference between the maximum and minimum values of the abscissa and the ordinate of all pixel points in (a).
Again, F 0 is processed by using the existing sparse matching method based on descriptors, such as SIFT, lightglue and the like, and feature points and descriptors are extracted from the whole image and are positioned in the target areaThe internal feature points and descriptors are noted as
Then, defining a target areaStoring four vertices and center pointsExpressed asWhereinFrom the following componentsIt can be seen that the representationUpper left, upper right, lower left, lower right.
Then, the motion V 0→1 E of the two latter image frames F 1 and F 2 is predicted by using an optical flow algorithm,V0→2And calculates each pixel point on the virtual mark according to the optical flow resultTrajectories on F 1 and F 2:
wherein Interp (∙) represents a sampling process to obtain a corresponding set of target points AndAnd referring to formula (1), acquiring a target regionAnd
Finally, the F 1、F2 correspondence is noted as key frame K 1、K2, which together with K 0 constitutes the initial key frame pool p= { K 0,K1,K2 }.
In step S33, for the current frame image, traversing the keyframe pool, performing similar keyframe query by using a sparse matching method, and obtaining an optimal template based on a sparse matching result between the queried similar keyframe and the current frame image, wherein the method specifically comprises:
acquiring characteristic points and characteristic descriptions of the current frame image F i;
Traversing each key frame K j in the key frame pool, and acquiring a matching point between F i and K j by using a sparse matching method;
counting the number n ji of the matching points of the target area on the K j, selecting three key frames with the maximum n ji as similar key frames K simi={Ka,Kb,Kc } ⊆ P,
The following formula is adopted to obtain the optimal template:
Wherein K j represents the j-th key frame in the key frame pool, j has the value of a, b or c, and n ji represents the target area on K j The number of matching points with the current frame image F i, sigma is a first threshold value, and K simi represents all the queried similar key frames.
It should be noted that, as shown in fig. 3, based on the number of similar key frames card (K best) in the best template, the current frame image F i may be divided into two states:
(1) If the card (K best) >0, namely the optimal template is not empty, the fact that the current frame has a region with high similarity with the target region on the key frame and a mark needing to be tracked exists, at the moment, the current frame is considered as a normal frame or a candidate key frame, the next step is carried out for tracking the target region, and accurate positioning is carried out to the pixel position of the operation path.
It will be appreciated that setting this condition identifies situations that may occur after the surgical path is lost.
(2) If card (K best) =0, i.e. the best template is empty, it indicates that the similarity between the current frame and all key frames is not high, the mark has moved out of view of the surgical path, there is no mark to be traced, the current frame is considered as a frame not to be traced, and the next frame image is directly read.
Then, the embodiment of the invention adopts a two-stage tracking method, when tracking the operation path, adopts two-stage matching from thick to thin, firstly determines similar key frames through traversing a key frame pool and sparse matching, firstly determines a local area by utilizing a matching result, and then tracks the local area to reduce error tracking, and particularly, the invention refers to steps S34-S35:
In step S34, when the best template is not an empty set, a tracking area on the current frame image is obtained based on the target area on each similar key frame in the best template;
this step determines the similar region of the current frame image and the best template frame, noting that tracking is established on the region, not the entire image.
Correspondingly, acquiring the tracking area on the current frame image specifically comprises the following steps:
First, a target region is determined based on a homography transform or a neighborhood similarity transform Four vertices ofCorresponding position on F i Comprising:
Based on the matching point between K j⊆Kbest and F i, the accuracy of estimating homography matrix H ji,Hji by using RANSAC is determined by N in is the number of inner points;
When (when) Not less than lambda, 0< lambda <1, lambda being a second threshold, for the target regionFour vertices ofHomography conversion is carried out to obtainCorresponding position on F i :
Otherwise, carrying out vertex coordinate conversion by using a neighborhood similarity transformation method:
Extracting a set of matches of F i and K j from sparse matching results ,ximAnd y jm EIs a two-dimensional column vector representing the feature point locations;
Searching for target area vertices N neighbors of (a)Obtaining corresponding matching pointsTo obtain a dotCorresponding position on F i :
Wherein the method comprises the steps ofW t andAndThe distance between them is inversely proportional.
Illustratively, N may be four herein.
Then, based on vertex coordinatesAn initial region is determined, expressed as:
Wherein, Respectively representCorresponding position on F i U min、umax represents the maximum and minimum values of the abscissa of the four vertices, respectively, and v min、vmax represents the maximum and minimum values of the ordinate of the four vertices, respectively.
Then, a tracking area is determined based on coordinates of four vertexes of the initial areaIs the center point of (2);
Wherein, Respectively represent tracking areasIs the center point of (2)And the abscissa and ordinate of (c).
Finally, referring to formula (1), a tracking area on the current frame image is obtained
In step S35, based on the two-dimensional target point set in each target area, predicting the motion trail of each point by using an optical flow algorithm to obtain the two-dimensional target point set on the corresponding tracking area, merging the two-dimensional target point sets on each tracking area to obtain the final two-dimensional target point set, and taking the final two-dimensional target point set as the tracking path on the current frame image.
The step of carrying out mark tracking on the target area and the tracking area comprises the following steps:
Two-dimensional target point set Coordinates of each point p j,s in the whole frame K j and relative target areaCoordinates of (c)Has the following mapping relation:
Optical flow in target area And tracking areaIn the above, the optical flow prediction result is V j→i, and the sparse matching result and the optical flow prediction result should be consistent, and the difference between the two is calculated by the following formula:
if d is less than or equal to gamma, and gamma is a third threshold value, the tracking area Target point set onP ji,s is obtained by the following method:
Otherwise the first set of parameters is selected, Not participate inIs calculated by the computer.
Merging target point sets on each tracking areaObtaining a final target point setExpressed as:
wherein lambda ji is the weight, and =1, Λ ji is proportional to n ji.
It should be noted that in practice, the embodiment of the present invention may obtain relatively accurate results by selecting one or two keyframes with the highest similarity. By judgingAnd if each pixel point in the image is on the image, covering the complete tracking path or the local tracking path on the current frame image, so that real-time guidance can be provided for doctors.
In step S36, when the current frame image is determined to be a new key frame based on a preset condition, the current frame image is put into the key frame pool, wherein the preset condition includes:
(1) A complete tracking path is obtained compared to the surgical path, expressed as The keyframes must contain the complete surgical path to have sufficient information.
(2) Acquiring the final two-dimensional target point setTracking area whereThe center of the tracking areaAlways within the preset range of the current frame image F i, expressed asWhere l is a fourth threshold. In order to eliminate the redundancy of the device,AndNeeds to meet the requirements ofΗ 1 is the fifth threshold and η 1 <1. This condition is intended to maintain a certain difference between the target area and the key frame, the target area on the key frame being as far as possible offset from the edge of the image.
(3) If the similar keyframes in the best template are greater than one, the distance places the last new keyframe into the keyframe pool, and the first preset number of image frames have been exceeded.
Illustratively, the first predetermined number here is ten, denoted as card (K best) >1, i-f >10, where f is the index of the most recently added keyframe, inserted from the last keyframe to the now 10 frames.
If the similar key frame in the optimal template is equal to one, placing the last new key frame into the key frame pool, wherein the distance exceeds a second preset number of image frames, and the ratio of the length of the tracking path to the length of the operation path is consistent with the scale change of the image frames where the tracking path and the operation path are positioned;
The second preset number is here, for example, four, correspondingly denoted as card (K best) =1, i-f >4, and Wherein η 2、η3 is a sixth threshold and a seventh threshold respectively, and the calculation formulas of η 2<1<η3, r are r=max { H 0,h1},h0、h1 respectively represent the first two diagonal elements of the homography matrix H ji of F i and K j, and H 0、h1 respectively corresponds to the scaling factors of the picture in the horizontal direction and the vertical direction.
(4) The maximum capacity of the keyframe pool is not exceeded, and n max,nmax, expressed as card (P). Ltoreq.n, is the maximum capacity of keyframe pool P, where n max is sixteen, for example. To prevent the key frames from repeatedly adding unnecessary computation, the key frames are considered to have saved key information at various perspectives when a certain capacity is reached.
Therefore, the embodiment of the invention can also process the condition of target loss and re-tracking, and realize the robustness of tracking.
The embodiment of the invention successfully obtains the tracking path represented by the two-dimensional target point set.
In step S4, a three-dimensional surgical path is obtained based on the three-dimensional point cloud and the two-dimensional target point set.
In the step, a three-dimensional point cloud V d based on the depth estimation of the laparoscopic image is used, and a two-dimensional target point set obtained by combining a virtual mark dynamic tracking method is usedA three-dimensional virtual marker point set V r can be obtained.
Wherein, pi (∙) is a function of obtaining a three-dimensional virtual mark according to the three-dimensional point cloud and the two-dimensional target point set, V r∈Vd.
In step S5, searching a point cloud position closest to the tail end position of the surgical instrument on the three-dimensional surgical path, introducing an additional investigation radius, obtaining an artificial vector of the tail end position of the surgical instrument, and establishing a guiding type virtual force field along the three-dimensional surgical path based on the artificial vector so as to guide a doctor to operate the surgical robot system through a force feedback mechanism, wherein the method comprises the following steps:
searching the end position of the surgical instrument on the three-dimensional surgical path by using a K-nearest neighbor algorithm Nearest point cloud locationThe relative distance x d between the two points is obtained, thereby judging whether the surgical instrument deviates from the preset surgical path.
Providing that all points on the three-dimensional surgical path are contained in a set of virtual marker points V r for the surgical instrument tip positionDistance on three-dimensional surgical pathPoint-closest point cloud locationFind the set of sub-points of V r WhereinTo additionally investigate the radius;
The surgical instrument end position The corresponding artificial vector is expressed in the form:
The artificial vector length function is calculated as follows:
Wherein, Is the position of the tail end of the surgical instrumentThe distance between the virtual mark point v and the virtual mark point v, R and R are upper and lower boundaries of the distance respectively, when the distance is smaller than or equal to R, the maximum weight q max is allocated to the virtual mark point v, when the distance is larger than R and smaller than R, the weight of the virtual mark point v is between q max and 0, and when the distance is larger than or equal to R, the weight of the virtual mark point v is 0; The representation is composed of Vector pointing to point v, and card (∙) represents the cardinality or number of elements of the collection.
Then, based on the artificial vector, establishing a virtual force field containing attractive force and viscous resistance along the three-dimensional operation path, namely a guiding type virtual clamp, so as to guide a doctor to operate the operation robot system through a force feedback mechanism, wherein the guiding type virtual force field is expressed as:
Wherein K vf represents the proportionality coefficient of attraction force and artificial vector, D vf(xd) represents the damping coefficient Representation ofIs a function of the speed of the machine.
Thus, the embodiment of the invention establishes a man-machine touch force interaction mechanism of the doctor and the robot through a force control technology, establishes a guiding type virtual force field in space, realizes accurate guiding of the movement track of the doctor, and ensures that the doctor can operate the surgical robot according to a preset track.
Example 2:
the embodiment of the invention provides a surgical operation force feedback guiding system based on virtual mark tracking and instrument pose, which comprises the following components:
the depth estimation module is used for acquiring a three-dimensional point cloud of the in-vivo scene based on the depth estimation;
the pose estimation module is used for acquiring the optimal 6D pose of the instrument of the current frame image based on a 3D Gaussian sputtering instrument 6D pose estimation method so as to estimate the tail end position of the surgical instrument;
The mark tracking module is used for tracking the operation path marked on the first image frame by a doctor based on a virtual mark dynamic tracking method guided by the key frame to acquire a tracking path of the current frame image, wherein the tracking path is represented by a two-dimensional target point set;
The path conversion module is used for acquiring a three-dimensional operation path based on the three-dimensional point cloud and the two-dimensional target point set;
The force field establishment module is used for searching a point cloud position closest to the tail end position of the surgical instrument on the three-dimensional surgical path, introducing an additional investigation radius, acquiring an artificial vector of the tail end position of the surgical instrument, and establishing a guiding type virtual force field along the three-dimensional surgical path based on the artificial vector so as to guide a doctor to operate the surgical robot system through a force feedback mechanism.
Example 3:
Embodiments of the present invention provide a storage medium storing a computer program for surgical operation force feedback guidance based on virtual marker tracking and instrument pose, wherein the computer program causes a computer to execute the surgical operation force feedback guidance method as described above.
Example 4:
the embodiment of the invention provides electronic equipment, which comprises:
The surgical force feedback guidance system comprises one or more processors, memory, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the programs comprising instructions for performing the surgical force feedback guidance method as described above.
It may be understood that the surgical operation force feedback guiding system, the storage medium and the electronic device based on the virtual mark tracking and the instrument pose provided by the embodiment of the present invention correspond to the surgical operation force feedback guiding method based on the virtual mark tracking and the instrument pose provided by the embodiment of the present invention, and the explanation, the examples, the beneficial effects and other parts of the relevant content may refer to the corresponding parts in the surgical operation force feedback guiding method, which are not repeated herein.
In summary, compared with the prior art, the method has the following beneficial effects:
1. In the embodiment of the invention, the real-time three-dimensional point cloud reconstruction is carried out on the in-vivo scene, the virtual mark dynamic tracking method based on key frame guidance is used for tracking the operation path marked on the first image frame by a doctor, the three-dimensional operation path is obtained based on the real-time three-dimensional point cloud, the 6D pose method of the operation instrument based on 3D Gaussian tracking is combined, and finally, the guiding type virtual clamp along the dynamic three-dimensional operation path is established, so that the doctor can be effectively helped to complete the operation along the preset operation path through a force feedback mechanism, and the man-machine co-fusion operation conforming to the operation intention of the doctor is realized.
2. According to the embodiment of the invention, no additional equipment or data is needed, the complete three-dimensional reconstruction of the surgical scene is realized by only using the binocular endoscope image, and the surgical planning and navigation are assisted by the real-time estimated three-dimensional information of the instrument and the tissue. Based on the research thought of reconstruction-optimization, the pose estimation of the instrument is estimated by aligning the reconstructed instrument model and the image features, so that the 6D pose of the instrument can be accurately and robustly recovered in a complex operation scene. And the frame based on 3D Gaussian sputtering realizes the estimation of the instrument pose, and can realize the quick recovery of the instrument pose in a few iteration times
3. The embodiment of the invention realizes the pixel-level precision tracking of the two-dimensional operation path on the endoscope image, and ensures the intuitiveness and accuracy of information transmission. When the operation path is tracked, two-stage matching from thick to thin is adopted, similar key frames are determined through traversing of a key frame pool and sparse matching, a local area is determined by utilizing a matching result, and then the local area is tracked, so that error tracking is reduced. And the diversity of the key frame pool is enriched by dynamically updating the key frame pool, so that the robustness and accuracy of dynamic tracking of the virtual mark are further improved.
4. According to the embodiment of the invention, on the basis of improving the operation precision, a doctor of a main knife does not need to wear other equipment, so that the interference to the operation of the doctor is prevented.
It is noted that relational terms such as first and second, and the like are used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Moreover, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one does not exclude the presence of other like elements in a process, method, article, or apparatus that comprises an element.
The foregoing embodiments are merely for illustrating the technical solution of the present invention, but not for limiting the same, and although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those skilled in the art that modifications may be made to the technical solution described in the foregoing embodiments or equivalents may be substituted for parts of the technical features thereof, and that such modifications or substitutions do not depart from the spirit and scope of the technical solution of the embodiments of the present invention in essence.

Claims (10)

1.一种基于虚拟标记追踪与器械位姿的手术操作力反馈引导方法,其特征在于,包括:1. A surgical operation force feedback guidance method based on virtual marker tracking and instrument posture, characterized by comprising: 基于深度估计获取体内场景的三维点云;Acquire the three-dimensional point cloud of the internal scene based on depth estimation; 基于3D高斯溅射的器械6D位姿估计方法,获取当前帧图像的器械最优6D位姿,以估计手术器械末端位置;The 6D pose estimation method of the instrument based on 3D Gaussian sputtering obtains the optimal 6D pose of the instrument in the current frame image to estimate the end position of the surgical instrument; 基于关键帧引导的虚拟标记动态追踪方法,对医生在首个图像帧上标注的手术路径进行追踪,获取当前帧图像的追踪路径;其中所述追踪路径采用二维目标点集进行表示;Based on the key frame-guided virtual marker dynamic tracking method, the surgical path marked by the doctor on the first image frame is tracked to obtain the tracking path of the current frame image; wherein the tracking path is represented by a two-dimensional target point set; 基于所述三维点云与二维目标点集,获取三维手术路径;Based on the three-dimensional point cloud and the two-dimensional target point set, obtaining a three-dimensional surgical path; 搜索所述三维手术路径上距离手术器械末端位置最近的点云位置,引入额外考察半径,获取所述手术器械末端位置的人工矢量,并基于所述人工矢量,建立沿着所述三维手术路径的引导型虚拟力场,以通过力反馈机制引导医生操控手术机器人系统。The point cloud position closest to the end position of the surgical instrument on the three-dimensional surgical path is searched, an additional investigation radius is introduced, an artificial vector of the end position of the surgical instrument is obtained, and based on the artificial vector, a guiding virtual force field along the three-dimensional surgical path is established to guide the doctor to control the surgical robot system through a force feedback mechanism. 2.如权利要求1所述的手术操作力反馈引导方法,其特征在于,所述基于3D高斯溅射的器械6D位姿估计方法,获取当前帧图像的器械最优6D位姿,以估计手术器械末端位置;包括:2. The surgical operation force feedback guidance method according to claim 1, characterized in that the instrument 6D posture estimation method based on 3D Gaussian sputtering obtains the optimal 6D posture of the instrument in the current frame image to estimate the end position of the surgical instrument; comprising: 读取当前帧图像,通过语义分割算法获取器械掩膜;以及利用双目深度估计算法对当前帧图像进行深度估计,并结合所述器械掩膜分别获取组织深度图和器械深度图;Reading the current frame image, obtaining the device mask through a semantic segmentation algorithm; and performing depth estimation on the current frame image using a binocular depth estimation algorithm, and respectively obtaining a tissue depth map and a device depth map in combination with the device mask; 初始化器械位姿,利用预先重构的器械高斯模型进行3D高斯溅射,获取对应位姿下的视图、投影掩膜以及深度图;Initialize the device pose, use the pre-reconstructed device Gaussian model to perform 3D Gaussian sputtering, and obtain the view, projection mask and depth map under the corresponding pose; 根据渲染结果与当前帧图像及其器械掩膜、器械深度图,构建总的损失函数;并通过反向传播算法更新和学习器械位姿,最小化所述总的损失函数,获取当前帧图像的器械最优6D位姿;According to the rendering result, the current frame image, the device mask, and the device depth map, a total loss function is constructed; and the device posture is updated and learned through the back propagation algorithm to minimize the total loss function and obtain the optimal 6D posture of the device in the current frame image; 基于当前帧图像的器械最优6D位姿,确定器械末端对应高斯核的位置,并作为所述手术器械末端位置。Based on the optimal 6D posture of the instrument in the current frame image, the position of the Gaussian kernel corresponding to the end of the instrument is determined and used as the end position of the surgical instrument. 3.如权利要求2所述的手术操作力反馈引导方法,其特征在于,3. The surgical operation force feedback guidance method according to claim 2, characterized in that: 所述初始化器械位姿;包括:The initialization device posture comprises: 对于第一帧图像,结合相机模型获取器械高斯模型在图像平面上的渲染视图,通过移动腔镜使得图像与渲染视图所在区域相重合,并将此时的器械位姿作为第一帧图像的初始器械位姿;For the first frame of image, a rendering view of the instrument Gaussian model on the image plane is obtained in combination with the camera model, the laparoscope is moved so that the image coincides with the area where the rendering view is located, and the instrument posture at this time is used as the initial instrument posture of the first frame of image; 对于除了第一帧以外的图像,基于前一帧图像的最优器械位姿和连续两帧的光流信息,对当前帧图像的器械位姿进行初始化;For images other than the first frame, the device pose of the current frame is initialized based on the optimal device pose of the previous frame and the optical flow information of two consecutive frames; 和/或and/or 所述器械高斯模型的重构过程;包括:The reconstruction process of the Gaussian model of the device includes: 获取若干已知相机位姿且包含器械全部视角的图像,以重构器械的稀疏点云;Acquire several images with known camera poses and containing all views of the device to reconstruct a sparse point cloud of the device; 将所述稀疏点云的每个点初始化为一个高斯,获取初始化的器械高斯模型;Initialize each point of the sparse point cloud as a Gaussian to obtain an initialized Gaussian model of the instrument; 结合前述若干已知相机位姿且包含器械全部视角的图像构成的图像序列及其对应的相机位姿,对所述初始化的器械高斯模型进行迭代优化,获取所述重构的器械高斯模型;Combining the aforementioned image sequence consisting of the images with known camera poses and containing all viewing angles of the instrument and their corresponding camera poses, iteratively optimizing the initialized instrument Gaussian model to obtain the reconstructed instrument Gaussian model; 和/或and/or 所述结合前述若干已知相机位姿且包含器械全部视角的图像构成的图像序列及其对应的相机位姿,对所述初始化的器械高斯模型进行迭代优化,获取所述重构的器械高斯模型;包括:The image sequence composed of the images of the aforementioned known camera poses and containing all viewing angles of the instrument and the corresponding camera poses is combined to iteratively optimize the initialized instrument Gaussian model to obtain the reconstructed instrument Gaussian model; comprising: 构建3D高斯渲染视图和仅保留器械部分的原图像之间的色彩一致性损失函数,如下所示:The color consistency loss function between the 3D Gaussian rendering view and the original image that only retains the instrument part is constructed as follows: 其中,分别表示任意一个高斯的中心位置、旋转、大小;L为图像序列包括的图像数量;L 1表示一范数损失;表示相机位姿为R l时进行3D高斯可微渲染图像的过程;G i表示第i个高斯点,I为高斯的个数;表示源自图像序列中仅保留器械部分的原图像;in, Respectively represent the center position, rotation, and size of any Gaussian; , L is the number of images in the image sequence; L 1 represents a norm loss; represents the process of 3D Gaussian differentiable rendering of the image when the camera pose is R l ; Gi represents the i- th Gaussian point, , I is the number of Gaussians; It indicates the original image from the image sequence where only the instrument part is retained; 引入多区域对齐损失函数,约束重构的器械高斯模型上每个铰接部分在图像平面的投影形状和原图像上的器械掩膜M lc分别对齐;所述多区域对齐损失函数,如下所示:A multi-region alignment loss function is introduced to constrain the projection shape of each articulated part on the reconstructed instrument Gaussian model on the image plane and the instrument mask M lc on the original image to be aligned respectively; the multi-region alignment loss function is as follows: 其中,C表示器械的铰接部件数量;表示相机位姿为R l时进行3D高斯可微渲染投影图的过程;in, , C represents the number of articulated parts of the device; It represents the process of 3D Gaussian differentiable rendering projection when the camera pose is R l ; 所述器械高斯模型重构过程的总的损失函数为:The total loss function of the instrument Gaussian model reconstruction process is: 其中,α表示相应的权重系数;Among them, α represents the corresponding weight coefficient; 通过最小化总损失函数L recons,获取所述重构的器械高斯模型,如下所示:By minimizing the total loss function L recons , the reconstructed instrument Gaussian model is obtained , as shown below: 其中,表示第i个重构的高斯点;表示函数值最小时变量的取值;in, represents the i- th reconstructed Gaussian point; Indicates the value of the variable when the function value is minimum; 和/或and/or 所述根据渲染结果与当前帧图像及其器械掩膜、器械深度图,构建总的损失函数;并通过反向传播算法更新和学习器械位姿,最小化所述总的损失函数,获取当前帧图像的器械最优6D位姿;包括:The method constructs a total loss function based on the rendering result, the current frame image, its device mask, and the device depth map; and updates and learns the device posture through a back propagation algorithm to minimize the total loss function and obtain the optimal 6D posture of the device in the current frame image; including: 构建3D高斯渲染视图和当前帧图像之间的色彩一致性损失函数,如下所示:Construct the color consistency loss function between the 3D Gaussian rendering view and the current frame image as follows: 其中,t表示当前时刻;T t表示当前帧器械位姿;表示渲染视图;表示进行3D高斯可微渲染图像的过程;M t表示当前图像分割的器械掩膜;Where, t represents the current time; T t represents the device posture in the current frame; Represents a rendered view; , represents the process of performing 3D Gaussian differentiable rendering of the image; M t represents the instrument mask of the current image segmentation; 构建3D高斯可微渲染投影图和当前帧图像的器械掩膜之间的区域对齐损失函数,如下所示:Construct the regional alignment loss function between the 3D Gaussian differentiable rendering projection image and the device mask of the current frame image as follows: 其中,分别表示器械高斯模型上第c个铰接部分在图像平面的投影形状和当前帧图像的器械掩膜;表示进行3D高斯可微渲染投影图的过程;in, They respectively represent the projection shape of the cth hinged part on the Gaussian model of the device on the image plane and the device mask of the current frame image; , Represents the process of performing 3D Gaussian differentiable rendering projection; ; 构建渲染的深度图和前帧图像的器械深度图之间的深度一致性损失,如下所述:Construct the depth consistency loss between the rendered depth map and the device depth map of the previous frame as follows: 其中,分别表示渲染的深度图和前帧图像的器械深度图;表示渲染器械高斯模型深度图的过程;in, Respectively represent the rendered depth map and the device depth map of the previous frame image; , Represents the process of rendering the depth map of the Gaussian model of the device; 构建总的损失函数为:The total loss function is constructed as: 其中,α 1α 2α 3分别表示相应的权重系数;Among them, α 1 , α 2 , and α 3 represent the corresponding weight coefficients respectively; 之后通过反向传播算法对器械位姿T t迭代更新,以最小化L pose,获取当前帧图像的器械最优6D位姿估计结果,如下所示:Then, the device pose T t is iteratively updated through the back propagation algorithm to minimize L pose and obtain the optimal 6D pose estimation result of the device in the current frame image. , as shown below: 4.如权利要求1所述的手术操作力反馈引导方法,其特征在于,所述基于关键帧引导的虚拟标记动态追踪方法,对医生在首个图像帧上标注的手术路径进行追踪,获取当前帧图像的追踪路径;包括:4. The surgical operation force feedback guidance method according to claim 1, characterized in that the virtual marker dynamic tracking method based on key frame guidance tracks the surgical path marked by the doctor on the first image frame to obtain the tracking path of the current frame image; comprising: 接收医生在首个图像帧上标注的手术路径;其中所述手术路径采用像素级别的二维目标点集进行表示;Receiving a surgical path marked by a doctor on a first image frame, wherein the surgical path is represented by a two-dimensional target point set at a pixel level; 获取首个图像帧上的二维目标点集所在的目标区域,并利用光流算法预测各个点的轨迹,获取在后多个图像帧上的二维目标点集及其所在的目标区域,以初始化关键帧池;Obtain the target area where the two-dimensional target point set on the first image frame is located, and use the optical flow algorithm to predict the trajectory of each point, obtain the two-dimensional target point set and the target area where it is located on the subsequent multiple image frames to initialize the key frame pool; 针对当前帧图像,遍历所述关键帧池,利用稀疏匹配方法进行相似关键帧查询,并基于查询到的相似关键帧与当前帧图像之间的稀疏匹配结果,获取最佳模板;For the current frame image, the key frame pool is traversed, similar key frames are queried using a sparse matching method, and the best template is obtained based on the sparse matching results between the queried similar key frames and the current frame image; 在所述最佳模板不为空集时,基于所述最佳模板中每一相似关键帧上的目标区域,获取当前帧图像上的追踪区域;When the best template is not an empty set, obtaining a tracking area on the current frame image based on the target area on each similar key frame in the best template; 基于每一所述目标区域中的二维目标点集,利用光流算法预测各个点的运动轨迹,获取相应追踪区域上的二维目标点集;合并每一所述追踪区域上的二维目标点集,获取最终的二维目标点集,并作为当前帧图像上的追踪路径;Based on the two-dimensional target point set in each of the target areas, the motion trajectory of each point is predicted using an optical flow algorithm to obtain a two-dimensional target point set on the corresponding tracking area; the two-dimensional target point set on each of the tracking areas is merged to obtain a final two-dimensional target point set, and the final two-dimensional target point set is used as the tracking path on the current frame image; 基于预设的条件,判定当前帧图像为新的关键帧时,将其放入所述关键帧池中;其中所述预设的条件包括:When the current frame image is determined to be a new key frame based on a preset condition, it is placed in the key frame pool; wherein the preset condition includes: (1)相较于所述手术路径,获取了完整的追踪路径;(1) Compared with the surgical path, a complete tracking path was obtained; (2)获取所述最终的二维目标点集所在的追踪区域,该追踪区域的中心始终在当前帧图像的预设范围内;(2) obtaining a tracking area where the final two-dimensional target point set is located, wherein the center of the tracking area is always within a preset range of the current frame image; (3)若所述最佳模板中的相似关键帧大于一,距离将上一新的关键帧放入所述关键帧池,已过去超过第一预设数量的图像帧;(3) if the number of similar key frames in the best template is greater than one, the number of image frames that have passed since the last new key frame was put into the key frame pool exceeds a first preset number; 若所述最佳模板中的相似关键帧等于一,距离将上一新的关键帧放入所述关键帧池,已过去超过第二预设数量的图像帧,且所述追踪路径与手术路径的长度之比,与各自所在图像帧的尺度变化保持一致;If the similar keyframe in the best template is equal to one, more than a second preset number of image frames have passed since the last new keyframe was put into the keyframe pool, and the ratio of the length of the tracking path to the surgical path is consistent with the scale change of the respective image frames; (4)未超出所述关键帧池的最大容量。(4) The maximum capacity of the keyframe pool is not exceeded. 5.如权利要求4所述的手术操作力反馈引导方法,其特征在于,5. The surgical operation force feedback guidance method according to claim 4, characterized in that: 所述初始化关键帧池;具体包括:The initialization key frame pool specifically includes: 采用表示关键帧,其中j为关键帧池中的关键帧索引;表示RGB值,W、H分别表示宽度和高度;表示目标区域;分别表示利用基于描述符的稀疏匹配方法提取的中的特征点和特征描述;表示二维目标点集;use represents a key frame, where j is the key frame index in the key frame pool; Represents RGB value, W and H represent width and height respectively; Indicates the target area; , They represent the extracted Feature points and feature descriptions in; represents a two-dimensional target point set; 将第0个图像帧F 0指定为关键帧K 0K,定义其上的二维目标点集为表示第s个像素点的坐标,n表示像素点的个数,表示实数集;The 0th image frame F 0 is designated as the key frame K 0K , and the two-dimensional target point set on it is defined as , represents the coordinates of the sth pixel, n represents the number of pixels, represents the set of real numbers; 为中心的区域由下式生成:by The central area Generated by: 其中,中心点中所有像素点的横坐标的最大值和最小值的算数平均值,中所有像素点的纵坐标的最大值和最小值的算数平均值;L为目标区域的边长,L不小于中所有像素点的横坐标和纵坐标的最大值和最小值的差;Among them, the center point , for The arithmetic mean of the maximum and minimum values of the horizontal coordinates of all pixels in , for The arithmetic mean of the maximum and minimum values of the ordinates of all pixels in the target area; L is the side length of the target area, L is not less than The difference between the maximum and minimum values of the horizontal and vertical coordinates of all pixels in; 定义目标区域存储四个顶点和中心点,表示为,其中可知,表示的四个顶点:左上,右上,左下,右下;Define the target area Store four vertices and the center point , expressed as ,in Depend on It can be seen that The four vertices of: upper left, upper right, lower left, lower right; 利用光流算法预测在后两个图像帧F 1F 2的运动V 0→1V 0→2,并根据光流结果计算所述手术路径上的每个像素点F 1F 2上的轨迹:The optical flow algorithm is used to predict the motion V 0→1 ∈ in the next two image frames F 1 and F 2 , V 0→2 , and calculate each pixel point on the surgical path according to the optical flow result Trajectory on F1 and F2 : 其中,Interp(∙)表示采样过程,得到相应的目标点集,并参考公式(1),获取目标区域Among them, Interp (∙) represents the sampling process, and the corresponding target point set is obtained and , and refer to formula (1) to obtain the target area and ; F 1F 2对应记为关键帧K 1 、K 2,与K 0一起构成初始的关键帧池P={ K 0, K 1, K 2}; F 1 and F 2 are recorded as key frames K 1 and K 2 respectively, and together with K 0, they constitute the initial key frame pool P = { K 0 , K 1 , K 2 }; 和/或and/or 所述针对当前帧图像,遍历所述关键帧池,利用稀疏匹配方法进行相似关键帧查询,并基于查询到的相似关键帧与当前帧图像之间的稀疏匹配结果,获取最佳模板;具体包括:The method of traversing the key frame pool for the current frame image, performing a similar key frame query using a sparse matching method, and obtaining an optimal template based on a sparse matching result between the queried similar key frame and the current frame image, specifically includes: 获取当前帧图像F i的特征点与特征描述;Get the feature points and feature description of the current frame image Fi ; 遍历所述关键帧池中的每一个关键帧K j,利用稀疏匹配方法获取F iK j之间的匹配点;Traversing each key frame K j in the key frame pool, and obtaining matching points between F i and K j using a sparse matching method; 统计K j上目标区域的匹配点个数n ji,并选取n ji最大的三个关键帧作为相似关键帧K simi={K a ,K b ,K c }⊆PCount the number of matching points n ji in the target area on K j , and select the three key frames with the largest n ji as similar key frames K simi ={ Ka , K b , K c }⊆ P , 采用如下公式获取最佳模板:The best template is obtained using the following formula: 其中,K j表示关键帧池中的第j个关键帧,j的取值为a、bc;n ji表示K j上的目标区域与当前帧图像F i之间的匹配点个数;σ为第一阈值;K simi表示查询到的所有相似关键帧;Where Kj represents the jth keyframe in the keyframe pool, j is a, b or c; nji represents the target area on Kj The number of matching points between the current frame image F i ; σ is the first threshold; K simi represents all similar key frames queried; 和/或and/or 所述基于所述最佳模板中每一相似关键帧上的目标区域,获取当前帧图像上的追踪区域;具体包括:The step of obtaining a tracking area on the current frame image based on the target area on each similar key frame in the best template specifically includes: 基于K jK bestF i之间的匹配点,利用RANSAC估计单应性矩阵H jiH ji的精度由得到,n in为内点的个数;Based on the matching points between K jK best and F i , RANSAC is used to estimate the homography matrix H ji . The accuracy of H ji is given by We get, n in is the number of interior points; ≥λ,0<λ<1λ为第二阈值,则对目标区域上的四个顶点进行单应变换,得到F i上的对应位置when ≥λ , 0<λ<1 , λ is the second threshold, then for the target area The four vertices on Perform homography transformation and get The corresponding position on F i : 否则,利用邻域相似变换方法进行顶点坐标转换:Otherwise, the vertex coordinates are transformed using the neighborhood similarity transformation method: 从稀疏匹配结果中提取F iK j的一组匹配x imy jm是表示特征点位置的二维列向量;Extract a set of matches of F i and K j from the sparse matching results xi m and y jm is a two-dimensional column vector representing the position of the feature point; 搜索目标区域顶点N个近邻,获取对应的匹配点,以得到点F i上的对应位置Search target area vertices The N nearest neighbors of , get the corresponding matching point , to get the point The corresponding position on F i : 其中w t之间的距离成反比;in , w t and and is inversely proportional to the distance between them; 基于顶点坐标,确定一个初始区域,表示为:Based on vertex coordinates , determine an initial region, expressed as: 其中,分别表示F i上的对应位置的横坐标和纵坐标; u min 、u max分别表示四个顶点的横坐标的最大值和最小值;v min 、v max分别表示四个顶点的纵坐标的最大值和最小值;in, , Respectively The corresponding position on F i The horizontal and vertical coordinates of the four vertices are as follows: u min and u max represent the maximum and minimum values of the horizontal coordinates of the four vertices, respectively; v min and v max represent the maximum and minimum values of the vertical coordinates of the four vertices, respectively; 基于所述初始区域的四个顶点的坐标,确定追踪区域的中心点Based on the coordinates of the four vertices of the initial area, determine the tracking area The center point ; 其中, 分别表示追踪区域的中心点的横坐标和纵坐标;in, , Respectively represent the tracking area The center point The horizontal and vertical coordinates of 参考公式(1),获取当前帧图像上的追踪区域Refer to formula (1) to obtain the tracking area on the current frame image ; 和/或and/or 所述基于每一所述目标区域中的二维目标点集,利用光流算法预测各个点的运动轨迹,获取相应追踪区域上的二维目标点集;合并每一所述追踪区域上的二维目标点集,获取最终的二维目标点集;具体包括:The method of predicting the motion trajectory of each point based on the two-dimensional target point set in each target area by using an optical flow algorithm to obtain a two-dimensional target point set on the corresponding tracking area; merging the two-dimensional target point set on each tracking area to obtain a final two-dimensional target point set specifically includes: 二维目标点集中的每个点p j,s在整个帧K j上的坐标与相对目标区域的坐标具有如下映射关系:2D target point set The coordinates of each point pj ,s in the entire frame Kj and the relative target area Coordinates The mapping relationship is as follows: 光流在目标区域和追踪区域上进行,光流预测结果为V j→i,所述稀疏匹配结果与光流预测结果应一致,用下式计算两者之差:Optical flow in the target area and tracking area The optical flow prediction result is V j→i . The sparse matching result should be consistent with the optical flow prediction result. The difference between the two is calculated using the following formula: dγγ为第三阈值,则追踪区域上的目标点集中的每个点p ji,s通过以下方法获得:If dγ , γ is the third threshold, then the tracking area The target point set on Each point p ji,s in is obtained by the following method: 否则,不参与的计算;otherwise, Not participating Calculation of 合并每一所述追踪区域上的二维目标点集,获取最终的二维目标点集,表示为:Merge the two-dimensional target point set on each tracking area , get the final two-dimensional target point set , expressed as: 其中,λ ji为权重,且=1,λ jin ji成正比。Among them, λ ji is the weight, and =1, λ ji is proportional to n ji . 6.如权利要求1所述的手术操作力反馈引导方法,其特征在于,所述搜索所述三维手术路径上距离手术器械末端位置最近的点云位置,引入额外考察半径,获取所述手术器械末端位置的人工矢量;包括:6. The surgical operation force feedback guidance method according to claim 1, characterized in that the step of searching for the point cloud position closest to the end position of the surgical instrument on the three-dimensional surgical path, introducing an additional investigation radius, and obtaining an artificial vector of the end position of the surgical instrument comprises: 利用K近邻算法,搜索所述三维手术路径上距离手术器械末端位置最近的点云位置,得到两点之间相对距离x dUsing the K nearest neighbor algorithm, search for the distance from the end of the surgical instrument on the three-dimensional surgical path Nearest point cloud location , get the relative distance x d between the two points; 设所述三维手术路径上的所有点包含在虚拟标记点集V r中,对于所述手术器械末端位置以及三维手术路径上距离点最近的点云位置,求取V r的子点集;其中为额外考察半径;Assume that all points on the three-dimensional surgical path are included in the virtual marker point set V r . and the distance along the 3D surgical path Point cloud location closest to the point , find the sub-point set of V r ;in For additional investigation radius; 则所述手术器械末端位置对应的人工矢量表示为如下形式:The end position of the surgical instrument The corresponding artificial vector is expressed as follows: 为人工矢量长度函数,计算如下: is an artificial vector length function, calculated as follows: 其中,为手术器械末端位置与虚拟标记点v之间的距离,rR分别为距离的上下界,当距离小于或等于r时,为虚拟标记点v分配最大的权重q max,当距离大于r且小于R时,令虚拟标记点v的权重处于q max至0之间,当距离大于或等于R时,令虚拟标记点v的权重为0;表示由指向点v的向量;card(∙)表示集合的基数或者元素个数。in, End position of surgical instrument The distance between the virtual marker point v , r and R are the upper and lower bounds of the distance respectively. When the distance is less than or equal to r , the virtual marker point v is assigned the maximum weight q max . When the distance is greater than r and less than R , the weight of the virtual marker point v is between q max and 0. When the distance is greater than or equal to R , the weight of the virtual marker point v is 0. Indicated by A vector pointing to a point v ; card (∙) represents the cardinality or number of elements in a set. 7.如权利要求6所述的手术操作力反馈引导方法,其特征在于,所述引导型虚拟力场表示为:7. The surgical operation force feedback guiding method according to claim 6, wherein the guiding virtual force field is represented as: 其中,K vf表示引力与人工矢量的比例系数,D vf(x d )表示阻尼系数表示的速度。Where K vf represents the proportionality coefficient between gravity and artificial vector, and D vf ( x d ) represents the damping coefficient express speed. 8.一种基于虚拟标记追踪与器械位姿的手术操作力反馈引导系统,其特征在于,包括:8. A surgical operation force feedback guidance system based on virtual marker tracking and instrument posture, characterized by comprising: 深度估计模块,用于基于深度估计获取体内场景的三维点云;A depth estimation module, used to obtain a three-dimensional point cloud of an in-vivo scene based on depth estimation; 位姿估计模块,用于基于3D高斯溅射的器械6D位姿估计方法,获取当前帧图像的器械最优6D位姿,以估计手术器械末端位置;The pose estimation module is used for the instrument 6D pose estimation method based on 3D Gaussian sputtering to obtain the optimal 6D pose of the instrument in the current frame image to estimate the end position of the surgical instrument; 标记追踪模块,用于基于关键帧引导的虚拟标记动态追踪方法,对医生在首个图像帧上标注的手术路径进行追踪,获取当前帧图像的追踪路径;其中所述追踪路径采用二维目标点集进行表示;A marker tracking module is used to track the surgical path marked by the doctor on the first image frame based on a virtual marker dynamic tracking method guided by a key frame, and obtain the tracking path of the current frame image; wherein the tracking path is represented by a two-dimensional target point set; 路径转换模块,用于基于所述三维点云与二维目标点集,获取三维手术路径;A path conversion module, used for acquiring a three-dimensional surgical path based on the three-dimensional point cloud and the two-dimensional target point set; 力场建立模块,用于搜索所述三维手术路径上距离手术器械末端位置最近的点云位置,引入额外考察半径,获取所述手术器械末端位置的人工矢量,并基于所述人工矢量,建立沿着所述三维手术路径的引导型虚拟力场,以通过力反馈机制引导医生操控手术机器人系统。The force field establishment module is used to search for the point cloud position on the three-dimensional surgical path that is closest to the end position of the surgical instrument, introduce an additional investigation radius, obtain an artificial vector of the end position of the surgical instrument, and establish a guiding virtual force field along the three-dimensional surgical path based on the artificial vector, so as to guide the doctor to control the surgical robot system through a force feedback mechanism. 9.一种存储介质,其特征在于,其存储有用于基于虚拟标记追踪与器械位姿的手术操作力反馈引导的计算机程序,其中,所述计算机程序使得计算机执行如权利要求1~6任一项所述的手术操作力反馈引导方法。9. A storage medium, characterized in that it stores a computer program for force feedback guidance of surgical operations based on virtual marker tracking and instrument posture, wherein the computer program enables a computer to execute the surgical operation force feedback guidance method as described in any one of claims 1 to 6. 10.一种电子设备,其特征在于,包括:10. An electronic device, comprising: 一个或多个处理器;存储器;以及一个或多个程序,其中所述一个或多个程序被存储在所述存储器中,并且被配置成由所述一个或多个处理器执行,所述程序包括用于执行如权利要求1~6任一项所述的手术操作力反馈引导方法。One or more processors; a memory; and one or more programs, wherein the one or more programs are stored in the memory and are configured to be executed by the one or more processors, and the programs include a method for executing a surgical operation force feedback guidance method as described in any one of claims 1 to 6.
CN202411580088.9A 2024-11-07 2024-11-07 Surgical operation force feedback guiding method based on virtual mark tracking and instrument pose Active CN119055358B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202411580088.9A CN119055358B (en) 2024-11-07 2024-11-07 Surgical operation force feedback guiding method based on virtual mark tracking and instrument pose

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202411580088.9A CN119055358B (en) 2024-11-07 2024-11-07 Surgical operation force feedback guiding method based on virtual mark tracking and instrument pose

Publications (2)

Publication Number Publication Date
CN119055358A true CN119055358A (en) 2024-12-03
CN119055358B CN119055358B (en) 2025-02-14

Family

ID=93634010

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202411580088.9A Active CN119055358B (en) 2024-11-07 2024-11-07 Surgical operation force feedback guiding method based on virtual mark tracking and instrument pose

Country Status (1)

Country Link
CN (1) CN119055358B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119970235A (en) * 2025-01-15 2025-05-13 东南大学 A force feedback device and method for assisting surgical robot clamping operation
CN120070775A (en) * 2025-04-28 2025-05-30 首都医科大学附属北京积水潭医院 Three-dimensional reconstruction method for dynamic scene in operation robot operation
CN120431135A (en) * 2025-05-07 2025-08-05 山东大学齐鲁医院 A real-time tracking system and method for spinal motion

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3125808A1 (en) * 2014-03-28 2017-02-08 Intuitive Surgical Operations, Inc. Quantitative three-dimensional visualization of instruments in a field of view
CN110279467A (en) * 2019-06-19 2019-09-27 天津大学 Ultrasound image under optical alignment and information fusion method in the art of puncture biopsy needle
US20200078097A1 (en) * 2016-10-25 2020-03-12 GYS Tech, LLC d/b/a Cardan Robotics Methods and Systems for Robot-Assisted Surgery
CN115349952A (en) * 2022-07-04 2022-11-18 安徽医科大学第二附属医院 A Guided 3D Virtual Fixture Generation Method for Dynamic Surgical Path Guidance
CN117958968A (en) * 2024-02-29 2024-05-03 珠海横乐医疗科技有限公司 Multi-source sensing and motion control method of respiratory intervention robot
CN118398162A (en) * 2024-03-18 2024-07-26 常州锦瑟医疗信息科技有限公司 Pose determination method, display method and system of preoperative virtual model

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3125808A1 (en) * 2014-03-28 2017-02-08 Intuitive Surgical Operations, Inc. Quantitative three-dimensional visualization of instruments in a field of view
US20200078097A1 (en) * 2016-10-25 2020-03-12 GYS Tech, LLC d/b/a Cardan Robotics Methods and Systems for Robot-Assisted Surgery
CN110279467A (en) * 2019-06-19 2019-09-27 天津大学 Ultrasound image under optical alignment and information fusion method in the art of puncture biopsy needle
CN115349952A (en) * 2022-07-04 2022-11-18 安徽医科大学第二附属医院 A Guided 3D Virtual Fixture Generation Method for Dynamic Surgical Path Guidance
CN117958968A (en) * 2024-02-29 2024-05-03 珠海横乐医疗科技有限公司 Multi-source sensing and motion control method of respiratory intervention robot
CN118398162A (en) * 2024-03-18 2024-07-26 常州锦瑟医疗信息科技有限公司 Pose determination method, display method and system of preoperative virtual model

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
杨健等: ""多模态图像引导手术导航进展"", 《光学学报》, 31 October 2023 (2023-10-31), pages 27 - 45 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN119970235A (en) * 2025-01-15 2025-05-13 东南大学 A force feedback device and method for assisting surgical robot clamping operation
CN120070775A (en) * 2025-04-28 2025-05-30 首都医科大学附属北京积水潭医院 Three-dimensional reconstruction method for dynamic scene in operation robot operation
CN120431135A (en) * 2025-05-07 2025-08-05 山东大学齐鲁医院 A real-time tracking system and method for spinal motion
CN120431135B (en) * 2025-05-07 2025-11-11 山东大学齐鲁医院 A real-time spinal motion tracking system and method

Also Published As

Publication number Publication date
CN119055358B (en) 2025-02-14

Similar Documents

Publication Publication Date Title
CN119055358B (en) Surgical operation force feedback guiding method based on virtual mark tracking and instrument pose
Laidlow et al. DeepFusion: Real-time dense 3D reconstruction for monocular SLAM using single-view depth and gradient predictions
Shen et al. Context-aware depth and pose estimation for bronchoscopic navigation
Mountney et al. A probabilistic framework for tracking deformable soft tissue in minimally invasive surgery
Wei et al. Stereo dense scene reconstruction and accurate localization for learning-based navigation of laparoscope in minimally invasive surgery
US20110282151A1 (en) Image-based localization method and system
WO2016195698A1 (en) Method and system for simultaneous scene parsing and model fusion for endoscopic and laparoscopic navigation
Luo et al. Unsupervised learning of depth estimation from imperfect rectified stereo laparoscopic images
CN112562081A (en) Visual map construction method for visual layered positioning
US12482122B2 (en) Self-supervised image depth estimation method based on channel self-attention mechanism
WO2024098240A1 (en) Gastrointestinal endoscopy visual reconstruction navigation system and method
Yang et al. Reconstruct dynamic soft-tissue with stereo endoscope based on a single-layer network
CN115349952A (en) A Guided 3D Virtual Fixture Generation Method for Dynamic Surgical Path Guidance
Guo et al. Free-surgs: Sfm-free 3d gaussian splatting for surgical scene reconstruction
WO2022192540A1 (en) System for and method of real-time nonrigid mosaicking of laparoscopy images
CN120070755A (en) 3D Gaussian-based endoscopic surgery scene real-time reconstruction method
CN119091060A (en) Three-dimensional reconstruction method and system for laparoscopic surgery scenes based on three-dimensional Gaussian sputtering
CN111161408A (en) Method for realizing augmented reality, application thereof and computing equipment
Amara et al. Augmented Reality localisation using 6 DoF phantom head Pose Estimation-based generalisable Deep Learning model
CN115619727A (en) Establishment method and application of endoscopic image parallax prediction model based on two-way supervision
Sui et al. Binocular-based dense 3D reconstruction for robotic assisted minimally invasive laparoscopic surgery
Lin et al. Superpm: A large deformation-robust surgical perception framework based on deep point matching learned from physical constrained simulation data
CN117218074A (en) Training method, soft tissue deformation estimation method, device, equipment and storage medium
Kshirsagar et al. Depth estimation using stereo medical imaging
Guo et al. Endo3r: Unified online reconstruction from dynamic monocular endoscopic video

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant