US10909771B2 - System and method for image processing - Google Patents
System and method for image processing Download PDFInfo
- Publication number
- US10909771B2 US10909771B2 US15/510,175 US201415510175A US10909771B2 US 10909771 B2 US10909771 B2 US 10909771B2 US 201415510175 A US201415510175 A US 201415510175A US 10909771 B2 US10909771 B2 US 10909771B2
- Authority
- US
- United States
- Prior art keywords
- dimensional image
- display
- volume
- control element
- mask
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/20—Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/05—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves
- A61B5/055—Detecting, measuring or recording for diagnosis by means of electric currents or magnetic fields; Measuring using microwaves or radio waves involving electronic [EMR] or nuclear [NMR] magnetic resonance, e.g. magnetic resonance imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0033—Features or image-related aspects of imaging apparatus, e.g. for MRI, optical tomography or impedance tomography apparatus; Arrangements of imaging apparatus in a room
- A61B5/004—Features or image-related aspects of imaging apparatus, e.g. for MRI, optical tomography or impedance tomography apparatus; Arrangements of imaging apparatus in a room adapted for image acquisition of a particular organ or body part
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0033—Features or image-related aspects of imaging apparatus, e.g. for MRI, optical tomography or impedance tomography apparatus; Arrangements of imaging apparatus in a room
- A61B5/004—Features or image-related aspects of imaging apparatus, e.g. for MRI, optical tomography or impedance tomography apparatus; Arrangements of imaging apparatus in a room adapted for image acquisition of a particular organ or body part
- A61B5/0042—Features or image-related aspects of imaging apparatus, e.g. for MRI, optical tomography or impedance tomography apparatus; Arrangements of imaging apparatus in a room adapted for image acquisition of a particular organ or body part for the brain
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient; User input means
- A61B5/742—Details of notification to user or communication with user or patient; User input means using visual displays
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01R—MEASURING ELECTRIC VARIABLES; MEASURING MAGNETIC VARIABLES
- G01R33/00—Arrangements or instruments for measuring magnetic variables
- G01R33/20—Arrangements or instruments for measuring magnetic variables involving magnetic resonance
- G01R33/44—Arrangements or instruments for measuring magnetic variables involving magnetic resonance using nuclear magnetic resonance [NMR]
- G01R33/48—NMR imaging systems
- G01R33/54—Signal processing systems, e.g. using pulse sequences ; Generation or control of pulse sequences; Operator console
- G01R33/56—Image enhancement or correction, e.g. subtraction or averaging techniques, e.g. improvement of signal-to-noise ratio and resolution
- G01R33/5608—Data processing and visualization specially adapted for MR, e.g. for feature analysis and pattern recognition on the basis of measured MR data, segmentation of measured MR data, edge contour detection on the basis of measured MR data, for enhancing measured MR data in terms of signal-to-noise ratio by means of noise filtering or apodization, for enhancing measured MR data in terms of resolution by means for deblurring, windowing, zero filling, or generation of gray-scaled images, colour-coded images or images displaying vectors instead of pixels
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2576/00—Medical imaging apparatus involving image processing or analysis
- A61B2576/02—Medical imaging apparatus involving image processing or analysis specially adapted for a particular organ or body part
- A61B2576/026—Medical imaging apparatus involving image processing or analysis specially adapted for a particular organ or body part for the brain
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0062—Arrangements for scanning
- A61B5/0066—Optical coherence imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
- A61B6/032—Transmission computed tomography [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/08—Clinical applications
- A61B8/0808—Clinical applications for diagnosis of the brain
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
Definitions
- the specification relates generally to medical imaging, and specifically to a computing device, system and method for image processing.
- the planning and execution of surgical procedures may require the gathering and organization of large volumes of information, including various medical images of the patient.
- images can include, for example, MRI scans.
- Accessing such information, particularly during a surgical procedure, may require extensive preparation of different image views prior to the procedure; in other cases, significant portions of the images may simply not be available during the procedure, or may require additional operators and time-consuming programming and computational efforts to produce.
- a computing device comprising: a memory; an input device; a display; and a processor interconnected with the memory, the input device and the display, the processor configured to: acquire a three-dimensional image of an anatomical structure of a patient and store the three-dimensional image in the memory; render on the display (i) an initial volume of the three-dimensional image corresponding to an initial portion of the anatomical structure, and (ii) a moveable control element; the initial volume having an initial outer surface defined by a position of the control element; receive, from the input device, input data updating the position of the control element on the display relative to the initial volume; responsive to receiving the input data, render on the display, in place of the initial volume, a further volume of the three-dimensional image, corresponding to a further portion of the anatomical structure and having a further outer surface defined by the updated position of the control element; the processor configured to select the further volume by identifying a portion of the three-dimensional images that intersects with the at least one plane or volume
- method of processing images in a computing device having a memory, an input device, a display and a processor interconnected with the memory, the input device and the display, the method comprising: acquiring a three-dimensional image of an anatomical structure of a patient and storing the three-dimensional image in the memory; at the processor, rendering on the display (i) an initial volume of the three-dimensional image corresponding to an initial portion of the anatomical structure, and (ii) a moveable control element; the initial volume having an initial outer surface defined by a position of the control element; receiving, at the processor from the input device, input data updating the position of the control element on the display relative to the initial volume; responsive to receiving the input data, controlling the display at the processor to render, in place of the initial volume, a further volume of the three-dimensional image, corresponding to a further portion of the anatomical structure and having a further outer surface defined by the updated position of the control element.
- FIG. 1 depicts an operating theatre, according to a non-limiting embodiment
- FIG. 2 depicts a computing device for deployment in the operating theatre of FIG. 1 , according to a non-limiting embodiment
- FIG. 3 depicts a method of processing images, according to a non-limiting embodiment
- FIG. 4 depicts a method of rendering initial and further volumes in the method of FIG. 3 , according to a non-limiting embodiment
- FIG. 5 depicts an example of the performance of block 310 of the method of FIG. 3 , according to a non-limiting embodiment
- FIG. 6 depicts another example of the performance of block 310 of the method of FIG. 3 , according to a non-limiting embodiment
- FIG. 7 depicts a further example of the performance of block 310 of the method of FIG. 3 , according to a non-limiting embodiment
- FIG. 8 depicts examples of performances of block 320 , according to a non-limiting embodiment
- FIG. 9 depicts other examples of performances of block 320 , according to a non-limiting embodiment.
- FIG. 10 depicts further examples of performances of block 320 , according to a non-limiting embodiment.
- the terms, “comprises” and “comprising” are to be construed as being inclusive and open ended, and not exclusive. Specifically, when used in the specification and claims, the terms, “comprises” and “comprising” and variations thereof mean the specified features, steps or components are included. These terms are not to be interpreted to exclude the presence of other features, steps or components.
- FIG. 1 depicts a surgical operating theatre 100 in which a healthcare worker 102 (e.g. a surgeon) operates on a patient 104 .
- a healthcare worker 102 e.g. a surgeon
- surgeon 102 is shown conducting a minimally invasive surgical procedure on the brain of patient 104 .
- Minimally invasive brain surgery involves the insertion and manipulation of instruments into the brain through an opening that is significantly smaller than the portions of skull removed to expose the brain in traditional brain surgery techniques.
- Access port 106 typically includes a hollow cylindrical device with open ends.
- an introducer (not shown) is generally inserted into access port 106 .
- the introducer is typically a cylindrical device that slidably engages the internal surface of access port 106 and bears a conical atraumatic tip to allow for insertion of access port 106 into the sulcal folds of the brain.
- the introducer may be removed, and access port 106 may then enable insertion and bimanual manipulation of surgical tools into the brain. Examples of such tools include suctioning devices, scissors, scalpels, cutting devices, imaging devices (e.g. ultrasound sensors) and the like.
- FIG. 1 Also shown in FIG. 1 is an equipment tower 108 supporting a computing device (not shown) such as a desktop computer, as well as one or more displays 110 connected to the computing device for displaying images provided by the computing device.
- the computing device, display 110 , or both, may also be supported by other structures. Indeed, the computing device may be located outside of operating theatre 100 , and where multiple displays are provided, at least one display may also be located outside of operating theatre 100 .
- Equipment tower 108 may also support a tracking system 112 .
- Tracking system 112 when included, is generally configured to track the positions of one or more reflective markers (not shown) mounted on access port 106 , any of the above-mentioned surgical tools, or any combination thereof.
- markers also referred to as fiducial markers
- Tracking system 112 may therefore include a camera (e.g. a stereo camera) and a computing device (either the same device as mentioned above or a separate device) configured to locate the fiducial markers in the images captured by the camera, and determine the spatial positions of those markers within the operating theatre.
- the spatial positions may be provided by tracking system 112 to the computing device in equipment tower 108 for subsequent use.
- An example of tracking system 112 is the “Polaris” system available from Northern Digital Inc.
- an automated articulated arm 114 also referred to as a robotic arm, carrying an external scope 116 (i.e. external to patient 104 ).
- External scope 116 may be positioned over access port 106 by robotic arm 114 , and may capture images of the brain of patient 104 for presentation on display 110 .
- the movement of robotic arm 114 to place external scope 116 correctly over access port 106 may be guided by tracking system 112 and the computing device in equipment tower 108 .
- the images from external scope 116 presented on display 110 may be overlaid with other images, including images obtained prior to the surgical procedure.
- the images presented on display 110 may also display virtual models of surgical instruments present in the field of view of tracking system 112 (the positions and orientations of the models having been determined by tracking system 112 from the positions of the markers mentioned above).
- images of anatomical structures within patient 104 may be obtained using various imaging modalities.
- images of the brain of patient 104 may be obtained using Magnetic Resonance Imaging (MRI), Optical Coherence Tomography (OCT), ultrasound, Computed Tomography (CT), optical spectroscopy and the like.
- MRI Magnetic Resonance Imaging
- OCT Optical Coherence Tomography
- CT Computed Tomography
- images of the brain of patient 104 may be obtained using Magnetic Resonance Imaging (MRI), Optical Coherence Tomography (OCT), ultrasound, Computed Tomography (CT), optical spectroscopy and the like.
- MRI Magnetic Resonance Imaging
- OCT Optical Coherence Tomography
- CT Computed Tomography
- optical spectroscopy optical spectroscopy
- a computing device 200 is depicted, including a central processing unit (also referred to as a microprocessor or simply a processor) 202 interconnected with a non-transitory computer readable storage medium such as a memory 204 .
- a central processing unit also referred to as a microprocessor or simply a processor
- a non-transitory computer readable storage medium such as a memory 204 .
- Processor 202 and memory 204 are generally comprised of one or more integrated circuits (ICs), and can have a variety of structures, as will now occur to those skilled in the art (for example, more than one CPU can be provided).
- Memory 204 can be any suitable combination of volatile (e.g. Random Access Memory (“RAM”)) and non-volatile (e.g. read only memory (“ROM”), Electrically Erasable Programmable Read Only Memory (“EEPROM”), flash memory, magnetic computer storage device, or optical disc) memory.
- RAM Random Access Memory
- ROM read only memory
- EEPROM Electrically Erasable Programmable Read Only Memory
- flash memory magnetic computer storage device, or optical disc
- memory 204 includes both a volatile memory and a non-volatile memory.
- Other types of non-transitory computer readable storage medium are also contemplated, such as compact discs (CD-ROM, CD-RW) and digital video discs (DVD).
- Computing device 200 can also include a network interface 206 interconnected with processor 200 .
- Network interface 206 allows computing device 200 to communicate with other computing devices via a network (e.g. a local area network (LAN), a wide area network (WAN) or any suitable combination thereof).
- Network interface 206 thus includes any necessary hardware for communicating over such networks, such as radios, network interface controllers (NICs) and the like.
- Computing device 200 can also include an input/output interface 208 , including the necessary hardware for interconnecting processor 202 with various input and output devices.
- Interface 208 can include, among other components, a Universal Serial Bus (USB) port, an audio port for sending and receiving audio data, a Video Graphics Array (VGA), Digital Visual Interface (DVI) or other port for sending and receiving display data, and any other suitable components.
- USB Universal Serial Bus
- I/O interface 208 connects computing device 200 to “local” input and output devices
- network interface 206 connects computing device 200 to “remote” computing devices, which may themselves be connected to additional input and output devices. This arrangement may be varied, however.
- any suitable combination of the input and output devices to be discussed below may be connected to computing device 200 via network interface 206 rather than I/O interface 208 .
- I/O interface 208 may be omitted entirely, while in other embodiments network interface 206 may be omitted entirely.
- computing device 200 can be connected to input devices including a keyboard and mouse 210 , a microphone 212 , as well as scope 116 and tracking system 112 , mentioned above. Also via interface 208 , computing device 200 can be connected to output devices including illumination or projection components 214 (e.g. lights, projectors and the like), as well as display 110 and robotic arm 114 mentioned above. It is contemplated that other combinations of devices may also be present, omitting one or more of the above devices, including other input (e.g. touch screens) and output (e.g. speakers, printers) devices, and the like.
- input devices e.g. touch screens
- output e.g. speakers, printers
- Computing device 200 stores, in memory 204 , an image manipulation application 216 (also referred to herein as application 216 ) comprising a plurality of computer readable instructions executable by processor 202 .
- application 216 also referred to herein as application 216
- processor 202 executes the instructions of application 216 (or, indeed, any other application stored in memory 204 )
- processor 202 performs various functions implemented by those instructions, as will be discussed below.
- Processor 202 or computing device 200 more generally, is therefore said to be “configured” or “operating” to perform those functions via the execution of application 216 .
- Patient data repository 218 can contain a surgical plan defining the various steps of the minimally invasive surgical procedure to be conducted on patient 104 , as well as images of patient 104 (e.g. MRI and CT scans).
- computing device 200 is configured, via the execution of application 216 by processor 202 , to perform various functions related to presenting and manipulating images of patient 104 on display 110 . Those functions will be described in further detail below.
- Method 300 will be discussed in conjunction with its performance on computing device 200 as deployed in operating theatre 100 . It will be apparent to those skilled in the art, however, that method 300 can also be implemented on other computing devices in other systems.
- computing device 200 is configured to acquire at least one three-dimensional image of an anatomical structure of patient 104 and store the three-dimensional image in memory 204 (for example, in patient data repository 218 ).
- the anatomical structure is the brain of patient 104 , but a variety of other anatomical structures may also be imaged instead of, or in addition to, the brain.
- the mechanism of acquisition of the image is not particularly limited.
- computing device 200 can obtain the image directly from an imaging device (not shown), such as an MRI scanner. In other examples, computing device 200 can obtain the image from another computing device which itself obtained the image from an imaging device.
- the three-dimensional image is assumed to be an MRI scan.
- the collection of two-dimensional MRI slices, together representing a three-dimensional scan of the brain, is referred to herein as the three-dimensional image.
- other imaging modalities may be employed instead of, or in addition to, MRI.
- computing device 200 is configured to control display 110 (e.g. via I/O interface 208 ) to render an interface for presenting and manipulating the three-dimensional image.
- the interface includes a rendering of an initial volume of the three-dimensional image corresponding to an initial portion of the anatomical structure (that is, the brain of patient 104 ).
- the interface rendered at block 310 also includes at least one moveable control element, which will be discussed in greater detail below.
- the initial volume rendered on display 110 has an initial outer surface defined by a position of the control element relative to the initial volume.
- the three-dimensional image contains data depicting a given volume of the anatomical structure.
- the three-dimensional image may in fact depict a volume of patient 104 that is larger than the brain itself.
- An MRI scan of the brain for example, can depict the skull as well as the brain.
- the initial volume rendered on display 110 at block 310 need not contain the entire volume of the three-dimensional image (although it may contain the entire volume).
- Which portion of the three-dimensional image is rendered at block 310 is computed by processor 204 based on the positions of the above-mentioned control elements.
- control elements define geometrical planes or volumes having positions relative to the three-dimensional image (e.g. coordinates within a coordinate system associated with the three-dimensional image). At those positions, the planes or volumes defined by the control elements intersect with the volume of the three-dimensional image. Such intersections define the outer surface of the initial volume to be rendered from the complete three-dimensional image. That is, the intersections of the control elements with the three-dimensional image define the boundaries of the initial volume to be rendered (i.e. what portion of the image will be rendered and what portion will be excluded from the rendering).
- FIG. 4 depicts a method 400 , per formed by computing device 200 (via the execution of application 216 , discussed earlier), of performing blocks 310 and 320 (that is, the rendering steps) of method 300 .
- Method 400 will be described with reference to FIGS. 4, 5 and 6 , which depicts examples of the interfaces generated at block 310 (and, as will be seen later herein, at block 320 ).
- computing device 200 is configured to select one or more control elements.
- the definitions of the control elements are stored in memory 204 , for example as part of application 216 .
- the selection at block 405 can be automatic—for example, application 216 can contain instructions to select a default control element, or set of control elements—or the selection can be received as input data, for example from keyboard/mouse 210 .
- control elements are contemplated, though it will be understood that these are only examples—other types of control elements will occur to those skilled in the art in light of the present description. As seen in FIG. 4 , any combination of the three types of control elements may be selected. For simplicity, the control elements will be discussed individually, however. Thus, when computing device 200 determines that the plane control elements have been selected at block 410 , performance of method 400 proceeds to block 415 .
- the plane control elements include a plurality of planes.
- three orthogonal planes are contemplated, although other numbers of planes, disposed at other angles relative to each other, may also be employed.
- the planes each have an initial default location relative to the three-dimensional image, such that each plane intersects the three-dimensional image.
- the planes by intersecting the three-dimensional image, divide the image into quadrants (in particular, eight quadrants in the case of three orthogonal planes).
- computing device 200 is configured to select one of the quadrants, and clip the intersection of that selected quadrant and the three-dimensional image, before proceeding to block 310 to render the resulting initial volume on display 110 .
- FIG. 5 depicts an interface 500 generated on display 110 that includes a rendering of an initial volume 504 and control elements 508 , 512 and 516 in the form of orthogonal planes.
- control elements 508 , 512 and 516 are the sagittal, coronal, and transverse anatomical planes, respectively.
- two-dimensional slices of the three-dimensional image are also illustrated, each corresponding to the position of one of the three planes within the three-dimensional image. The two-dimensional views may be omitted in other embodiments.
- initial volume 504 includes the entire volume of the three-dimensional image (that is, the MRI scan of patient 104 's head), with the exception of the portion intersected by one of the quadrants defined by planes 508 , 512 and 516 .
- the intersecting quadrant has been cut away, or clipped, and therefore is not rendered in interface 500 , which allows certain interior portions of the brain to be rendered.
- the position of the three planes defines the outer surface of initial volume 504 .
- computing device 200 may also (or alternatively) determine that a cone control element has been selected at block 405 . If the determination at block 420 is affirmative, performance of method 400 proceeds to block 425 .
- the cone control element is a volume in the shape of a cone, or in some embodiments, a truncated cone, having an initial default position relative to the three-dimensional image.
- computing device 200 is configured to clip the portion of the three-dimensional image that intersects with the conical volume, before proceeding to block 310 .
- Interface 600 includes an initial volume 604 and a control element 608 .
- Control element 608 corresponds to a cone 612 , which may be omitted from interface 600 (in the present example, cone 612 is shown in FIG. 6 only for illustrative purposes, and does not appear on interface 600 ). More particularly, control element 608 comprises an axis 616 along which a model 620 of an access port may be positioned at various depths relative to the three-dimensional image (and by extension, to initial volume 604 ). In other words, input data received at processor 204 may act to slide model 620 along axis 616 . As seen in FIG.
- the summit of cone 612 coincides with the point of model 620 .
- cone 612 will begin to intersect with the three-dimensional image.
- Computing device 200 is configured to clip any portions of the three-dimensional image that intersect with cone 612 .
- the default initial position for cone 612 does not intersect with the three-dimensional image, and initial volume 604 has not had any portions clipped therefrom by computing device 200 .
- computing device 200 may also (or alternatively) determine that a mask control element has been selected at block 405 . If the determination at block 430 is affirmative, performance of method 400 proceeds to block 435 .
- the mask control element is an irregular surface having an initial default position relative to the three-dimensional image.
- the initial position of the mask is determined by computing device 200 by any suitable algorithm or combination of algorithms to identify the border between skull and brain.
- images such as MRI scans generally include data depicting the entire head of patient 104 , including skull and facial features.
- the mask control element is an estimate of the outer surface of the patient 104 's brain, and divides the three-dimensional image into an “outer” part representing non-brain tissues, and an “inner” part representing brain tissues.
- computing device 200 is configured to clip the portion of the three-dimensional image that intersects with the above-mentioned outer part, before proceeding to block 310 .
- the mask control element defines the outer surface of the initial volume.
- Interface 700 includes an initial volume 704 and a control element 708 .
- Control element 708 corresponds to the mask, an irregular surface not illustrated in conjunction with initial volume 704 .
- the mask may be illustrated in a second two-dimensional view as an outline 712 delineating the boundary between the outer part (excluded from initial volume 704 ) and the inner part (included in initial volume 704 ).
- the nature of control element 708 is not particularly limited, and in the present example comprises an axis 716 and a depth indicator 720 indicating the depth of the mask along axis 716 .
- interface 500 does not show the application of a mask (thus, the patient 104 's skull and ears are visible in initial volume 504 ), interface 600 does apply the mask, in addition to cone 612 .
- computing device 200 is configured to proceed to block 315 .
- processor 204 is configured to receive input data updating the position of the control element(s) rendered at block 310 , relative to the initial volume rendered at block 310 .
- the position of the control element(s) within the coordinate system of the three-dimensional image may change in response to input data.
- the input data received at block 315 can include the selection and dragging, or other repositioning, or the control elements shown therein.
- processor 204 may receive input data from keyboard/mouse 210 representing a selection and moving operation performed on any one or more of planes 508 , 512 and 516 .
- processor 204 may receive input data from keyboard/mouse 210 representing a change in depth of model 620 along axis 616 , a change in angle of axis 616 , and the like.
- processor 204 may receive input data from keyboard/mouse 210 representing a change in depth of depth indicator 720 along axis 716 .
- input data can also be received form a touchscreen interface, a joystick input, a gesture control or a voice control interface.
- computing device 200 is configured, at block 320 , to control display 110 to present an updated interface.
- the updated interface includes a rendering of a further volume of the three-dimensional image, as well as the control elements rendered in block 310 , but in their updated positions.
- the generation of an updated interface is performed as described above in connection with method 400 , substituting the updated control element positions for the previous (e.g. initial) control element positions.
- computing device 200 is configured to repeat blocks 315 and 320 until an exit command is received at block 325 .
- FIGS. 8A and 8B illustrate the interfaces rendered in two subsequent performances of blocks 315 and 320 .
- FIG. 8A illustrates an interface 800 in which the position of plane 516 has been updated to raise plane 516 in the superior direction (that is, towards the top of patient 104 's head as depicted by the three-dimensional image).
- the clipped quadrant that intersects with the three-dimensional image has changed in dimensions.
- the outer surface of a further volume 802 of the three-dimensional image has changed.
- the example above shows only a translation of plane 516 , it is contemplated that any of the planes may also be rotated, angled and the like.
- FIG. 8B illustrates another interface 804 in which plane 516 has been returned to its previous position (as shown in FIG. 5 ), but in which a further volume 806 has been rotated on display 110 .
- computing device 200 is configured to select one of the quadrants defined by the intersecting planes. In the present example, computing device 200 is configured to automatically select the quadrant of which the greatest proportion is visible on display 110 . Thus, for the same plane positions, different quadrants may be selected for clipping based on the position of the illustrated volume on the display.
- Interface 900 in FIG. 9A depicts a further volume 902 in which control element 608 has been relocated in the inferior direction (that is, model 620 has been moved down axis 616 ).
- a portion of the three-dimensional image intersects with cone 612 , and the intersection portion has been clipped, resulting a cone-shaped cavity in further volume 902 .
- Interface 904 in FIG. 9B depicts a further volume 906 which has been rotated in comparison to further volume 902 .
- the angle of control element 608 has been altered.
- Interface 1000 in FIG. 10A depicts a further volume 1002 of the three-dimensional image, in which depth marker 720 has been raised “outwards” from the brain along axis 716 , in comparison with FIG. 7 .
- the mask defined by control element 708 defines an outer surface of further volume 1002 that lies outside the outer surface of the three-dimensional image for the majority of the patient 104 's skull.
- the skull and facial features are visible in further volume 1002 , as they do not intersect with the “outer” part mentioned earlier.
- FIG. 10B depicts an interface 1004 in which depth marker 720 has been relocated inwardly along axis 716 , contracting the mask and thus adjusting the outer surface of a further volume 1006 .
- the boundary of the mask is clearly visible in FIG. 10B on the two-dimensional pane.
- the adjustment of the mask depth as shown in FIGS. 10A and 10B may be implemented in a variety of ways.
- the brain/skull boundary detection parameters are not altered. Rather, those parameters are set once, and computing device is configured to shift each point of the mask inwards or outwards by a number of pixels (or voxels, or any other unit of measurement within the three-dimensional image) proportional to the distance travelled by depth marker 720 along axis 716 .
- the three-dimensional image may be supplemented by other three-dimensional images acquired by computing device 200 .
- patient data repository may contain one or more images of fluid flow tracts for patient 104 , one or more images or models of a tumour within the brain of patient 104 , and the like. Such images can be overlaid on the initial and further volumes discussed above. In addition, such images can be exempt from the clipping behaviour discussed above.
- FIG. 10B shows an image 1008 of a tumour in conjunction with further volume 1006 . As seen in FIG. 10B , the tumour is not subject to the masking behaviour exhibited with further volume 1006 . It will now be apparent to those skilled in the art that FIG. 5 also shows an image 524 of a tumour.
- FIG. 9B shows, in addition to further volume 906 , an image 908 of fluid tracts.
- Image 908 can be a portion selected from a larger image of fluid tracts (for example, the selected portion can be limited to those tracts that would be intersected by the access port if the access port were inserted to the illustrated position of model 620 ).
- Image overlays such as fluid tracts and tumours may be enabled and disabled by way of input data received at processor 204 (e.g. from keyboard/mouse 210 ).
- the interfaces discussed above may include one or more selectable toggle elements for enabling and disabling such overlays.
- Other types of overlays contemplated can display or hide different types of tissue.
- the three-dimensional image can include identifiers of which type of tissue each voxel (or group of voxels) depicts. Such identifiers may be added to the three-dimensional image manually, or by execution of a tissue identification algorithm.
- An interface 100 on display 110 may then include selectable elements that disable or enable the display of various portions of the three-dimensional image.
- an additional selection of data from the three-dimensional image can occur at or before block 310 , depending on which tissue types are selected for display.
- certain tissue types may be identified as being exempt from the “clipping” behaviour discussed above (similar to the illustration of the tumour model in FIG. 10B , which is exempt from the clipping imposed by the mask).
- the volumes mentioned above can correspond to particular tissue types, and computing device 200 can render additional volumes whose outer surfaces are not defined by control element positions (as they are exempt from clipping).
- the corresponding initial or further volume may include an illustration of the plane from which the two-dimensional view is taken. This is shown in FIG. 5 ; a similar mechanism can be applied to other interfaces, such as that shown in FIGS. 7, 10A and 10B .
- computing device 200 may colorize control elements to enhance their visibility on display 110 .
- the “outer” part beyond the boundary of the mask shown in FIGS. 7, 10A and 10B may be colorized differently than the inner part.
- each one of planes 508 , 512 and 516 may be assigned a colour.
- the planes as illustrated on initial volume 504 and any subsequently presented further volumes, as well as in the accompanying two-dimensional views, may bear the same colours.
- control elements may be configurable.
- the radius of cone 612 may be altered by processor 204 in response to input data.
Landscapes
- Health & Medical Sciences (AREA)
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- General Health & Medical Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- Veterinary Medicine (AREA)
- Public Health (AREA)
- Surgery (AREA)
- Molecular Biology (AREA)
- Medical Informatics (AREA)
- Heart & Thoracic Surgery (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- General Physics & Mathematics (AREA)
- High Energy & Nuclear Physics (AREA)
- Computer Graphics (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computer Hardware Design (AREA)
- Condensed Matter Physics & Semiconductors (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Architecture (AREA)
- Neurology (AREA)
- Magnetic Resonance Imaging Apparatus (AREA)
Abstract
Description
Claims (6)
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/CA2014/000690 WO2016041045A1 (en) | 2014-09-15 | 2014-09-15 | System and method for image processing |
Related Parent Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CA2014/000690 A-371-Of-International WO2016041045A1 (en) | 2014-09-15 | 2014-09-15 | System and method for image processing |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/120,722 Continuation US12039682B2 (en) | 2014-09-15 | 2020-12-14 | System and method for image processing |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20170249786A1 US20170249786A1 (en) | 2017-08-31 |
| US10909771B2 true US10909771B2 (en) | 2021-02-02 |
Family
ID=55532365
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US15/510,175 Active 2035-02-13 US10909771B2 (en) | 2014-09-15 | 2014-09-15 | System and method for image processing |
| US17/120,722 Active 2036-10-23 US12039682B2 (en) | 2014-09-15 | 2020-12-14 | System and method for image processing |
Family Applications After (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/120,722 Active 2036-10-23 US12039682B2 (en) | 2014-09-15 | 2020-12-14 | System and method for image processing |
Country Status (3)
| Country | Link |
|---|---|
| US (2) | US10909771B2 (en) |
| CA (1) | CA2960889C (en) |
| WO (1) | WO2016041045A1 (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20210166494A1 (en) * | 2014-09-15 | 2021-06-03 | Synaptive Medical Inc. | System and method for image processing |
Families Citing this family (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190076195A1 (en) * | 2015-11-11 | 2019-03-14 | Think Surgical, Inc. | Articulating laser incision indication system |
| US10780507B2 (en) | 2018-06-17 | 2020-09-22 | Arevo, Inc. | Embedding fiducial marks into articles of manufacture with non-trivial dimensional variations |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20080177182A1 (en) * | 2007-01-24 | 2008-07-24 | Kabushiki Kaisha Toshiba | Ultrasonic imaging apparatus and method for acquiring ultrasonic image |
| US20110107270A1 (en) * | 2009-10-30 | 2011-05-05 | Bai Wang | Treatment planning in a virtual environment |
| US8083680B2 (en) | 2007-03-20 | 2011-12-27 | Medison Co., Ltd. | Ultrasound system and method for forming an ultrasound image |
| US20120078102A1 (en) | 2010-09-24 | 2012-03-29 | Samsung Medison Co., Ltd. | 3-dimensional (3d) ultrasound system using image filtering and method for operating 3d ultrasound system |
| US20130021336A1 (en) * | 2011-07-19 | 2013-01-24 | Toshiba Medical Systems Corporation | Image processing system, image processing device, image processing method, and medical image diagnostic device |
| US20140063208A1 (en) | 2011-05-23 | 2014-03-06 | Toshiba Medical Systems Corporation | Medical image diagnostic apparatus, image processing apparatus, and ultrasonic diagnostic apparatus |
| US20140071132A1 (en) | 2011-05-16 | 2014-03-13 | Toshiba Medical Systems Corporation | Image processing system, image processing apparatus, and image processing method |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2007015365A1 (en) * | 2005-08-01 | 2007-02-08 | National University Corporation NARA Institute of Science and Technology | Information processing device and program |
| US20100259263A1 (en) * | 2007-11-14 | 2010-10-14 | Dominic Holland | Longitudinal registration of anatomy in magnetic resonance imaging |
| AU2009202377A1 (en) * | 2009-06-15 | 2011-01-06 | Canon Kabushiki Kaisha | Combining overlapping objects |
| WO2011066689A1 (en) * | 2009-12-04 | 2011-06-09 | Shenzhen Institute Of Advanced Technology | Method and device for detecting bright brain regions from computed tomography images |
| US9659364B2 (en) * | 2010-03-11 | 2017-05-23 | Koninklijke Philips N.V. | Probabilistic refinement of model-based segmentation |
| JP5871705B2 (en) * | 2012-04-27 | 2016-03-01 | 株式会社日立メディコ | Image display apparatus, method and program |
| US10561861B2 (en) * | 2012-05-02 | 2020-02-18 | Viewray Technologies, Inc. | Videographic display of real-time medical treatment |
| RU2523929C1 (en) * | 2013-03-18 | 2014-07-27 | Корпорация "САМСУНГ ЭЛЕКТРОНИКС Ко., Лтд." | System and method for automated planning of views in 3d brain images |
| WO2016041045A1 (en) * | 2014-09-15 | 2016-03-24 | Synaptive Medical (Barbados) Inc. | System and method for image processing |
-
2014
- 2014-09-15 WO PCT/CA2014/000690 patent/WO2016041045A1/en not_active Ceased
- 2014-09-15 CA CA2960889A patent/CA2960889C/en active Active
- 2014-09-15 US US15/510,175 patent/US10909771B2/en active Active
-
2020
- 2020-12-14 US US17/120,722 patent/US12039682B2/en active Active
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20080177182A1 (en) * | 2007-01-24 | 2008-07-24 | Kabushiki Kaisha Toshiba | Ultrasonic imaging apparatus and method for acquiring ultrasonic image |
| US8083680B2 (en) | 2007-03-20 | 2011-12-27 | Medison Co., Ltd. | Ultrasound system and method for forming an ultrasound image |
| US20110107270A1 (en) * | 2009-10-30 | 2011-05-05 | Bai Wang | Treatment planning in a virtual environment |
| US20120078102A1 (en) | 2010-09-24 | 2012-03-29 | Samsung Medison Co., Ltd. | 3-dimensional (3d) ultrasound system using image filtering and method for operating 3d ultrasound system |
| US20140071132A1 (en) | 2011-05-16 | 2014-03-13 | Toshiba Medical Systems Corporation | Image processing system, image processing apparatus, and image processing method |
| US20140063208A1 (en) | 2011-05-23 | 2014-03-06 | Toshiba Medical Systems Corporation | Medical image diagnostic apparatus, image processing apparatus, and ultrasonic diagnostic apparatus |
| US20130021336A1 (en) * | 2011-07-19 | 2013-01-24 | Toshiba Medical Systems Corporation | Image processing system, image processing device, image processing method, and medical image diagnostic device |
Non-Patent Citations (3)
| Title |
|---|
| International Preliminary Report on Patentability dated Sep. 3, 2015 issued from the Canadian Intellectual Property Office relating to PCT International Application No. PCT/CA2014/000690. |
| International Search Report dated May 15, 2015 issued from the Canadian Intellectual Property Office relating to PCT International Application No. PCT/CA2014/000690. |
| Written Opinion dated May 15, 2015 issued from the Canadian Intellectual Property Office relating to PCT International Application No. PCT/CA2014/000690. |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20210166494A1 (en) * | 2014-09-15 | 2021-06-03 | Synaptive Medical Inc. | System and method for image processing |
| US12039682B2 (en) * | 2014-09-15 | 2024-07-16 | Synaptive Medical Inc. | System and method for image processing |
Also Published As
| Publication number | Publication date |
|---|---|
| US12039682B2 (en) | 2024-07-16 |
| WO2016041045A1 (en) | 2016-03-24 |
| US20210166494A1 (en) | 2021-06-03 |
| CA2960889A1 (en) | 2016-03-24 |
| US20170249786A1 (en) | 2017-08-31 |
| CA2960889C (en) | 2022-04-19 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12462468B2 (en) | Treatment procedure planning system and method | |
| US20190057555A1 (en) | Method, system and apparatus for rendering medical image data | |
| US10074176B2 (en) | Method, system and apparatus for displaying surgical engagement paths | |
| CN102422335B (en) | Systems, methods and devices for interactive preoperative assessment | |
| US12039682B2 (en) | System and method for image processing | |
| US12023107B2 (en) | Method, system and apparatus for maintaining patient registration in a surgical navigation system | |
| US20210290309A1 (en) | Method, system and apparatus for surface rendering using medical imaging data | |
| CN120826201A (en) | Depth-based mixed reality image generation | |
| HK1227126A1 (en) | Dynamic and interactive navigation in a surgical environment |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| AS | Assignment |
Owner name: SYNAPTIVE MEDICAL (BARBADOS) INC., BARBADOS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GALLOP, DAVID;HODGES, WES;REEL/FRAME:053985/0617 Effective date: 20140924 |
|
| AS | Assignment |
Owner name: SYNAPTIVE MEDICAL INC., CANADA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SYNAPTIVE MEDICAL (BARBADOS) INC.;REEL/FRAME:054206/0570 Effective date: 20200902 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| AS | Assignment |
Owner name: ESPRESSO CAPITAL LTD., CANADA Free format text: SECURITY INTEREST;ASSIGNOR:SYNAPTIVE MEDICAL INC.;REEL/FRAME:054922/0791 Effective date: 20201223 |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2551); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY Year of fee payment: 4 |