[go: up one dir, main page]

WO2011163414A2 - Mécanisme de génération et d'édition de structures avancées - Google Patents

Mécanisme de génération et d'édition de structures avancées Download PDF

Info

Publication number
WO2011163414A2
WO2011163414A2 PCT/US2011/041520 US2011041520W WO2011163414A2 WO 2011163414 A2 WO2011163414 A2 WO 2011163414A2 US 2011041520 W US2011041520 W US 2011041520W WO 2011163414 A2 WO2011163414 A2 WO 2011163414A2
Authority
WO
WIPO (PCT)
Prior art keywords
image
structures
map
deformation
updated
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/US2011/041520
Other languages
English (en)
Other versions
WO2011163414A3 (fr
Inventor
Michael Huber
Michael Waschbuesch
Lasse Toimela
Patrik Kunz
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Siemens Healthineers International AG
MURABITO ANTHONY C
Original Assignee
Varian Medical Systems International AG
MURABITO ANTHONY C
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US12/821,977 external-priority patent/US9401051B2/en
Priority claimed from US12/821,985 external-priority patent/US10311585B2/en
Application filed by Varian Medical Systems International AG, MURABITO ANTHONY C filed Critical Varian Medical Systems International AG
Priority to EP11798881.6A priority Critical patent/EP2584970A4/fr
Priority to CN201180039643.1A priority patent/CN103079467B/zh
Priority to JP2013516750A priority patent/JP5926252B2/ja
Publication of WO2011163414A2 publication Critical patent/WO2011163414A2/fr
Publication of WO2011163414A3 publication Critical patent/WO2011163414A3/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/174Segmentation; Edge detection involving the use of two or more images
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5211Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
    • A61B6/5229Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5211Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
    • A61B6/5229Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
    • A61B6/5235Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from the same or different ionising radiation imaging techniques, e.g. PET and CT
    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B6/00Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
    • A61B6/52Devices using data or image processing specially adapted for radiation diagnosis
    • A61B6/5211Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
    • A61B6/5229Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image
    • A61B6/5235Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from the same or different ionising radiation imaging techniques, e.g. PET and CT
    • A61B6/5241Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data combining image data of a patient, e.g. combining a functional image with an anatomical image combining images from the same or different ionising radiation imaging techniques, e.g. PET and CT combining overlapping images of the same imaging modality, e.g. by stitching
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10072Tomographic images
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20092Interactive image processing based on input by user
    • G06T2207/20104Interactive definition of region of interest [ROI]

Definitions

  • the present invention pertains generally to radiation therapy and, in particular, involves a mechanism for manipulating images generated by radiotherapy machines used in radiation diagnostic and treatment applications.
  • an imaging device such as an X-ray device, Computer Tomography (CT), or Magnetic Resonance Imaging (MR) device is used to generate one or more initial scans or images of the area of interest. These initial scans may be acquired by focusing a beam of radiation into a target volume and collecting the traversing beams in an imager. The beams collected by the imagers are used to generate a display (i.e., one or more images) of the targeted volume that may be used to diagnose or monitor an afflicted area (e.g., a tumor or lesion or a surrounding area).
  • CT Computer Tomography
  • MR Magnetic Resonance Imaging
  • critical structures e.g ., regions or organs
  • Conventional medical imaging techniques include techniques for automatically identifying ("segmenting") organs and large structures. These techniques often include delineating adjacent structures by derived radiodensities and classifying the structures according to their relative positions and derived densities with known values.
  • identification of these regions often also includes tracing the outline ("contouring") of these or other structures.
  • radiation targeting a specific organ or a portion of an organ may require specific identification and/or demarcation of the portion(s) of the organ to receive treatment.
  • tumors can also be specifically contoured and identified for treatment. For certain treatment plans, it may be preferred to designate these identified portions by specifically contouring the circumference of the area.
  • CBCT cone-beam computerized tomography system
  • 2D slices are reconstructed from one dimensional projections of the patient, and these slices may be combined to form a three dimensional (3D) image of the patient.
  • a cone beam computerized tomography system is similar to that of a conventional computerized tomography system, with the exception that an entire volumetric image is acquired through rotation of the source and imager, and a fully 3D image is reconstructed from a plurality of 2D projections.
  • specific anatomies may change (sometimes drastically) over time and/or as a result of receiving radiation treatment over the course of a treatment plan.
  • the target volumes may expand or reduce in size, depending on the ailment and the efficacy of the treatment plan .
  • treatment plans designed around diagnostic images generated during an initial CT scan may be ineffective, inefficient, or even dangerous to treat patients.
  • updated images of the treated area are sometimes acquired periodically by generating additional images during the treatment process to ensure an appropriate positioning of the therapeutic radiation beam and to determine the effectiveness of the treatment regimen.
  • Recently developed treatment machines allow for the detection of anatomical changes by employing advanced imaging acquisition techniques such as cone-beam computer tomography (CBCT) at the site of the treatment machine, often either immediately before or after treatment is administered to the patient.
  • CBCT cone-beam computer tomography
  • anatomical structures which have been defined on one image can be automatically propagated to another 3D image data set (e.g., CBCT acquired on the treatment machine), given the condition that the two data sets have been pre-registered to each other.
  • a method is provided for intelligent automatic propagation of manual or automatic contouring across linked (e.g., registered) images and image data sets. As provided, the method includes acquiring one or more images of one or more image data sets, determining the correlation between the images with respect to identified structures, generating a deformation map that establishes a correspondence for each point in the source image with a point in the target image.
  • the intelligent propagation mechanism applies this deformation map individually to each structure of the source image and propagates the deformed structure to the target image.
  • This allows automatic propagation of generated contouring and local structural changes to structures in a given data set or between disparate data sets while accounting for content deviation as a result of time and/or treatment that may exist between the images.
  • the advantages provided by performing this method include enabling structure copy functionality that incorporates the knowledge from the deformation map to provide a more accurate match to the actual anatomy in the target image.
  • the same tools designed for correction of the deformation field from image registration can also be used to correct features simultaneously on the two or more registered images to enable simultaneous contouring on multiple registered data sets.
  • the efficiency of editing of multiple structures simultaneously may be realized.
  • Figure 1 depicts a flowchart of a method for automatically propagating defined structures across a plurality of associated data sets, in accordance with embodiments of the present invention.
  • Figure 2 depicts a flowchart of a method for propagating updated contouring data between updated images in a single data set, in accordance with embodiments of the present invention.
  • Figure 3 is an illustration of an exemplary propagation between related images in a data set, in accordance with embodiments of the present invention.
  • Figure 4 depicts a flowchart of a method for automatically editing structures a plurality of related images, in accordance with embodiments of the present invention.
  • Figure 5 is an illustration of an exemplary structural editing in a single image, in accordance with embodiments of the present invention.
  • Figure 6 is an illustration of an exemplary structural editing and propagation in multiple, related images, in accordance with embodiments of the present invention.
  • Figure 7 depicts an exemplary computing environment, in accordance with embodiments of the present invention.
  • Computer-usable media may comprise computer storage media and communication media.
  • Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer-readable instructions, data structures, program modules or other data.
  • Computer storage media includes, but is not limited to, random access memory (RAM), read only memory (ROM), electrically erasable programmable ROM (EEPROM), flash memory or other memory technology, compact disk ROM (CD-ROM), digital versatile disks (DVDs) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium that can be used to store the desired information.
  • Communication media can embody computer-readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism and includes any information delivery media.
  • modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
  • communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, radio frequency (RF), infrared and other wireless media. Combinations of any of the above should also be included within the scope of computer-readable media.
  • Embodiments include a method for automatically propagating manually contoured effects from a source image to associated target images while adjusting for differences between the referenced images.
  • Figure 1 is a flowchart 1 00 of a method for automatically propagating manually or automatically contoured structures across a plurality of images in accordance with one embodiment.
  • Steps 101 -107 describe exemplary steps comprising the process depicted in flowchart 100 in accordance with the various embodiments herein described .
  • the flowchart 100 is implemented as computer-executable instructions stored in a computer-readable medium and performed by a computing device executing a process for automatically propagating manually contoured effects among data sets.
  • a first image comprising a first plurality of structures and one or more manually or automatically contoured effects is accessed.
  • the first image may comprise, for example, a first data scan of a CT data set generated by a medical imaging device.
  • the first image may also consist of the display of data acqu ired during an initial diagnostic CT scan.
  • data for the entire data set may be pre-imaged and stored in a data storage repository (such as a memory), which is accessed at step 101 .
  • the first image may itself comprise a plurality of anatomical features or structures. These features may include, but are not limited to, organs, tumors, lesions and the like. Some or all of these features may be automatically segmented according to various identification and segmentation techniques implemented as software programs.
  • an image may display features which may comprise a plurality of contoured effects, such as outlined regions or portions of structures.
  • a second image comprising a second plurality of structures is accessed .
  • the second image may comprise a display of the same anatomical region(s) and for the same (or even different) patient.
  • the second image may comprise, for example, a second data scan of a data set generated by a medical imaging device.
  • the medical imaging device may comprise the same imaging device as the imaging device which generated the data for the first data set.
  • the second image may be of the same data set as the first image.
  • other imaging devices may also be used to generate the second data set.
  • a radiation therapy machine may be equipped with CBCT or other imaging equipment.
  • a patient receiving treatment from a radiation therapy machine may also be imaged by the imaging equipment.
  • Other devices, such as magnetic resonance imaging devices or other similar medical imaging devices may also be used to acquire image data.
  • a computing device performing step 103 may be communicatively coupled to the treatment and/or imaging equipment, thereby allowing the data comprising the second image, once acquired, to be accessed immediately by the executing computing device.
  • data for one image or the entire data set may be likewise pre-imaged and stored in the data storage repository that was accessed at step 101 .
  • the second image may, as with the first image, comprise a plurality of anatomical features. These features may include, but are not limited to all or a portion of the structures displayed in the first image.
  • the second image may consist of a display of an equivalent or substantially equivalent general anatomical region displayed in the first image, with an equivalent or substantially similar orientation, axes, dimensions, scope.
  • the second image may comprise a display of a different plurality of anatomical features, wherein only portions of the anatomy displayed in the first image are visible, and which may differ in orientation, or other visual configurations and conditions with the display of the features in the first image.
  • the second image may be pre-associated with the first image. Pre-association may include registration within a system.
  • any disparities of specifically identified content common to both the first image accessed at step 101 and the second image accessed at step 103 are identified and the relationships between points of disparities are mapped.
  • Mapping the dissimilarities of the common features may be performed by generating a "deformation map" of the two images.
  • the deformation map may be generated by establishing a correspondence for each pixel of a common structure or anatomical region in the first or "source” image with an equivalent pixel in the second or “target” image. The positions of the pixels are compared and the relative displacement between a pair of corresponding pixels is then determined to derive a relativity between the pixels.
  • corresponding pixels are identified by comparing the respective pixel intensities relative to the rest of the image. This correspondence may be implemented as a three dimensional vector field, for example. Alternatively, the correspondence may also be implemented by a plurality of mathematical functions which express the relativity between two corresponding points.
  • contoured effects identified in the first image accessed at step 101 are automatically propagated to the second image accessed at step 103 according to the deformation map generated at step 105.
  • the contoured effects may be propagated automatically to every image registered to the first image (or every data set registered to the data set of the first image).
  • Propagating the contoured effects may be performed, for example, by applying the deformation map to the contoured effects and copying the resulting output on to the second image. That is, rather than an explicit 1 : 1 propagation of identically contoured effects, any manually or automatically contoured effects from the first image will be propagated to the second image while being adjusted to the mapped disparities between the two images.
  • an area surrounding the tumor which has been contoured for specific targeting would, ideally, be also modified correspondingly for optimal targeting of treatment.
  • any contouring of a target may be inappropriately sized (e.g., either of insufficient size or specificity), in a non- corresponding position, or a non-conforming and non-ideal shape.
  • Figure 2 depicts a flowchart of a method for propagating updated structural contouring data among a plurality of data sets, in accordance with embodiments of the present invention .
  • Steps 201 -213 describe exemplary steps comprising the process depicted in flowchart 200 in accordance with the various embodiments herein described .
  • the flowchart 200 is implemented as computer-executable instructions stored in a computer-readable medium and performed by a computing device executing a process for automatically updating structural contouring data among a plurality of data sets.
  • initial image data of a target subject is acquired by an imaging device.
  • the image data comprises a display of a portion of the target subject's anatomy.
  • Acquisition of the initial image data may be performed by generating image data of a plurality of images in a data set in an imaging device such as an X-ray, MRI, or other medical imaging device.
  • the initial image data may be acquired during, or in furtherance of, a diagnostic procedure.
  • the initial image data may be stored for future reference.
  • initial image data may be stored by, accessed from, and manipulated in a dedicated software application. User input received through a graphical user interface of the software application may direct the storage, access, and manipulation of the file(s) comprising the image data.
  • the initial image data acquired by the imaging device in step 201 may comprise a plurality of anatomical structures (e.g., organs, tumors, lesions, etc.) Some or all of these structures may be automatically segmented according to various identification and segmentation techniques implemented as software programs. In some embodiments, additional contouring corresponding to user input may be received at step 203, for the purposes of treatment planning, for example.
  • anatomical structures e.g., organs, tumors, lesions, etc.
  • additional contouring corresponding to user input may be received at step 203, for the purposes of treatment planning, for example.
  • the contouring may be manually added, through user input received in a graphical user interface executing on a computer system, for example.
  • the contouring may be automatically derived from contouring algorithms. Examples of contouring may include effects such as explicit outlining of regions of all or a portion one or more portions of structures.
  • user input may be acquired by displaying an image corresponding to the image data, and receiving user input via cursors and other user input devices, the input indicating the desired manual contouring on the display.
  • Updated image data may consist of a display of the target area of the same subject acquired the initial image data at a subsequent point in time.
  • Updated image data may be acquired, for example, after a diagnostic process and during the course of treatment.
  • the image data may be acquired by a treatment device equipped with CT or CBCT imaging device.
  • Updated image data may also be acquired in subsequent sessions of the same diagnostic process in which the initial image data was acquired, and generated by the same imaging device as the imaging device which generated the initial image data.
  • the same or a substantial majority of structures depicted in the image according to the initial image data may also be depicted in the updated image data.
  • the updated image data may represent a display of the same or substantially equivalent anatomical region(s) for a target subject (e.g ., a patient or a portion of a patient's anatomy).
  • a target subject e.g ., a patient or a portion of a patient's anatomy.
  • Alternate embodiments may include less or greater amounts of structures, or a like number of structures, with varying degrees of orientation, position, and axial disposition, such that at least one structure is in common between the updated image data and the initial image data.
  • a relationship between the updated image data and the initial image data is established.
  • establishing a relationship between the updated image data and the initial image data may be performed by registering the respective data in an image software application. Registration may be performed by receiving manual input from a user (through a user interface, for example) which associates a file comprising the updated image data with a file comprising the initial image data within an image editing or viewing application .
  • the relationship may be pre-defined prior to the acquisition of the updated image data. For example, images identified for a particular subject or patient may be automatically related (e.g. registered) within the application once the image data is acquired. According to some embodiments, multiple images may be registered together simultaneously.
  • a data set comprising multiple images may be registered automatically.
  • entire data sets may be related to each other, such that each image of each data set is mutually associated with every other image of every data set within the application .
  • the initial image data may be edited at step 209 to include the contouring received at step 203. Editing the initial image data may be performed by, for example, adding the data corresponding to the contouring to the initial image data. In further embodiments, the edited initial image data may be stored for future reference.
  • a mapping mechanism is generated for cataloguing any content deviation determined between the initial and updated image data.
  • the mapping mechanism may be implemented as a deformation map derived by comparing the updated image data with the initial image data and determining the presence of deviations of common image features between targeted and identified features common to both images.
  • a deformation map may be generated for each image in a pairing, wherein the deviation of common features may be mapped while non common features may be disregarded during the mapping procedure.
  • a distinct deformation map for each common feature within an image data pairing may be generated.
  • content deviation may be detected by comparing pixel data of the pixels comprising one or more common features in the initial image data to pixel data of the common features in the updated image data. For example, relative pixel intensities of the pixels comprising a common feature of the initial image data may be compared to pixel intensities of the pixels comprising the same feature on the updated image data. In such fashion, a correspondence between pixels for the same features on different images may be derived, and a deformation map (or other such mechanism) may be generated. This correspondence may be represented as a three dimensional vector field and/or expressed by a plurality of mathematical functions which define the relativity between two corresponding points.
  • the manually (or automatically) contoured effects identified in the initial image data received at step 203 are automatically propagated to the updated image data acquired at step 205 according to the mapping mechanism (e.g., deformation map) generated at step 211 .
  • the contoured effects may be propagated automatically to every image registered to the initial image (or every data set registered to the data set of the initial image).
  • a generated deformation map is invertible, and, accordingly, once a deformation map has been generated, contoured effects received for updated image data may be propagated to the initial image data according to the same process.
  • Propagating the contoured effects may be performed according to step 107 described above with respect to Figure 1 .
  • any contoured effects may be propagated by applying the deformation map (or other mapping mechanism acquired at step 211 ) to the contoured effects in the initial image and adding the resulting output to the data of the updated image (or vice versa).
  • any contoured effects from the initial image will be propagated to the updated image while automatically being adjusted to the mapped disparities between the two images, therefore resulting in a propagation of the contoured effects that is specific and customized to the more recent image data.
  • a propagation of contoured effects from an updated image to the initial image will create corresponding contoured effects on the initial image that are customized for the initial image.
  • Figure 3 is an illustration of an exemplary propagation between related images in a data set.
  • a first image 301 and an associated second image 317 may represent images of a portion of a subject anatomy generated from a medical imaging system, for example, such as CT images or CBCT images.
  • These images may include structures such as organs, or vessels or other anatomical units.
  • these structures may be manually (e.g., through a user interface) or automatically (e.g ., through a software procedure) delineated and identified.
  • the first image may comprise an earlier "source” image generated from (and graphically representing) a portion of a subject anatomy and the second image may comprise a chronologically later "target" image of the same or similar portion of the subject anatomy.
  • the source image may be generated from a subject during an earlier diagnostic period, whereas the target image may be generated from the same subject at a later date, after a treatment or routine has been applied.
  • the target image may also be generated from a different subject (e.g., patient) but comprise a display of the same or substantially similar portion of an anatomy.
  • the image data generated by medical imaging devices may be enhanced by manual or automatic contouring.
  • the contouring may be used, for example, to delineate, emphasize or target specific portions of the image.
  • manually or automatically contoured effects in the first or "source” image 301 may be automatically propagated to a second, associated "target” image 317 through the execution of the methods described above with respect to Figures 1 and 2.
  • the association between the first and second image may be predefined within an application, such as an image manipulation and/or image display application .
  • the association may be definitively and explicitly established through received manual input (e.g., from a user through an on-screen user interface).
  • the associations may be automatically established once certain preconditions are met (e.g., same identified subject, same identified storage location, etc.)
  • images within a data set may further comprise one or more layers.
  • the first image 301 is presented with a plurality of layers (e.g., layers 303, 305).
  • the identified features may be grouped and/or arranged among the one or more layers.
  • organs may be presented on one layer
  • the cardiovascular system may be presented on a second layer
  • the skeletal system may be presented on a third
  • contoured effects may be separated from other features and arranged within an exclusive layer.
  • the layers comprising the first image 301 may correspond to layers of the second image 317.
  • contour layer 319 of the second image 317 corresponds to contour layer 303 of the first image 301
  • feature layer 321 of the second image 317 corresponds to feature layer 305 of the first image 301 .
  • like- identified layers among associated images may be automatically associated within an application or platform.
  • user-defined associations may also be created.
  • image 303 includes a feature layer 305 that includes a feature (e.g., feature 307).
  • the feature may, for example, represent an anatomical organ or other region in a target anatomy.
  • the same anatomical organ or region may also be represented in the feature layer 321 of the second image 317 as feature 325.
  • feature 325 appears smaller than feature 307.
  • the specific pixel disparities between two features or units within a pair (or more) of images may be mapped by a deformation mechanism (e.g., deformation map 311 ).
  • mapping may be performed by determining correspondences in the pixels comprising one or more features (e.g., feature 309 and feature 325).
  • the correspondences may be mapped by generating a map of pixels for each image.
  • Each pixel map (e.g., deformation map) may be generated specifically for each image, and plots the spatial relativity between the features (via pixels) comprising the image in the deformation map.
  • a correspondence between the deformation map of the first image 301 and the deformation map of the second image 317 may be mapped by, for example, determining relative pixel intensities of the pixels comprising the feature on each image, determining a correspondence (e.g., an equivalence) between a pixel in the first image 301 and a pixel in the second image 317 based on the pixel intensities, and determining the relative displacement between related pixels in the respective deformation maps of the images.
  • a correspondence e.g., an equivalence
  • the pixel intensity for any pixel in structure 309 relative to adjacent pixels may be determined and associated with a pixel in structure 325 with the same or substantially equivalent relative pixel intensity.
  • a one to one mapping may be generated for each pixel of the structure(s) comprising the images 301 , 317. Once the pixels comprising each feature are associated with an equivalent pixel in a related image, the relative displacement between each pixel of the source image 301 and its equivalent pixel in the target image 31 7 may be determined and mapped.
  • This relative displacement may be implemented as a registration map (e.g ., 311 ) mapping the relativities between a plurality of deformation maps representing each respective image (301 , 31 7).
  • a registration map e.g ., 311
  • the specific deformation between each pixel in deformation map 313 (corresponding to image 301 ) and deformation map 315 (corresponding to image 317) may be determined as a vector, with the aggregated relativities comprising a vector field.
  • Other points of data (e.g., pixel) within image 301 may be thus similarly modified for image 317 by applying an equivalent or substantially equivalent vector.
  • the underlying algebraic equations that express the vectors comprising the vector field may be used to determine deformation (displacement).
  • contoured effects in one image may be propagated to another associated image.
  • the contoured structure 307 in the contour layer 303 may be propagated in the contour layer 31 9 of image 31 7.
  • the replicated effects may be modified according to the deformation mechanism 311 to more accurately reflect the subject anatomies.
  • the relativity between feature 309 and feature 325 includes a change (e.g., a change in dimension, shape, axis, orientation, etc.)
  • an equivalent change may be experienced by the contoured effect, once propagated.
  • exemplary feature 325 comprises a smaller total area than feature 309.
  • the contoured effect 307, once adapted by the deformation mechanism 311 may also be reflected as a smaller total area in the second image as contoured effect 323, thereby providing automatic propagation that provides an adaptive output responsive to any changes over time.
  • FIG. 4 depicts a flowchart of a method for editing local structures in one or more data sets, in accordance with embodiments of the present invention.
  • Steps 401 -409 describe exemplary steps comprising the process depicted in flowchart 400 in accordance with the various embodiments herein described.
  • the flowchart 400 is implemented as computer-executable instructions stored in a computer- readable medium and performed by a computing device executing a process for automatically editing structural data among one or more data sets.
  • a first image is accessed by, for example, a computing device executing an image manipulation software application.
  • the image data represents a graphical display of a portion of the target subject's anatomy, such as data acquired according to a diagnostic procedure (e.g ., X-ray, MRI, etc.)
  • anatomical structures e.g., organs, vessels, system units
  • the image may have a corresponding deformation mechanism that maintains the relativity between the image and other registered images.
  • the deformation mechanism may be implemented to include an identity map and a registration map. According to such embodiments, the identity map may be used to map the relative positions of the anatomical structures within the same image.
  • the identity map may be implemented as, for example, a coordinate grid representing a two or three dimensional space, wherein a plurality of structures occupies space within the grid .
  • a specific position of any point within such a space may be expressed as a set of values or coordinates, with each structure being the aggregate of a plurality of points.
  • the identity map may be implemented by mapping the relative position of specific points comprising each structure to the points of adjacent structures.
  • the image data may be stored by and accessed from the same image manipulation software application .
  • User input received through a graphical user interface of the software application may direct the storage, access, and manipulation of the file(s) comprising the image data within the computing device executing the application, or other communicatively coupled computing devices.
  • the registration map may be implemented as a linked series of identity maps corresponding to pre-associated images. Common structures shared between an identity map of an image and an identity map corresponding to another associated image may be mapped together within the registration map.
  • an input corresponding to a manual adjustment or editing of one or more structures of the first image (local structures) may be received at step 403.
  • the structures which are "local" to (e.g., disposed on) the image being viewed or edited, may be adjusted for accuracy, by correcting an artifact or to clarify one or more generated structures, for example.
  • previous manual contouring performed according to user input may be corrected or adjusted.
  • the structural editing may be received through a graphical user interface generated by a software application.
  • user input may be acquired by displaying an image corresponding to the image data, and receiving user input via one or more cursors and other user input devices, the input indicating the desired structural on the display. For example, an outline or shape of a structure may be thus edited.
  • the deformation mechanism corresponding to the first image is referenced. Referencing the deformation mechanism may be achieved by the executing software application on data stored in memory of a computing system, for example.
  • the structure corresponding to the user input received at step 403 is edited by editing the identity map of the first image in the deformation mechanism according to the user input.
  • the plurality of pixels that comprise the structure to be edited may be adjusted in the identity map to conform to, and represent, the structure as edited to take into account the user input received at step 403.
  • an additional identity map may be created for the edited structure, and mapped to the original identity map corresponding to the first image.
  • the unedited structures may be duplicated in the new identity map and modified to conform to the same relativity mapped between the structure prior to editing in the original identity map and the edited structure in the newly created identity map.
  • Modifying the structures may be performed by, for example, modifying the values corresponding to one or more points comprised in the modified structure in the identity map.
  • the manual structural edits to the first image received as user input at step 403 are automatically propagated to the other structures of the first image according to the identity map corresponding to the first image.
  • the effects of the structural edit may be propagated automatically to every other image registered to the first image (or every data set registered to the data set of the first image), as described below with respect to Figure 6.
  • Propagating the structural edits may be performed by applying the deformation mechanism (e.g., the identity map and/or the registration map) to the first image and adjusting the remaining structures to account for the effect of the edited structure according to the identity map.
  • the deformation mechanism e.g., the identity map and/or the registration map
  • the structures in the first image will retain their relative sizes and positions (e.g., orientation, proportionality). Therefore resulting in a propagation of the effect of the initial structural edits to the other local structures that is specific and customized to the original image data.
  • FIG. 5 is an illustration of simultaneous editing of a plurality of structures in an image 501 .
  • an image 501 comprising a plurality of layers (e.g., contour layer 503 and structure layer 505), each with one or more effects (e.g., contour effect 507 and structure 509, respectively).
  • the image 501 may be the image 301 described above with respect to Figure 3.
  • data for the image 501 may be a pre-generated image acquired at an earlier point in time and stored as a computer readable medium, for example.
  • the other remaining structures comprising the image 501 may be likewise modified to account for the edited structures, such that a specific, simultaneous editing may be automatically achieved.
  • the image 501 may be accessed (e.g ., via an image manipulation or image viewing application) and displayed (e.g., through a graphical user interface) to a user.
  • the one or more structures of the image 501 may be edited after the image has been referenced .
  • the structures may be edited to add, modify, refine or remove contouring or other structures for example.
  • the editing may be performed through the user interface of the same image manipulation or image viewing application used to view the application.
  • a new identity map 513 is created and a registration map 527 maps the identity map 51 3 of the updated image to the identity map 511 of the original image.
  • the unedited structures in the original image may be modified to reflect and/or account for the edited structures.
  • the modification may be achieved by, for example, applying to each unedited structure the adjustment made to the edited structures modified according to the mapped relationship between the identity maps (e.g ., identity map1 511 and identity map2 513) in the registration map 527 to consider the relativity of the structures, thus enabling an automatic and simultaneous editing of each of the structures comprising an image that generates a specific and adaptive output for each structure.
  • the resulting image (e.g., image 517) with modified structures (e.g., 523, 525) may include a corresponding number of layers (e.g., contouring layer 519, structure layer 521 ) as the first image (e.g., image 501 ).
  • the resulting image 517 may be stored with or, alternatively, may replace the original image (e.g., image 501 ) when referenced at a subsequent time.
  • the deformation mechanism may be implemented as a vector field.
  • the automatic propagation properties described above with reference to Figure 3 can be extended to include manually edited structures, such as the manually edited structures described above with respect to Figure 4.
  • Figure 6 is an illustration of automatic propagation of an edited plurality of structures in an image 601 . Accordingly, Figure 6 is depicted as combining features from a process of editing local structures (e.g., the process described above with respect to Figures 4 and 5) with a process for automatically propagating modifications to associated images (e.g., the process described above with respect to Figures 1 -3).
  • an image 601 comprising a plurality of layers (e.g., contour layer 603 and structure layer 605), each with one or more structures (e.g., contour effect 607 and structure 609, respectively).
  • the image 601 may be the image 301 described above with respect to Figure 3 and/or 5.
  • data for the image 601 may be a pre-generated image acquired at an earlier point in time and stored as a computer readable medium, for example.
  • the other remaining local structures comprising the image 601 may be likewise modified to account for the edited structures, such that a specific, simultaneous editing may be automatically achieved.
  • the structures of associated images may also be edited to conform to the edited structure, while accounting for pre-existing dissimilarities.
  • images of the same structures e.g., anatomy
  • images of the same structures taken over a period of time that exhibit progress (or regress) may depict those structures in drastically different sizes, shapes, and orientations. This may be particularly true of areas targeted for radiation treatment. Accordingly, editing a structure in one image (e.g., an image obtained during the treatment planning or diagnostic stage) may not produce completely appropriate and/or accurate results to apply a direct equivalent editing in a later image.
  • editing local structures of a first image may be performed as described above in Figure 5.
  • images of different data sets or even corresponding to different image standards may be likewise affected.
  • the image 601 may be generated from a computer tomography (CT) image device.
  • CT computer tomography
  • an image of the same or substantially similar portion of a target anatomy displayed in image 601 may be likewise generated by an image device of a separate standard or method.
  • CBCT cone-beam computer tomography
  • a CBCT image (e.g., image 623) associated with a CT image (e.g., image 601 ) and displaying common structures (e.g ., structure 625 and 627) may also be automatically modified to according to the same relativity (e.g., captured in a registration map 611 ) mapped between the structures the identity map1 613 of the first image 601 and the identity map2 615 of the image 623, resulting in a modified CBCT image (e.g., image 629) with edited structures (e.g., structure 631 and 633).
  • a modified CBCT image e.g., image 629 with edited structures (e.g., structure 631 and 633).
  • the identity map 613 mapping the structures of the first image 601 when modified (e.g ., by editing one or more structures of the first image 601 ), the identity map 613 may be compared to an identity map 619 generated from the modification to one or more structures of the first image 601 .
  • the edited structures may be subsequently mapped to corresponding structures of the first identity map 613 (e.g ., in a registration map 61 7).
  • Unedited structures from the first identity map 61 3 may be duplicated in the generated identity map 619 while being adjusted according to the same relativity mapped between the structures of the first identity map 613 and the manually edited counterparts in the new identity map 619.
  • a structure such as a tumor or mass that has been edited to be of a reduced size may cause an associated structure, such as a manually contoured target volume, to be of proportionally reduced size.
  • modifications to the identity map 613 may alter the image displayed (e.g., in a display device to the editing user). As presented, the modifications to the identity map 613 may be displayed as a resulting image 635, with corresponding edited structures (e.g., structure 637 corresponds to structure 607, and structure 639 corresponds to structure 609). Accordingly, the specific propagation of contoured effects and manually edited structures may be automatically propagated to other local structures as well as affiliated images.
  • an exemplary system upon which embodiments of the present invention may be implemented includes a general purpose computing system environment, such as computing system 700.
  • computing system 700 typically includes at least one processing unit 701 and memory, and an address/data bus 709 (or other interface) for communicating information.
  • memory may be volatile (such as RAM 702), non-volatile (such as ROM 703, flash memory, etc.) or some combination of the two.
  • Computer system 700 may also comprise an optional graphics subsystem 705 for presenting information to the computer user, e.g., by displaying information on an attached display device 710, connected by a video cable 711 .
  • the graphics subsystem 705 may be coupled directly to the display device 710 through the video cable 711 .
  • a graphical user interface of an image viewing software application executing in the computer system 700 may be generated in the graphics subsystem 705, for example, and displayed to the user in the display device 710.
  • display device 710 may be integrated into the computing system (e.g., a laptop or netbook display panel) and will not require a video cable 711 .
  • the processes 100, 200, and 300 may be performed, in whole or in part, by graphics subsystem 705 in conjunction with the processor 701 and memory 702, with any resulting output displayed in attached display device 710.
  • computing system 700 may also have additional features/functionality.
  • computing system 700 may also include additional storage (removable and/or non-removable) including, but not limited to, magnetic or optical disks or tape.
  • additional storage is illustrated in Figure 7 by data storage device 707.
  • Computer storage media includes volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data.
  • RAM 702, ROM 703, and data storage device 707 are all examples of computer storage media.
  • Computer system 700 also comprises an optional alphanumeric input device 706, an optional cursor control or directing device 707, and one or more signal communication interfaces (input/output devices, e.g., a network interface card) 708.
  • Optional alphanumeric input device 706 can communicate information and command selections to central processor 701 .
  • Optional cursor control or directing device 707 is coupled to bus 709 for communicating user input information and command selections to central processor 701 .
  • Signal communication interface (input/output device) 708, also coupled to bus 709, can be a serial port.
  • Communication interface 709 may also include wireless communication mechanisms.
  • computer system 700 can be communicatively coupled to other computer systems over a communication network such as the Internet or an intranet (e.g., a local area network), or can receive data (e.g., a digital television signal).

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Apparatus For Radiation Diagnosis (AREA)
  • Magnetic Resonance Imaging Apparatus (AREA)
  • Image Processing (AREA)
  • Image Analysis (AREA)

Abstract

Selon des modes de réalisation, la présente invention concerne des procédés et un mécanisme de manipulation d'images générées par des machines de radiothérapie utilisées dans des applications de radiodiagnostic et de traitement par rayonnement. Un mode de réalisation porte sur un procédé de propagation automatique intelligente de contours détectés manuellement ou automatiquement dans des images liées (par exemple superposées) et des ensembles de données d'image correspondants, par l'acquisition d'une ou de plusieurs images d'un ou de plusieurs ensembles de données d'image ; de détermination de la corrélation entre les images relativement à des structures identifiées, et de génération d'une carte de déformation qui établit une correspondance pour chaque point dans l'image source avec un point dans l'image cible. Ensuite, le mécanisme de propagation intelligente applique cette carte de déformation individuellement à chaque structure de l'image source et propage la structure déformée à l'image cible.
PCT/US2011/041520 2010-06-23 2011-06-22 Mécanisme de génération et d'édition de structures avancées Ceased WO2011163414A2 (fr)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP11798881.6A EP2584970A4 (fr) 2010-06-23 2011-06-22 Mécanisme de génération et d'édition de structures avancées
CN201180039643.1A CN103079467B (zh) 2010-06-23 2011-06-22 用于高级结构生成和编辑的机制
JP2013516750A JP5926252B2 (ja) 2010-06-23 2011-06-22 高度な構造生成および編集のためのメカニズム

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US12/821,977 2010-06-23
US12/821,977 US9401051B2 (en) 2010-06-23 2010-06-23 Mechanism for dynamically propagating real-time alterations of medical images
US12/821,985 2010-06-23
US12/821,985 US10311585B2 (en) 2010-06-23 2010-06-23 Mechanism for advanced structure generation and editing

Publications (2)

Publication Number Publication Date
WO2011163414A2 true WO2011163414A2 (fr) 2011-12-29
WO2011163414A3 WO2011163414A3 (fr) 2012-04-05

Family

ID=45372091

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2011/041520 Ceased WO2011163414A2 (fr) 2010-06-23 2011-06-22 Mécanisme de génération et d'édition de structures avancées

Country Status (4)

Country Link
EP (1) EP2584970A4 (fr)
JP (2) JP5926252B2 (fr)
CN (1) CN103079467B (fr)
WO (1) WO2011163414A2 (fr)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015011816A1 (fr) * 2013-07-25 2015-01-29 株式会社島津製作所 Appareil de fluoroscopie et procédé de fluoroscopie
GB2542666A (en) * 2015-07-21 2017-03-29 Canon Kk Image processing apparatus, image processing method, and storage medium
US9724540B2 (en) 2014-02-24 2017-08-08 National Institutes For Quantum And Radiology Science And Technology Moving-body tracking device for radiation therapy, irradiation region determining device for radiation therapy, and radiation therapy device

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA3000050C (fr) * 2015-11-03 2022-02-08 Ventana Medical Systems, Inc. Image de tissu composite mise en oeuvre par ordinateur avec une interface reglable en temps reel
CN109964249A (zh) * 2016-09-21 2019-07-02 皇家飞利浦有限公司 用于对身体部分的自适应轮廓勾画的装置
CN106843694B (zh) * 2017-02-13 2020-07-21 上海联影医疗科技有限公司 三维图像处理系统
FR3068813B1 (fr) * 2017-07-10 2021-07-02 Univ Aix Marseille Procede et dispositif de segmentation d'images par propagation automatique dans une (n+1)-ieme dimension d'une segmentation d'images initialisee en dimension n

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7460699B2 (en) * 2004-03-05 2008-12-02 Siemens Medical Solutions Usa, Inc. System and method for a semi-automatic quantification of delayed enchancement images
US20050251029A1 (en) * 2004-04-21 2005-11-10 Ali Khamene Radiation therapy treatment plan
US7352370B2 (en) * 2005-06-02 2008-04-01 Accuray Incorporated Four-dimensional volume of interest
US7376246B2 (en) * 2005-06-27 2008-05-20 Mitsubishi Electric Research Laboratories, Inc. Subspace projection based non-rigid object tracking with particle filters
WO2007007276A2 (fr) * 2005-07-14 2007-01-18 Koninklijke Philips Electronics Methode de prise en compte du deplacement d'une tumeur lors d'une radiotherapie
CN101529442A (zh) * 2005-07-22 2009-09-09 断层放疗公司 在形变图上放置约束的方法及实现该方法的系统
CN101820948A (zh) * 2007-10-25 2010-09-01 断层放疗公司 用于放疗实施的运动适应性优化的系统和方法

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of EP2584970A4 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2015011816A1 (fr) * 2013-07-25 2015-01-29 株式会社島津製作所 Appareil de fluoroscopie et procédé de fluoroscopie
JPWO2015011816A1 (ja) * 2013-07-25 2017-03-02 株式会社島津製作所 X線透視装置およびx線透視方法
US10117630B2 (en) 2013-07-25 2018-11-06 Shimadzu Corporation Fluoroscopy apparatus and fluoroscopy method
US9724540B2 (en) 2014-02-24 2017-08-08 National Institutes For Quantum And Radiology Science And Technology Moving-body tracking device for radiation therapy, irradiation region determining device for radiation therapy, and radiation therapy device
GB2542666A (en) * 2015-07-21 2017-03-29 Canon Kk Image processing apparatus, image processing method, and storage medium
US9871947B2 (en) 2015-07-21 2018-01-16 Canon Kabushiki Kaisha Image processing apparatus, image processing method, and storage medium
GB2542666B (en) * 2015-07-21 2018-12-05 Canon Kk Image processing apparatus, image processing method, and storage medium

Also Published As

Publication number Publication date
WO2011163414A3 (fr) 2012-04-05
CN103079467A (zh) 2013-05-01
EP2584970A4 (fr) 2017-08-02
EP2584970A2 (fr) 2013-05-01
JP2013529509A (ja) 2013-07-22
JP6397846B2 (ja) 2018-09-26
JP2016129796A (ja) 2016-07-21
JP5926252B2 (ja) 2016-05-25
CN103079467B (zh) 2016-06-29

Similar Documents

Publication Publication Date Title
US10311585B2 (en) Mechanism for advanced structure generation and editing
TWI663961B (zh) 物件定位裝置、物件定位方法、物件定位程式及輻射治療系統
JP7030050B2 (ja) 組織パラメータ推定を用いたmrデータからの疑似ct生成
JP6567179B2 (ja) 特徴回帰モデルを用いたmrデータからの疑似ct生成
CN106920234B (zh) 一种复合式自动放疗计划的方法
JP6397846B2 (ja) 高度な構造生成および編集のためのメカニズム
US8457372B2 (en) Subtraction of a segmented anatomical feature from an acquired image
WO2018119766A1 (fr) Système et procédé de traitement d'images multi-modales
JP7055872B2 (ja) 二次元/三次元画像コレジストレーションのためにコンピュータによって実現される情報処理方法、プログラム、医用画像分析システム、および、放射線療法治療システム
US10032316B2 (en) System for dynamically propagating real-time alterations of medical images
US9492124B2 (en) System and method for treatment planning of organ disease at the functional and anatomical levels
US10628963B2 (en) Automatic detection of an artifact in patient image
CN107865658A (zh) 用于修正合成电子密度图的方法和设备
US10034610B2 (en) System and method for registration of brain images
US11565129B2 (en) Binary tracking of an anatomical tracking structure on medical images
NL2016800B1 (en) Method of positioning an interventional device.
US20250339711A1 (en) A method for quantifying patient set up errors in radiotherapy
AU2023263690B2 (en) A method for quantifying patient set up errors in radiotherapy
Miandoab et al. Extraction of respiratory signal based on image clustering and intensity parameters at radiotherapy with external beam: A comparative study
Sewa Motion Determination Of Lung Tumours Based On Cine-MR Images
WO2025195577A1 (fr) Procédé de détermination de données de positionnement d'un sujet dans un système de radiothérapie
Starzynski et al. Conformance criteria for validation of target volume surface reconstructed from delineation
Schreibmann et al. -Atlas-Based Segmentation: Concepts and Applications

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 201180039643.1

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11798881

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE

ENP Entry into the national phase

Ref document number: 2013516750

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 2011798881

Country of ref document: EP