WO2019208037A1 - 画像解析方法、セグメンテーション方法、骨密度測定方法、学習モデル作成方法および画像作成装置 - Google Patents
画像解析方法、セグメンテーション方法、骨密度測定方法、学習モデル作成方法および画像作成装置 Download PDFInfo
- Publication number
- WO2019208037A1 WO2019208037A1 PCT/JP2019/011773 JP2019011773W WO2019208037A1 WO 2019208037 A1 WO2019208037 A1 WO 2019208037A1 JP 2019011773 W JP2019011773 W JP 2019011773W WO 2019208037 A1 WO2019208037 A1 WO 2019208037A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- organ
- ray
- drr
- subject
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5211—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data
- A61B6/5217—Devices using data or image processing specially adapted for radiation diagnosis involving processing of medical diagnostic data extracting a diagnostic or physiological parameter from medical diagnostic data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/02—Arrangements for diagnosis sequentially in different planes; Stereoscopic radiation diagnosis
- A61B6/03—Computed tomography [CT]
- A61B6/032—Transmission computed tomography [CT]
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/50—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment specially adapted for specific body parts; specially adapted for specific clinical applications
- A61B6/505—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment specially adapted for specific body parts; specially adapted for specific clinical applications for diagnosis of bone
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/52—Devices using data or image processing specially adapted for radiation diagnosis
- A61B6/5205—Devices using data or image processing specially adapted for radiation diagnosis involving processing of raw data to produce diagnostic data
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H30/00—ICT specially adapted for the handling or processing of medical images
- G16H30/40—ICT specially adapted for the handling or processing of medical images for processing medical images, e.g. editing
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10116—X-ray image
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30004—Biomedical image processing
- G06T2207/30008—Bone
Definitions
- the present invention relates to an image analysis method, a segmentation method, a bone density measurement method, a learning model creation method, and an image creation device.
- Patent Document 1 discloses collimating only a means for generating radiation, a single crystal lattice irradiated with the radiation, and radiation having a predetermined two reflection angles out of the radiation reflected by the crystal lattice.
- radiation detection means By means of simultaneously irradiating the subject with radiation of different energies, radiation detection means through which the radiation of these two energies passes through the subject, and analyzing the pulse height of the output of the radiation detection means, respectively
- An apparatus for quantitatively analyzing bone mineral which comprises a pulse height analyzing means for separating transmission data relating to radiation of different energy, and a means for calculating bone density by processing the separated data.
- This measurement of bone density is targeted at the bone density of the lumbar vertebrae and femur, which requires clinical attention.
- the femur has a large individual difference in shape, and in order to perform stable follow-up observation, it is important to specify the bone region of the subject.
- the operator manually specifies the region, which is not only complicated, but also has a problem that the region specified by the operator varies.
- the present invention has been made to solve the above problems, and an image analysis method capable of creating an image in which an organ region is accurately extracted from an X-ray image of a region including a subject's organ.
- An object of the present invention is to provide a segmentation method, a bone density measurement method, a learning model creation method, and an image creation device.
- the invention according to claim 1 is an image analysis method for performing segmentation for identifying a region of the organ by analyzing an image of the region including the organ of the subject, and machine learning is used as the segmentation technique. And a modified image creating step of creating a modified image in which the density of the region of the organ in the image including the organ of the subject is changed, and an image including the organ of the subject and the modified image creating step And a learning model creation step of creating a machine learning learning model by a learning process using the corrected image created in the above.
- the invention according to claim 2 is the learning model according to claim 1, wherein the learning model is used for an X-ray image of a region including the subject's organ obtained by X-ray imaging of the subject.
- An image representing the organ is created by performing conversion using the learning model created in the creation step.
- the invention according to claim 3 is the invention according to claim 1, wherein the image of the region including the organ of the subject is a DRR image created from CT image data of the subject, and the correction In the image creating step, the density of the CT image data is changed using the region where the CT value of the CT image data is a predetermined value as the region of the organ.
- a parameter including at least one of a projection coordinate and an angle of the geometric condition is changed, or an image is rotated.
- Image processing including at least one of deformation and enlargement / reduction is performed to create a plurality of DRR images.
- At least one of contrast change, noise addition, and edge enhancement is performed on the created DRR image.
- the invention according to claim 6 is the invention according to claim 1, wherein the image of the region including the organ of the subject is an X-ray image created by X-ray imaging of the subject.
- the density of the organ region is changed using the X-ray image and the image of the organ obtained by using dual energy subtraction.
- an X-ray image of a region including the subject's organ obtained by X-raying the subject and the learning model creating step The image representing the organ obtained by performing the conversion using the learning model created in the above is used for learning of the learning model by the learning unit.
- the invention according to claim 8 is the invention according to claim 1, wherein the organ has a symmetrical shape with respect to the body axis of the subject, and in the pre-learning model creation step, the right organ A machine learning learning model is created collectively for the left and right organ images by horizontally flipping either the left image or the left organ image.
- the region of the bone part is segmented using the image analysis method according to the first aspect, wherein the organ is the bone part of the subject.
- the bone density is measured for the bone region segmented by the segmentation method described in claim 9.
- the invention according to claim 11 is a learning model used when performing segmentation for specifying the region of the organ by analyzing an image of the region including the organ of the subject using machine learning.
- An image including the subject's organ and a modified image generated by changing the density of the region of the organ in the image including the subject's organ. Used to create a learning model by executing machine learning.
- the invention according to claim 12 is an image creation device for creating an image obtained by extracting an area of the organ from an X-ray image of the area including the organ of the subject, and X-rays the area including the organ.
- An X-ray image storage unit that stores a plurality of X-ray images obtained in this manner, a plurality of X-ray image teacher images for machine learning, and a DRR image generation unit that generates a DRR image of an area including the bone part
- a DRR image that stores a plurality of DRR images created by the DRR image creation unit and a plurality of machine learning DRR image teacher images created based on the DRR image created by the DRR image creation unit
- Machine learning is performed using the storage unit, the plurality of X-ray images stored in the X-ray image storage unit, and the plurality of X-ray image teacher images, and is stored in the DRR image storage unit.
- the plurality of DRR images X-rays of a region including the organ of the subject using a learning model for recognizing the organ, which is created in advance by performing machine learning using the plurality of teacher images for DRR images
- an image creation unit that creates an image representing the organ by performing conversion on the image.
- the DRR image creation unit may include a part of the plurality of DRR images as a part of the region including the bone part. Created as a DRR image in which the density of the region is changed.
- the invention described in claim 14 is the invention described in claim 11, wherein a part of the plurality of X-ray images stored in the X-ray image storage unit uses dual energy subtraction.
- This is an X-ray image in which the density of the organ region of the region including the organ is changed.
- the corrected image in which the density of the organ region of the subject is changed is used for machine learning, it can be applied to a subject having a low organ density. It is possible to create a learning model. For this reason, it becomes possible to improve the detection accuracy of an organ.
- the parameters including the projection coordinates and angles of the geometric perspective conditions are changed, or image processing including rotation, deformation, and enlargement / reduction of the image is performed.
- image processing including rotation, deformation, and enlargement / reduction of the image.
- the fifth aspect of the present invention since contrast change, noise addition, and edge enhancement are performed on the created DRR image, even when there is a difference in image quality between the DRR image and the X-ray image. In addition, the position of the bone part can be accurately detected.
- the seventh aspect of the present invention by reusing a plurality of X-ray images and an image representing an organ obtained by performing conversion using a learned learned model for learning of a learning model, It is possible to create a learning model with higher accuracy by expanding the learning image.
- the corrected image in which the density of the organ region of the subject is changed is used for machine learning, a learning model corresponding to the subject having a low organ density is also provided. It becomes possible to create.
- FIG. 1 is a schematic front view of a bone image creating apparatus according to an embodiment of the present invention that also functions as an X-ray imaging apparatus.
- 1 is a schematic side view of a bone image creating apparatus according to an embodiment of the present invention that also functions as an X-ray imaging apparatus.
- It is a block diagram which shows the control system of the bone part image creation apparatus which concerns on embodiment of this invention.
- It is a schematic diagram for demonstrating the process of producing the bone part image of a subject using machine learning with the bone part image creation apparatus which concerns on embodiment of this invention.
- FIG. 3 is a schematic diagram of an X-ray image 101 created by an X-ray image creation unit 81.
- FIG. FIG. 6 is a schematic diagram of a teacher bone image for X-ray image 102 created by an X-ray image creation unit 81. It is explanatory drawing which shows typically the state which produces a DRR image by the virtual projection which simulated the geometric condition of the X-ray irradiation part 11 and the X-ray detection part 12 which are shown in FIG.
- FIG. 6 is a schematic diagram of a DRR image 103 created by a DRR image creation unit 83.
- FIG. FIG. 10 is a schematic diagram of a DRR image 104 in which the density of the bone region created by the DRR image creating unit 83 is changed to a small value.
- 6 is a schematic diagram of a DRR image teacher bone image 105 created by a DRR image creation unit 83.
- FIG. 3 is a schematic diagram of an X-ray image 106 created by an X-ray image creation unit 81.
- FIG. 6 is a schematic diagram of a DRR image 107 created by a DRR image creation unit 83.
- FIG. 1 is a schematic front view of a bone image creating apparatus according to an embodiment of the present invention that also functions as an X-ray imaging apparatus
- FIG. 2 is a schematic side view thereof.
- the present invention is applied to a bone image creating apparatus that creates an image of a bone part of a subject among organs such as a bone part and an organ.
- This bone image creating apparatus that also functions as an X-ray imaging apparatus is also referred to as an X-ray fluoroscopic imaging table, and includes a top plate 13, an X-ray tube holding member 15, and an X-ray tube holding member 15.
- X-ray irradiation unit 11 disposed at the tip, and X-rays such as a flat panel detector and an image intensifier (II) disposed on the opposite side of the X-ray irradiation unit 11 with respect to the top plate 13
- an X-ray detection unit 12 having a detector.
- the top plate 13, the X-ray tube holding member 15, the X-ray irradiation unit 11, and the X-ray detection unit 12 are shown in FIG. 1 and FIG. 2 by the action of a rotation mechanism 16 incorporating a motor (not shown). It is possible to rotate between a recumbent position where the surface of 13 faces the horizontal direction and a standing position where the surface of the top plate 13 faces the vertical direction. Further, the rotation mechanism 16 itself can be moved up and down with respect to the main column 17 erected on the base plate 18.
- top 13 When the top 13 is in the prone position, X-ray imaging is performed on the subject in the prone position. At this time, the subject is placed on the top board 13. When the top 13 is in the standing position, X-ray imaging is performed on the subject in the standing position. At this time, the subject stands up in front of the top board 13.
- FIG. 3 is a block diagram showing a control system of the bone image creating apparatus according to the embodiment of the present invention.
- This bone part image creating apparatus is for creating a bone part image obtained by extracting a bone region from an X-ray image of a region including a bone part of a subject, and a CPU as a processor for executing a logical operation And a ROM that stores an operation program necessary for controlling the apparatus, a RAM that temporarily stores data and the like during control, and a control unit 80 that controls the entire apparatus.
- the control unit 80 includes an X-ray image creation unit 81 for creating an X-ray image, and a plurality of X-ray images obtained by X-ray imaging of a region including a bone part such as a subject.
- X-ray image storage unit 82 that stores a plurality of X-ray image teacher bone images for machine learning, and X-ray imaging of a subject with respect to CT image data of a region including the bone portion
- DRR image creation unit 83 a plurality of DRR images created by DRR image creation unit 83, and a plurality of machines created based on the DRR image created by DRR image creation unit 83
- a DRR image storage unit 84 for storing a training DRR image teacher bone image, a plurality of X-ray images and a plurality of X-ray image teacher bone images stored in the X-ray image storage unit 82 are used.
- Machine learning and recognizing a bone part by executing machine learning using a plurality of DRR images and a plurality of DRR image teacher bone images stored in the DRR image storage unit 84.
- a bone image creating unit 86 for creating The control unit 80 is composed of a computer in which software is installed. The functions of each unit included in the control unit 80 are realized by executing software installed in the computer.
- the learning unit 85 may perform machine learning at the stage before delivery of the device and store the result in advance, and additionally machine learning after delivery of the device to a medical institution or the like. May be executed. At this time, the learning unit 85 creates a discriminator by various learnings using arbitrary methods such as FCN (Fully Convolutional Networks), a neural machine network, a support vector machine (SVM), and boosting.
- FCN Full Convolutional Networks
- SVM support vector machine
- the control unit 80 is connected to the X-ray irradiation unit 11 and the X-ray detection unit 12 described above.
- the control unit 80 is connected to a display unit 21 configured by a liquid crystal display panel or the like and displaying various images including an X-ray image, and an operation unit 22 having various input means such as a keyboard and a mouse. Yes.
- the control unit 80 is connected online or offline to a CT image storage unit 70 that stores a CT image obtained by CT imaging of the subject.
- the CT image storage unit 70 may be included in a CT imaging apparatus, or may be included in a treatment planning apparatus that creates a treatment plan for a subject.
- FIG. 4 is a schematic diagram for explaining a process of creating a bone image of a subject using machine learning by the bone image creating apparatus according to the embodiment of the present invention.
- a learning model is created.
- an X-ray image and a DRR image of a region including a bone part are used as an input layer
- an X-ray image teacher bone image indicating a bone part and a DRR image teacher bone part image are used as an output layer
- a convolutional layer used as a learning model is learned by machine learning.
- a bone part image is created.
- An image showing a partial image is created.
- FIG. 5 is a flowchart showing an operation of creating a bone portion image obtained by extracting a bone region from an X-ray image of a region including a bone portion of a subject by the bone portion image creating apparatus according to the embodiment of the present invention. .
- an X-ray image creating step is executed (step S1).
- an X-ray image of the subject on the top 13 is obtained by using the X-ray irradiation unit 11 and the X-ray detection unit 12 shown in FIG. 1 by the X-ray image creation unit 81 shown in FIG.
- a plurality of X-ray images are created.
- an X-ray image may be obtained by taking an image taken by another X-ray imaging apparatus, or may be created by taking an X-ray image of a phantom instead of the subject.
- the created X-ray image is stored in the X-ray image storage unit 82 shown in FIG. 3 (step S2).
- a teacher bone image for X-ray image used for machine learning is created (step S3).
- This X-ray image teacher bone part image is created by the X-ray image creation part 81 by trimming the region of the bone part of the subject with respect to the previously created X-ray image. Further, when the X-ray image teacher bone image is created, an image obtained by slightly translating, rotating, deforming, and enlarging / reducing the trimmed X-ray image is also created.
- An image obtained by translating, rotating, transforming, and enlarging / reducing the trimmed X-ray image is also used for learning because the subject moves during X-ray imaging described later, or the X-ray irradiation unit 11 and the X-ray detection unit This is to cope with a case where 12 moves.
- the created X-ray image teacher bone image is stored in the X-ray image storage unit 82 shown in FIG. 3 (step S4).
- FIG. 6 is a schematic diagram of the X-ray image 101 created by the X-ray image creation unit 81
- FIG. 7 is a schematic diagram of the teacher bone image 102 for the X-ray image created by the X-ray image creation unit 81. It is.
- a femur 51, a pelvis 52, and a soft part region 53 are displayed. Further, the femur 51 and the pelvis 52 are displayed in the teacher bone image 102 for X-ray images.
- step S5 a plurality of DRR images showing the region including the bone part are created (step S5), and the DRR image is stored in the DRR image storage unit 84 (step S6).
- step S7 A plurality of teacher bone part images for DRR images showing a region including the image are created (step S7), and the teacher bone part images for DRR images are stored in the DRR image storage unit 84 (step S8).
- a DRR image teacher bone image is created with a region having a CT value equal to or greater than a certain value as a bone region.
- a DRR image teacher bone image is created by identifying a region having a CT value of 200 HU (Hounsfield Unit) or more as a bone region.
- FIG. 8 is an explanatory diagram schematically showing a state in which a DRR image is created by virtual projection simulating the geometric conditions of the X-ray irradiation unit 11 and the X-ray detection unit 12 shown in FIG.
- reference numeral 300 indicates CT image data.
- the CT image data 300 is three-dimensional voxel data that is a set of a plurality of two-dimensional CT image data.
- the CT image data 300 has a structure in which, for example, about 200 two-dimensional images of 512 ⁇ 512 pixels are stacked in a direction crossing the subject (direction along the line segment L1 or L2 shown in FIG. 8). .
- the DRR image creating unit 83 When the DRR image creating unit 83 creates a DRR image, it virtually projects the CT image data 300. At this time, the three-dimensional CT image data 300 is arranged on the computer. Then, the geometry which is the geometric arrangement of the X-ray imaging system is reproduced on the computer. In this embodiment, the X-ray irradiation unit 11 and the X-ray detection unit 12 are disposed on both sides of the CT image data 300.
- the arrangement of the CT image data 300, the X-ray irradiation unit 11, and the X-ray detection unit 12 is such that the subject when performing X-ray imaging, the X-ray irradiation unit 11, and the X-ray detection unit 12 are arranged. It has the same geometry as the arrangement.
- the term “geometry” means a geometric arrangement relationship between the imaging target, the X-ray irradiation unit 11 and the X-ray detection unit 12.
- a large number of line segments L connecting the X-ray irradiation unit 11 and each pixel of the X-ray detection unit 12 via each pixel of the CT image data 300 are set.
- two line segments L1 and L2 are shown for convenience of explanation.
- a plurality of calculation points are set on the line segment L, and the CT value of each calculation point is calculated.
- interpolation is performed using the CT value in CT data voxels around the calculation point.
- the CT values of the calculation points on the line segment L are accumulated. This accumulated value is converted into a line integral of a line attenuation coefficient, and a DRR image is created by calculating attenuation of X-rays.
- the DRR image is created by changing the parameters for creating the DRR image including at least one of the projection coordinates and the angle with respect to the CT image data 300.
- image processing including at least one of slight translation, rotation, deformation, and enlargement / reduction is executed.
- the parallel movement, rotation, deformation, and enlargement / reduction are executed in order to correspond to the case where the subject moves during the X-ray imaging described later, or the X-ray irradiation unit 11 and the X-ray detection unit 12 move. It is.
- contrast change is executed on the created DRR image.
- This contrast change, noise addition, and edge enhancement are performed in order to absorb the difference in image quality between the DRR image and the X-ray image and to more reliably recognize the bone region.
- the parameters including the projection coordinates and angle of the geometric perspective condition are changed under the same conditions, or the image is rotated, deformed, or enlarged. Image processing including reduction is performed under the same conditions.
- the DRR image creation part 83 selects a part of the DRR images from the plurality of DRR images as a bone in the region including the bone part. It is created as a DRR image in which the density of the partial area is changed. More specifically, the CT value of the bone region where the CT value is a certain value or more is set to a value smaller than the actual CT value. Thereby, it is possible to obtain a DRR image simulating a bone part having a reduced bone density.
- FIG. 9 is a schematic diagram of the DRR image 103 created by the DRR image creation unit 83.
- FIG. 10 shows a DRR image 104 in which the density of the bone region created by the DRR image creation unit 83 is changed to a small value.
- FIG. 11 is a schematic diagram of the DRR image teacher bone part image 105 created by the DRR image creation unit 83.
- the femur 51, the pelvis 52, and the soft part region 53 are displayed. Further, the femur 51 and the pelvis 52 are displayed in the DRR image teacher bone part image 105.
- the learning unit 85 executes machine learning using the X-ray image 101 shown in FIG. 6 as an input layer and the X-ray image teacher bone image 102 shown in FIG. 7 as an output layer.
- a learning model for recognizing the bone part (the femur 51 and the pelvis 52) is created (step S9).
- FCN is used for this machine learning.
- the convolutional neural network used in the FCN is configured as shown in FIG. That is, when creating a learning model, the input layer is an X-ray image 101 and DRR images 103 and 104, and the output layer is an X-ray image teacher bone image 102 and a DRR image teacher bone image 105.
- step S10 X-ray imaging is performed on the subject.
- step S11 the bone part image creation unit 86 converts the captured X-ray image by using the learning model (convolution layer) created earlier, thereby executing segmentation, and the bone part (femur) 51 and the pelvis 52) are created (step S11). That is, the learning model created previously is used for an X-ray image obtained by X-ray imaging, and an image representing a bone part is created as an output layer. Then, the bone density is measured by various methods using the bone region specified by the segmentation.
- segmentation is a concept including a process of specifying an outline of a bone or the like or a process of specifying an outline of a bone or the like in addition to the process of specifying a region such as a bone in this embodiment. is there.
- the operator corrects the created bone part image as necessary. Then, the corrected bone image and the original X-ray image are used for creating a learning model by the learning unit 85 or for re-learning. As a result, it is possible to create a learning model with higher accuracy by expanding learning images including failure examples.
- the extraction accuracy can be improved by extracting the bone region by machine learning.
- machine learning is performed using both the X-ray image and the DRR image
- the learning image can be expanded, and the collection of learning clinical data can be easily performed.
- a DRR image in which the density of the bone region is changed it is possible to perform machine learning with a DRR image simulating a bone portion having a reduced bone density, resulting in a decrease in bone density and osteoporosis. Bone extraction accuracy can be improved for patients including patients.
- the X-ray image may be input to the learning model after being blurred by a Gaussian filter or the like.
- a DRR image is created from a low-resolution CT image, it has a lower resolution than an X-ray image. For this reason, it is possible to more reliably identify bone parts by blurring the X-ray image, reducing noise in the X-ray image, and setting the resolution to be equivalent to that of the DRR image at the time of learning.
- the DRR image and the X-ray image input to the learning model may be input after performing contrast normalization in advance. Further, a local contrast normalization layer or a local response normalization layer may be added to the intermediate layer.
- the bone density is reduced by creating a part of the plurality of DRR images as a DRR image in which the density of the bone region of the region including the bone portion is changed.
- a DRR image simulating the bone part is created and used for machine learning.
- an X-ray image (high-voltage image) obtained by imaging a part of the plurality of X-ray images with a high voltage applied to the X-ray tube;
- dual energy subtraction that performs subtraction processing on an X-ray image (low-pressure image) taken with a low voltage applied to the X-ray tube, the bone region of the region including the bone portion
- An X-ray image having a changed density is used.
- an X-ray image taken with a high voltage applied to the X-ray tube and a low voltage applied to the X-ray tube A configuration is adopted in which bone density is measured by dual energy subtraction for performing subtraction processing on the X-ray image. Even at the time of specifying the bone part image, this dual energy subtraction was used to capture an X-ray image taken with a high voltage applied to the X-ray tube and a low voltage applied to the X-ray tube. After weighting the X-ray image, a differential energy subtraction image representing the bone part is created by taking the difference between them.
- an X-ray image used for machine learning either a high pressure image, a low pressure image, or a dual energy subtraction image may be used, or an image obtained by connecting these images in the channel direction may be used. Good.
- parameter adjustment is performed on the dual energy subtraction image, thereby simulating the bone portion having a reduced bone density. May be obtained.
- FIG. 12 is a schematic diagram of the X-ray image 106 created by the X-ray image creation unit 81
- FIG. 13 is a schematic diagram of the DRR image 107 created by the DRR image creation unit 83.
- FIG. 6 described above is a schematic diagram of the X-ray image 101 near the subject's right foot
- FIG. 9 is a schematic diagram of the DRR image 103 near the subject's right foot
- FIG. 12 is a schematic diagram of an X-ray image 106 near the left foot of the subject
- FIG. 13 is a schematic diagram of a DRR image 107 near the left foot of the subject.
- the learning unit 85 performs an image of the right bone part on the left side.
- the machine learning is executed on the left and right bone images collectively by flipping one of the two bone images horizontally.
- the X-ray image 106 near the subject's left foot shown in FIG. 12 is reversed left and right, and used together with the X-ray image 101 near the subject's right foot shown in FIG. 6 for machine learning.
- the DRR image 107 near the subject's left foot shown in FIG. 13 is reversed left and right to be used together with the DRR image 103 near the subject's right foot shown in FIG. 9 for machine learning.
- machine learning is performed using both X-ray images and DRR images.
- machine learning may be performed using either one of the X-ray image and the DRR image.
- a bone part is targeted as an organ, but an organ such as an organ may be targeted.
- an organ such as an organ may be targeted.
- the concentration of the organ region is low during X-ray imaging. According to the present invention, even in such a case, it is possible to create a learning model corresponding to a subject whose organ concentration is low. For this reason, it becomes possible to improve the detection accuracy of an organ.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Medical Informatics (AREA)
- Physics & Mathematics (AREA)
- Public Health (AREA)
- General Health & Medical Sciences (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Radiology & Medical Imaging (AREA)
- Biomedical Technology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Pathology (AREA)
- Theoretical Computer Science (AREA)
- Animal Behavior & Ethology (AREA)
- Optics & Photonics (AREA)
- High Energy & Nuclear Physics (AREA)
- Biophysics (AREA)
- Heart & Thoracic Surgery (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- Veterinary Medicine (AREA)
- General Physics & Mathematics (AREA)
- Software Systems (AREA)
- Data Mining & Analysis (AREA)
- Epidemiology (AREA)
- Primary Health Care (AREA)
- Orthopedic Medicine & Surgery (AREA)
- Quality & Reliability (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Physiology (AREA)
- Artificial Intelligence (AREA)
- Pulmonology (AREA)
- Evolutionary Computation (AREA)
- Dentistry (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Databases & Information Systems (AREA)
- Image Analysis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
12 X線検出部
13 天板
14 支柱
15 X線管保持部材
16 回転機構
17 主支柱
18 ベースプレート
21 表示部
22 操作部
70 CT画像記憶部
80 制御部
81 X線画像作成部
82 X線画像記憶部
83 DRR画像作成部
84 DRR画像記憶部
85 学習部
86 骨部画像作成部
300 CT画像データ
Claims (14)
- 被検者の器官を含む領域の画像を解析することにより前記器官の領域を特定するためのセグメンテーションを行う画像解析方法であって、
前記セグメンテーションの手法として機械学習を用いるとともに、
前記被検者の器官を含む画像における前記器官の領域の濃度を変化させた修正画像を作成する修正画像作成工程と、
前記被検者の器官を含む画像と前記修正画像作成工程で作成した修正画像とを用いた学習処理により機械学習の学習モデルを作成する学習モデル作成工程と、
を含むことを特徴とする画像解析方法。 - 請求項1に記載の画像解析方法において、
前記被検者をX線撮影して得た前記被検者の器官を含む領域のX線画像に対して、前記学習モデル作成工程で作成した学習モデルを利用して変換を行うことにより、前記器官を表す画像を作成する画像解析方法。 - 請求項1に記載の画像解析方法において、
前記被検者の器官を含む領域の画像は、前記被検者のCT画像データから作成されたDRR画像であり、
前記修正画像作成工程においては、前記CT画像データのCT値が所定の値となる領域を前記器官の領域としてその濃度を変化させる画像解析方法。 - 請求項3に記載の画像解析方法において、
DRR画像の作成時に、前記幾何学的条件の投影座標および角度の少なくとも一方を含むパラメータを変化させ、あるいは、画像の回転、変形および拡大縮小の少なくとも1つを含む画像処理を施して、複数のDRR画像を作成する画像解析方法。 - 請求項3に記載の画像解析方法において、
作成後のDRR画像に対して、コントラスト変化、ノイズ付加およびエッジ強調の少なくとも1つを実行する画像解析方法。 - 請求項1に記載の画像解析方法において、
前記被検者の器官を含む領域の画像は、前記被検者をX線撮影することにより作成されたX線画像であり、
前記修正画像作成工程においては、前記X線画像と、デュアルエナジーサブトラクションを利用して得られた前記器官の画像とを利用して前記器官の領域の濃度を変化させる画像解析方法。 - 請求項2に記載の画像解析方法において、
前記被検者をX線撮影して得た前記被検者の器官を含む領域のX線画像と、前記学習モデル作成工程で作成した学習モデルを利用して変換を行うことにより得た前記器官を表す画像とを、前記学習部による学習モデルの学習に利用する画像解析方法。 - 請求項1に記載の画像解析方法において、
前記器官は前記被検者の体軸に対して左右対称の形状を有し、前学習モデル作成工程においては、右側の器官の画像と左側の器官の画像のいずれか一方を左右反転することにより、左右の器官の画像に対して一括して機械学習の学習モデルを作成する画像解析方法。 - 前記器官は前記被検者の骨部である、請求項1に記載の画像解析方法を利用して前記骨部の領域をセグメンテーションするセグメンテーション方法。
- 請求項9に記載のセグメンテーション方法によりセグメンテーションされた骨部の領域に対して骨密度を測定する骨密度測定方法。
- 被検者の器官を含む領域の画像を、機械学習を利用して解析することにより、前記器官の領域を特定するためのセグメンテーションを行うときに用いられる学習モデルを作成する学習モデル作成方法であって、
前記被検者の器官を含む画像と、前記被検者の器官を含む画像における前記器官の領域の濃度を変化させることにより作成された修正画像とを用い、機械学習の学習を実行して学習モデルを作成することを特徴とする学習モデル作成方法。 - 被検者の器官を含む領域のX線画像から前記器官の領域を抽出した画像を作成する画像作成装置であって、
前記器官を含む領域をX線撮影して得た複数のX線画像と、機械学習用の複数のX線画像用教師画像とを記憶するX線画像記憶部と、
前記骨部を含む領域のDRR画像を作成するDRR画像作成部と、
前記DRR画像作成部により作成された複数のDRR画像と、前記DRR画像作成部により作成されたDRR画像に基づいて作成された複数の機械学習用のDRR画像用教師画像とを記憶するDRR画像記憶部と、
前記X線画像記憶部に記憶された前記複数のX線画像と前記複数のX線画像用教師画像とを使用して機械学習を実行するとともに、前記DRR画像記憶部に記憶された前記複数のDRR画像と前記複数のDRR画像用教師画像とを使用して機械学習を実行することによって予め作成された前記器官を認識するための学習モデルを使用して、前記被検者の器官を含む領域のX線画像に対して変換を行うことにより、前記器官を表す画像を作成する画像作成部と、
を備えたことを特徴とする画像作成装置。 - 請求項11に記載の画像作成装置において、
前記DRR画像作成部は、前記複数のDRR画像のうちの一部のDRR画像を、前記骨部を含む領域のうちの器官領域の濃度を変化させたDRR画像として作成する画像作成装置。 - 請求項11に記載の画像作成装置において、
前記X線画像記憶部に記憶される複数のX線画像のうちの一部のX線画像は、デユアルエナジーサブトラクションを利用することにより、前記器官を含む領域のうちの器官領域の濃度を変化させたX線画像である画像作成装置。
Priority Applications (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201980035078.8A CN112165900B (zh) | 2018-04-24 | 2019-03-20 | 图像解析方法、分割方法、骨密度测量方法、学习模型生成方法和图像生成装置 |
| JP2020516112A JP7092190B2 (ja) | 2018-04-24 | 2019-03-20 | 画像解析方法、セグメンテーション方法、骨密度測定方法、学習モデル作成方法および画像作成装置 |
| KR1020207032563A KR102527440B1 (ko) | 2018-04-24 | 2019-03-20 | 화상 해석 방법, 세그먼테이션 방법, 골밀도 측정 방법, 학습 모델 작성 방법 및 화상 작성 장치 |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2018083340 | 2018-04-24 | ||
| JP2018-083340 | 2018-04-24 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2019208037A1 true WO2019208037A1 (ja) | 2019-10-31 |
Family
ID=68293538
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2019/011773 Ceased WO2019208037A1 (ja) | 2018-04-24 | 2019-03-20 | 画像解析方法、セグメンテーション方法、骨密度測定方法、学習モデル作成方法および画像作成装置 |
Country Status (4)
| Country | Link |
|---|---|
| JP (1) | JP7092190B2 (ja) |
| KR (1) | KR102527440B1 (ja) |
| CN (1) | CN112165900B (ja) |
| WO (1) | WO2019208037A1 (ja) |
Cited By (15)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113496494A (zh) * | 2021-06-17 | 2021-10-12 | 北京理工大学 | 基于drr模拟数据生成的二维骨骼分割方法及装置 |
| EP4056120A1 (en) * | 2021-03-12 | 2022-09-14 | FUJI-FILM Corporation | Estimation device, estimation method, and estimation program |
| US20220287663A1 (en) * | 2021-03-11 | 2022-09-15 | Fujifilm Corporation | Estimation device, estimation method, and estimation program |
| JP2022139212A (ja) * | 2021-03-11 | 2022-09-26 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| JP2022163614A (ja) * | 2021-04-14 | 2022-10-26 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| WO2022224558A1 (ja) * | 2021-04-22 | 2022-10-27 | 日本装置開発株式会社 | X線検査装置 |
| WO2022244495A1 (ja) * | 2021-05-17 | 2022-11-24 | キヤノン株式会社 | 放射線撮像装置および放射線撮像システム |
| JP2022176882A (ja) * | 2021-05-17 | 2022-11-30 | キヤノン株式会社 | 放射線撮像装置および放射線撮像システム |
| JP2023047911A (ja) * | 2021-09-27 | 2023-04-06 | 富士フイルム株式会社 | 画像処理装置、方法およびプログラム |
| JP2023065028A (ja) * | 2021-10-27 | 2023-05-12 | 堺化学工業株式会社 | 教師データ生成方法、画像解析モデル生成方法、画像解析方法、教師データ生成プログラム、画像解析プログラムおよび教師データ生成装置 |
| EP4233000A1 (en) * | 2022-01-13 | 2023-08-30 | Brainlab AG | Detection of image structures via dimensionality-reducing projections |
| WO2023224022A1 (ja) * | 2022-05-20 | 2023-11-23 | 国立大学法人大阪大学 | プログラム、情報処理方法、及び情報処理装置 |
| US11963810B2 (en) | 2021-03-12 | 2024-04-23 | Fujifilm Corporation | Estimation device, estimation method, and estimation program |
| JP2025041498A (ja) * | 2023-09-13 | 2025-03-26 | 京セラ株式会社 | 情報処理システム、情報処理方法、画像生成装置および画像生成プログラム |
| WO2025164645A1 (ja) * | 2024-01-31 | 2025-08-07 | 京セラ株式会社 | 生成方法、学習方法、推定方法、生成装置、推定装置、制御プログラムおよび記録媒体 |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| KR102705884B1 (ko) * | 2022-03-22 | 2024-09-11 | 사회복지법인 삼성생명공익재단 | 대퇴골 x-레이 영상을 이용한 학습 모델 기반의 종양 분류 방법 및 분석장치 |
Citations (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH05184562A (ja) * | 1992-01-17 | 1993-07-27 | Fuji Photo Film Co Ltd | 放射線画像の撮影方向認識方法 |
| JP2002236910A (ja) * | 2001-02-09 | 2002-08-23 | Hitachi Medical Corp | 3次元画像作成方法 |
| JP2007044485A (ja) * | 2005-08-05 | 2007-02-22 | Ge Medical Systems Global Technology Co Llc | 脳内出血部位セグメンテーション方法および装置 |
| JP2008167949A (ja) * | 2007-01-12 | 2008-07-24 | Fujifilm Corp | 放射線画像処理方法および装置ならびにプログラム |
| JP2014158628A (ja) * | 2013-02-20 | 2014-09-04 | Univ Of Tokushima | 画像処理装置、画像処理方法、制御プログラム、および記録媒体 |
| US20150094564A1 (en) * | 2012-05-03 | 2015-04-02 | University Of Pittsburgh - Of The Commonwealth System Of Higher Education | Intelligent algorithms for tracking three-dimensional skeletal movement from radiographic image sequences |
| JP2017185007A (ja) * | 2016-04-05 | 2017-10-12 | 株式会社島津製作所 | 放射線撮影装置、放射線画像の対象物検出プログラムおよび放射線画像における対象物検出方法 |
| US20170323444A1 (en) * | 2016-05-09 | 2017-11-09 | Siemens Healthcare Gmbh | Method and apparatus for atlas/model-based segmentation of magnetic resonance images with weakly supervised examination-dependent learning |
Family Cites Families (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2638875B2 (ja) | 1988-01-31 | 1997-08-06 | 株式会社島津製作所 | 骨塩定量分析装置 |
| JP2008011901A (ja) * | 2006-07-03 | 2008-01-24 | Fujifilm Corp | 画像種類判別装置および方法並びにプログラム |
| JP2010246883A (ja) * | 2009-03-27 | 2010-11-04 | Mitsubishi Electric Corp | 患者位置決めシステム |
| JP6430238B2 (ja) * | 2014-12-24 | 2018-11-28 | 好民 村山 | 放射線撮影装置 |
| JP6815586B2 (ja) * | 2015-06-02 | 2021-01-20 | 東芝エネルギーシステムズ株式会社 | 医用画像処理装置、および治療システム |
| KR101928984B1 (ko) * | 2016-09-12 | 2018-12-13 | 주식회사 뷰노 | 골밀도 추정 방법 및 장치 |
-
2019
- 2019-03-20 JP JP2020516112A patent/JP7092190B2/ja active Active
- 2019-03-20 CN CN201980035078.8A patent/CN112165900B/zh active Active
- 2019-03-20 WO PCT/JP2019/011773 patent/WO2019208037A1/ja not_active Ceased
- 2019-03-20 KR KR1020207032563A patent/KR102527440B1/ko active Active
Patent Citations (8)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH05184562A (ja) * | 1992-01-17 | 1993-07-27 | Fuji Photo Film Co Ltd | 放射線画像の撮影方向認識方法 |
| JP2002236910A (ja) * | 2001-02-09 | 2002-08-23 | Hitachi Medical Corp | 3次元画像作成方法 |
| JP2007044485A (ja) * | 2005-08-05 | 2007-02-22 | Ge Medical Systems Global Technology Co Llc | 脳内出血部位セグメンテーション方法および装置 |
| JP2008167949A (ja) * | 2007-01-12 | 2008-07-24 | Fujifilm Corp | 放射線画像処理方法および装置ならびにプログラム |
| US20150094564A1 (en) * | 2012-05-03 | 2015-04-02 | University Of Pittsburgh - Of The Commonwealth System Of Higher Education | Intelligent algorithms for tracking three-dimensional skeletal movement from radiographic image sequences |
| JP2014158628A (ja) * | 2013-02-20 | 2014-09-04 | Univ Of Tokushima | 画像処理装置、画像処理方法、制御プログラム、および記録媒体 |
| JP2017185007A (ja) * | 2016-04-05 | 2017-10-12 | 株式会社島津製作所 | 放射線撮影装置、放射線画像の対象物検出プログラムおよび放射線画像における対象物検出方法 |
| US20170323444A1 (en) * | 2016-05-09 | 2017-11-09 | Siemens Healthcare Gmbh | Method and apparatus for atlas/model-based segmentation of magnetic resonance images with weakly supervised examination-dependent learning |
Cited By (32)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US12138092B2 (en) | 2021-03-11 | 2024-11-12 | Fujifilm Corporation | Estimation device, estimation method, and estimation program |
| JP7746507B2 (ja) | 2021-03-11 | 2025-09-30 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| US20220287663A1 (en) * | 2021-03-11 | 2022-09-15 | Fujifilm Corporation | Estimation device, estimation method, and estimation program |
| JP7628031B2 (ja) | 2021-03-11 | 2025-02-07 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| JP2022139211A (ja) * | 2021-03-11 | 2022-09-26 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| JP2022139212A (ja) * | 2021-03-11 | 2022-09-26 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| JP2025014050A (ja) * | 2021-03-11 | 2025-01-28 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| US12161495B2 (en) * | 2021-03-11 | 2024-12-10 | Fujifilm Corporation | Estimation device, estimation method, and estimation program |
| JP7582884B2 (ja) | 2021-03-11 | 2024-11-13 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| US11963810B2 (en) | 2021-03-12 | 2024-04-23 | Fujifilm Corporation | Estimation device, estimation method, and estimation program |
| JP7758830B2 (ja) | 2021-03-12 | 2025-10-22 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| EP4056120A1 (en) * | 2021-03-12 | 2022-09-14 | FUJI-FILM Corporation | Estimation device, estimation method, and estimation program |
| JP2022140050A (ja) * | 2021-03-12 | 2022-09-26 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| JP2025010394A (ja) * | 2021-03-12 | 2025-01-20 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| US12133752B2 (en) | 2021-03-12 | 2024-11-05 | Fujifilm Corporation | Estimation device, estimation method, and estimation program |
| US12148156B2 (en) | 2021-04-14 | 2024-11-19 | Fujifilm Corporation | Estimation device, estimation method, and estimation program |
| JP7686430B2 (ja) | 2021-04-14 | 2025-06-02 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| JP2022163614A (ja) * | 2021-04-14 | 2022-10-26 | 富士フイルム株式会社 | 推定装置、方法およびプログラム |
| WO2022224558A1 (ja) * | 2021-04-22 | 2022-10-27 | 日本装置開発株式会社 | X線検査装置 |
| JP2022167132A (ja) * | 2021-04-22 | 2022-11-04 | 日本装置開発株式会社 | X線検査装置 |
| US20240069222A1 (en) * | 2021-05-17 | 2024-02-29 | Canon Kabushiki Kaisha | Radiation imaging apparatus and radiation imaging system |
| WO2022244495A1 (ja) * | 2021-05-17 | 2022-11-24 | キヤノン株式会社 | 放射線撮像装置および放射線撮像システム |
| JP2022176882A (ja) * | 2021-05-17 | 2022-11-30 | キヤノン株式会社 | 放射線撮像装置および放射線撮像システム |
| CN113496494A (zh) * | 2021-06-17 | 2021-10-12 | 北京理工大学 | 基于drr模拟数据生成的二维骨骼分割方法及装置 |
| JP7728680B2 (ja) | 2021-09-27 | 2025-08-25 | 富士フイルム株式会社 | 画像処理装置、方法およびプログラム |
| JP2023047911A (ja) * | 2021-09-27 | 2023-04-06 | 富士フイルム株式会社 | 画像処理装置、方法およびプログラム |
| JP2023065028A (ja) * | 2021-10-27 | 2023-05-12 | 堺化学工業株式会社 | 教師データ生成方法、画像解析モデル生成方法、画像解析方法、教師データ生成プログラム、画像解析プログラムおよび教師データ生成装置 |
| EP4233000A1 (en) * | 2022-01-13 | 2023-08-30 | Brainlab AG | Detection of image structures via dimensionality-reducing projections |
| WO2023224022A1 (ja) * | 2022-05-20 | 2023-11-23 | 国立大学法人大阪大学 | プログラム、情報処理方法、及び情報処理装置 |
| JP2025041498A (ja) * | 2023-09-13 | 2025-03-26 | 京セラ株式会社 | 情報処理システム、情報処理方法、画像生成装置および画像生成プログラム |
| JP7664438B2 (ja) | 2023-09-13 | 2025-04-17 | 京セラ株式会社 | 情報処理システム、情報処理方法、画像生成装置および画像生成プログラム |
| WO2025164645A1 (ja) * | 2024-01-31 | 2025-08-07 | 京セラ株式会社 | 生成方法、学習方法、推定方法、生成装置、推定装置、制御プログラムおよび記録媒体 |
Also Published As
| Publication number | Publication date |
|---|---|
| KR20200142057A (ko) | 2020-12-21 |
| JPWO2019208037A1 (ja) | 2021-04-01 |
| CN112165900A (zh) | 2021-01-01 |
| KR102527440B1 (ko) | 2023-05-02 |
| JP7092190B2 (ja) | 2022-06-28 |
| CN112165900B (zh) | 2024-12-27 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| JP7092190B2 (ja) | 画像解析方法、セグメンテーション方法、骨密度測定方法、学習モデル作成方法および画像作成装置 | |
| JP6881611B2 (ja) | 画像作成装置及び学習済モデルの生成方法 | |
| EP2056255B1 (en) | Method for reconstruction of a three-dimensional model of an osteo-articular structure | |
| CN112308788B (zh) | 图像处理装置、图像处理方法以及x射线ct装置 | |
| US8660329B2 (en) | Method for reconstruction of a three-dimensional model of a body structure | |
| US12217430B2 (en) | Systems and methods for processing x-ray images | |
| CN109419526A (zh) | 用于数字乳房断层合成中的运动评估和校正的方法和系统 | |
| EP4284253B1 (en) | Adaptive collimation for interventional x-ray | |
| KR20110115762A (ko) | 환자 맞춤형 3차원 인체 뼈 모델 재구성 방법 | |
| US11963812B2 (en) | Method and device for producing a panoramic tomographic image of an object to be recorded | |
| EP3725227B1 (en) | Method of calibrating x-ray projection geometry in x-ray cone beam computed tomography | |
| EP4354395A1 (en) | Artificial intelligence-based dual energy x-ray image motion correction training method and system | |
| JP4416823B2 (ja) | 画像処理装置、画像処理方法、及びコンピュータプログラム | |
| JP2020127600A (ja) | 医用画像処理装置、x線診断装置及び医用情報処理システム | |
| CN118628432A (zh) | 医用图像处理方法、x射线诊断装置及训练完成模型的生成方法 | |
| EP4368109A1 (en) | Method for training a scatter correction model for use in an x-ray imaging system | |
| EP4428808A1 (en) | Medical image processing method, x-ray diagnosis apparatus, and generation method of trained model | |
| CN119497875A (zh) | 锥形束伪影减少 |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19793450 Country of ref document: EP Kind code of ref document: A1 |
|
| ENP | Entry into the national phase |
Ref document number: 2020516112 Country of ref document: JP Kind code of ref document: A |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| ENP | Entry into the national phase |
Ref document number: 20207032563 Country of ref document: KR Kind code of ref document: A |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 19793450 Country of ref document: EP Kind code of ref document: A1 |
|
| WWG | Wipo information: grant in national office |
Ref document number: 201980035078.8 Country of ref document: CN |