WO2022027595A1 - Method for reconstructing low-dose image by using multiscale feature sensing deep network - Google Patents
Method for reconstructing low-dose image by using multiscale feature sensing deep network Download PDFInfo
- Publication number
- WO2022027595A1 WO2022027595A1 PCT/CN2020/107822 CN2020107822W WO2022027595A1 WO 2022027595 A1 WO2022027595 A1 WO 2022027595A1 CN 2020107822 W CN2020107822 W CN 2020107822W WO 2022027595 A1 WO2022027595 A1 WO 2022027595A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- feature
- network model
- image
- dose
- convolutional layer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
Definitions
- the present invention relates to the technical field of medical image processing, and more particularly, to a method for reconstructing a low-dose image using a multi-scale feature-aware deep network.
- Computed tomography (CT) technology is widely used in the detection of tissues in the human body.
- CT Computed tomography
- the oral CT instrument emits X-ray beams from enough angles during the scanning process to obtain fully sampled projection data, and then calculates and reconstructs a clear intraoral image.
- the public has paid extensive attention to the damage to the human body caused by X-ray radiation in CT examinations, and patients wish to receive as little radiation as possible during examinations.
- the exposure time of the X-ray tube can be shortened by reducing the sampling angle, thereby reducing the radiation dose.
- the current low-dose image reconstruction methods mainly have the following defects: when the complete projection data is sampled, the radiation dose to the patient is relatively large; when the complete projection data is sampled, the X-ray tube exposure time is long, and the equipment loss is large; the existing low-dose image
- the reconstruction effect still needs to be improved.
- Xieshipeng et al. published an article "Artifact Removal using Improved GoogLeNet for Sparse-view CT Reconstruction” in the journal Scientific Reports in 2018 (using improved GoogLeNet to remove artifacts for sparse CT reconstruction).
- the article adds residual learning on the basis of GoogleNet, so that the network can learn the characteristics of artifacts.
- the network in the article contains 3 convolutional layers and 8 inception modules, each inception module contains 3 convolution kernels of different sizes, with sizes of 1 ⁇ 1, 3 ⁇ 3, 5 ⁇ 5, to improve feature extraction accuracy.
- the network is firstly a two-layer convolutional layer used to initially extract the artifact features of the input image, and then eight inception modules are used to process the artifact features more finely.
- the last layer of the network convolutional layer fuses all feature maps to form
- the artifact distribution map is output, and the sparsely reconstructed CT image is used to subtract the artifact distribution map to obtain the final corrected image.
- this method to obtain reconstructed images still needs to be further improved to restore more complete and clearer image details.
- the purpose of the present invention is to overcome the above-mentioned defects of the prior art, and provide a method for reconstructing a low-dose image using a multi-scale feature-aware depth network, and further process the low-dose image reconstructed from sparsely sampled projection data to obtain high quality image, the obtained image is close to the effect of reconstruction using fully sampled projection data.
- a method for reconstructing a low-dose image using a multi-scale feature-aware deep network includes the following steps:
- the network model includes multiple cascaded multi-scale feature perception modules, and the input and output of the network model have connections, each The multi-scale feature perception module includes a backbone structure, a first-level branch and a second-level branch for feature extraction of input feature maps of different scales, wherein the second-level branch adjusts the processed feature map size to the same size as the first-level branch.
- the branch is consistent and passed to the first-level branch to guide the feature extraction of the first-level branch; the first-level branch adjusts the size of the processed feature map to be consistent with the backbone structure and transmits it to the backbone structure to guide the backbone structure feature extraction; input the low-dose image to be reconstructed into the trained network model to obtain a reconstructed image.
- the present invention has the advantage that the feature map is processed into different scales by using a multi-scale feature perception module, and then the feature information is learned on different scales, so that the extracted feature information is more comprehensive.
- the processed feature information of different branches is transmitted layer by layer from bottom to top, and the feature information extracted by the upper branch can be corrected, thereby improving the accuracy of feature learning and making the artifact correction in the reconstructed image more accurate.
- the human tissue structure information is more complete.
- FIG. 1 is a flowchart of a method for reconstructing a low-dose image using a multi-scale feature-aware deep network according to an embodiment of the present invention
- FIG. 2 is a structural diagram of a multi-scale feature perception module according to an embodiment of the present invention.
- FIG. 3 is an overall framework diagram of a deep network model according to an embodiment of the present invention.
- FIG. 4 is a schematic diagram of a residual block in a multi-scale feature perception module according to an embodiment of the present invention.
- FIG. 5 is a schematic diagram of a fully sampled normal dose oral CT image according to an embodiment of the present invention.
- FIG. 6 is a schematic diagram of a sparsely sampled low-dose oral CT image according to an embodiment of the present invention.
- FIG. 7 is an effect diagram of a reconstructed oral CT image according to an embodiment of the present invention.
- the present invention provides a technical solution for reconstructing sparsely sampled low-dose maps using a multi-scale feature-aware deep network, which can be applied to various types of image data reconstruction, such as CT images, MRI images, PET images, and 3D images.
- image data reconstruction such as CT images, MRI images, PET images, and 3D images.
- the present invention provides a deep network model including a multi-scale feature perception module (or multi-scale feature perception module) to correct artifacts in sparsely sampled low-dose oral CT images to improve image quality.
- the deep network model is an end-to-end network, and the main body consists of multiple multi-scale feature perception modules and a long skip connection connecting the input and output.
- the multi-scale feature perception module is divided into a backbone structure and two branches according to the level. Two branches are used to reduce the input feature map to different sizes, so as to extract and learn the features of artifacts on different sizes.
- the size of the feature map processed by the second-level branch of the lowest level will be adjusted to be consistent with the first-level branch of the upper layer, and then input into the first-level branch to correct and guide the extraction and learning of the features of the first-level branch.
- the module backbone structure learns on the input feature map size.
- the feature map processed by the first-level branch will also adjust the size to be consistent with the backbone feature map, and then pass it to the backbone to make the feature map information more refined by the backbone. Accurate extraction and learning.
- Multiple multi-scale feature perception modules can be connected to facilitate the deep network model to generate higher quality CT images.
- the method for reconstructing a low-dose image using a multi-scale feature-aware deep network includes the following steps:
- step S110 a multi-scale feature perception module is constructed to perform feature extraction on feature maps of different scales.
- the multi-scale feature sensing module is divided into a backbone and two branches according to the level, which are respectively called the backbone structure (or backbone for short), the first-level branch and the second-level branch from top to bottom.
- the input feature map size is reduced to 1/2 and 1/4, respectively, and the number of feature map channels is increased to 4 times and 16 times, respectively, and then the reduced size feature maps are input to In the two-level branch.
- a layer of convolutional layer is used to fuse the feature information on some channels to reduce the number of channels and improve computational efficiency, and then use a residual block to further learn features.
- the residual block contains 3 convolutional layers, and the activation function LeakyReLU is used for activation after each convolutional layer.
- the processing result of the third convolution layer in the residual block is fused with the input of the residual block, and then output through the activation function LeakyReLU, so as to realize the transfer of low-level feature information to high-level, and improve the stability of network training.
- the shuffling operation is used to aggregate the feature information on different channels, the number of channels is reduced to 1/4 and the size of the feature map is enlarged by 2 times, and then the feature information is passed to the upper branch. middle.
- a convolutional layer is also used to fuse the feature information on some channels to reduce the number of channels.
- the primary branch first cascades the feature information from the secondary branch and fuses the information through the convolution layer, and then inputs it into the residual block for processing.
- the feature information is used to guide and correct the feature extraction and learning of the first-level branch.
- the output of the residual block is also processed by the convolution layer, and then the feature information on different channels is aggregated through the shuffling operation, the number of channels is reduced to 1/4 and the size of the feature map is enlarged to 2 times the original size, and then the feature information is passed to the backbone. among.
- the backbone structure is different from other branches.
- the backbone can be divided into two parts according to the cascading position of the feature information of the lower branch.
- Each of the two parts has 3 layers of convolutional layers, and the convolutional layers of the former part are followed by activation functions.
- ReLU the latter part is followed by the activation function LeakyReLU.
- the feature information extracted and learned from the former part is fused with the feature information of the lower branch for correction, and then input to the latter part for subsequent processing.
- the processed feature information is fused with the input of the multi-scale feature perception module to obtain the final result, which is then output outside the module.
- a multi-size feature perception module is constructed for learning feature maps of different sizes, which can extract more accurate feature information.
- the residual blocks that use more convolution layers, first-level branches and second-level branches use the same Or a different structure, use other activation functions, or include more levels of branches, etc.
- the cascade identified in FIG. 2 is used to indicate the location where the feature information of the subordinate branch is received, and is also called a cascade layer for the convenience of understanding the overall architecture of the network.
- Step S120 constructing an end-to-end network model including multiple multi-scale feature perception modules.
- the network model includes two convolution layers (the size of the convolution kernel can be set to 3 ⁇ 3), two multi-scale feature perception modules and a connection network model input and long jump connections at the output.
- the first convolutional layer in the network model is used to perform preliminary feature extraction on the input image, generate a feature map with 64 channels, for example, and then input it into the multi-scale feature perception module.
- Two multi-scale feature perception modules are used for more accurate extraction and learning of image feature information.
- the last layer of convolutional layer fuses the information of different channels in the feature map output by the multi-scale feature perception module to obtain a single-channel image, and fuses the single-channel image with the input image to obtain the final output of the network model. Shadow corrected image.
- the performance of the network model can be adjusted, so that the network model can be applied to different tasks, thereby improving the generalization ability.
- Step S130 designing the loss function of the network model.
- the loss function of the network model is divided into reconstruction loss and perceptual loss.
- the reconstruction loss uses L1loss to calculate the difference between the generated image and the normal dose image, which can be expressed as:
- Net represents the network model of the present invention
- Net( xi ) represents the output result of the network model
- the perceptual loss refers to using the network model VGG19 to generate high-level features of the image and the normal dose image, and then calculating the L2loss between the two, which can be expressed as:
- VGG 19 (Net(x i )) and VGG 19 (y i ) represent the output results of the generated image and normal dose image input to the VGG19 network, respectively.
- the purpose of training the network model is to minimize the value of the above two loss functions as the optimization goal, and update the parameters of the network model.
- Step S140 train the network model with the optimized loss function as the target.
- Step S150 the low-dose image is input into the trained network model to obtain a reconstructed image.
- the mapping between the low-dose oral CT image and the normal-dose oral CT image can be realized, that is, the sparsely sampled low-dose image is input into the trained network model, and the reconstructed image can be obtained.
- Fig. 5 is a normal-dose oral CT image of complete sampling
- Fig. 6 is a sparsely sampled low-dose oral CT image
- Fig. 7 is the rendering of the reconstructed oral CT image. It can be seen that the reconstructed image obtained by the present invention has clearer details and a more complete structure.
- CT reconstruction image quality evaluation parameters are shown in Table 1 below.
- CT reconstruction image quality evaluation parameter table (PSNR, SSIM, NMSE)
- the multi-scale feature perception module of the present invention processes the feature map into different scales, and then learns the feature information on different scales, so that the extracted feature information is more comprehensive.
- the processed feature information of different branches is transmitted from bottom to top layer by layer. After the feature information extracted by the lower layer branch is fused with the feature information of the upper layer branch, the feature information extracted by the upper layer branch can be corrected, which improves the performance of the upper layer branch.
- the accuracy of feature learning makes the artifact correction in the images output by the network model more accurate, and the human tissue structure information is more complete.
- shuffling and inverse shuffling operations are applied in the multi-size feature perception module, and the number of feature map channels is correspondingly increased and decreased when reducing and increasing the size of the feature map, so as to better preserve pixel information and avoid information loss.
- the present invention may be a system, method and/or computer program product.
- the computer program product may include a computer-readable storage medium having computer-readable program instructions loaded thereon for causing a processor to implement various aspects of the present invention.
- a computer-readable storage medium may be a tangible device that can hold and store instructions for use by the instruction execution device.
- the computer-readable storage medium may be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
- Non-exhaustive list of computer readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM) or flash memory), static random access memory (SRAM), portable compact disk read only memory (CD-ROM), digital versatile disk (DVD), memory sticks, floppy disks, mechanically coded devices, such as printers with instructions stored thereon Hole cards or raised structures in grooves, and any suitable combination of the above.
- RAM random access memory
- ROM read only memory
- EPROM erasable programmable read only memory
- flash memory static random access memory
- SRAM static random access memory
- CD-ROM compact disk read only memory
- DVD digital versatile disk
- memory sticks floppy disks
- mechanically coded devices such as printers with instructions stored thereon Hole cards or raised structures in grooves, and any suitable combination of the above.
- Computer-readable storage media are not to be construed as transient signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through waveguides or other transmission media (eg, light pulses through fiber optic cables), or through electrical wires transmitted electrical signals.
- the computer readable program instructions described herein may be downloaded to various computing/processing devices from a computer readable storage medium, or to an external computer or external storage device over a network such as the Internet, a local area network, a wide area network, and/or a wireless network.
- the network may include copper transmission cables, fiber optic transmission, wireless transmission, routers, firewalls, switches, gateway computers, and/or edge servers.
- a network adapter card or network interface in each computing/processing device receives computer-readable program instructions from a network and forwards the computer-readable program instructions for storage in a computer-readable storage medium in each computing/processing device .
- the computer program instructions for carrying out the operations of the present invention may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state setting data, or instructions in one or more programming languages.
- Source or object code written in any combination, including object-oriented programming languages, such as Smalltalk, C++, etc., and conventional procedural programming languages, such as the "C" language or similar programming languages.
- the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server implement.
- the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (eg, using an Internet service provider through the Internet connect).
- LAN local area network
- WAN wide area network
- custom electronic circuits such as programmable logic circuits, field programmable gate arrays (FPGAs), or programmable logic arrays (PLAs)
- FPGAs field programmable gate arrays
- PDAs programmable logic arrays
- Computer readable program instructions are executed to implement various aspects of the present invention.
- These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer or other programmable data processing apparatus to produce a machine that causes the instructions when executed by the processor of the computer or other programmable data processing apparatus , resulting in means for implementing the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
- These computer readable program instructions can also be stored in a computer readable storage medium, these instructions cause a computer, programmable data processing apparatus and/or other equipment to operate in a specific manner, so that the computer readable medium on which the instructions are stored includes An article of manufacture comprising instructions for implementing various aspects of the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
- Computer readable program instructions can also be loaded onto a computer, other programmable data processing apparatus, or other equipment to cause a series of operational steps to be performed on the computer, other programmable data processing apparatus, or other equipment to produce a computer-implemented process , thereby causing instructions executing on a computer, other programmable data processing apparatus, or other device to implement the functions/acts specified in one or more blocks of the flowcharts and/or block diagrams.
- each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more functions for implementing the specified logical function(s) executable instructions.
- the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
- each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in dedicated hardware-based systems that perform the specified functions or actions , or can be implemented in a combination of dedicated hardware and computer instructions. It is well known to those skilled in the art that implementation in hardware, implementation in software, and implementation in a combination of software and hardware are all equivalent.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
本发明涉及医学图像处理技术领域,更具体地,涉及一种利用多尺度特征感知深度网络重建低剂量图像的方法。The present invention relates to the technical field of medical image processing, and more particularly, to a method for reconstructing a low-dose image using a multi-scale feature-aware deep network.
计算机断层成像(CT)技术在人体内组织检测中应用广泛。以口腔CT图像为例,口腔CT仪器在扫描过程中从足够多的角度发射X射线束来获取完整采样的投影数据,进而计算重建得到清晰的口腔内部图像。近年来,大众广泛关注CT检查中X射线辐射对人体的损伤,病人在接受检查时希望所承受的辐射剂量尽可能的小。在检查过程中,可以通过减少采样角度来缩短X射线管的曝光时长,从而降低辐射剂量。但稀疏采样的投影数据重建得到的图像中会出现明显的伪影以及噪声,造成图像质量明显下降,对临床诊断造成影响。因此,研究从低剂量的口腔CT图像生成接近正常剂量效果的口腔CT图像,能够在有效减少病人所受的辐射剂量的情况下,获得满足临床诊断需求的高质量CT图像,有利于扩大CT技术的应用范围。Computed tomography (CT) technology is widely used in the detection of tissues in the human body. Taking an oral CT image as an example, the oral CT instrument emits X-ray beams from enough angles during the scanning process to obtain fully sampled projection data, and then calculates and reconstructs a clear intraoral image. In recent years, the public has paid extensive attention to the damage to the human body caused by X-ray radiation in CT examinations, and patients wish to receive as little radiation as possible during examinations. During the inspection process, the exposure time of the X-ray tube can be shortened by reducing the sampling angle, thereby reducing the radiation dose. However, there will be obvious artifacts and noise in the image reconstructed from the sparsely sampled projection data, which will cause the image quality to decrease significantly and affect the clinical diagnosis. Therefore, studying the generation of oral CT images with a near-normal dose effect from low-dose oral CT images can effectively reduce the radiation dose received by patients and obtain high-quality CT images that meet the needs of clinical diagnosis, which is conducive to expanding CT technology. scope of application.
然而,目前低剂量图像重建方法主要存在以下缺陷:完整投影数据采样时病人承受的辐射剂量较大;完整投影数据采样时X射线管曝光时间较长,设备损耗较大;现有的低剂量图像重建效果还有待提高等。例如,Xieshipeng等于2018年在Scientific Reports期刊上发表文章“Artifact Removal using Improved GoogLeNet for Sparse-view CT Reconstruction”(使用改进的GoogLeNet去除伪影以进行稀疏CT重建)。文章在GoogleNet的基础上加入残差学习,使网络学习伪影的特征。文章中的网络包含3层卷积层和8个inception模块,每个inception模块包含3个不同尺寸的卷积核,尺寸分别为1×1,3×3,5×5,用以提升特征提取的准确性。网络首先是两层卷积层用于对输入图像进行伪影特征的初步提取,然后是8 个inception模块对伪影特征进行更为精细地处理,网络最后一层卷积层融合全部特征图形成伪影分布图并输出,使用稀疏重建的CT图像减去伪影分布图得到最终的校正图像。但利用该方法获得重建图像的效果仍有待于进一步改善,以恢复出更完整、更清晰的图像细节。However, the current low-dose image reconstruction methods mainly have the following defects: when the complete projection data is sampled, the radiation dose to the patient is relatively large; when the complete projection data is sampled, the X-ray tube exposure time is long, and the equipment loss is large; the existing low-dose image The reconstruction effect still needs to be improved. For example, Xieshipeng et al. published an article "Artifact Removal using Improved GoogLeNet for Sparse-view CT Reconstruction" in the journal Scientific Reports in 2018 (using improved GoogLeNet to remove artifacts for sparse CT reconstruction). The article adds residual learning on the basis of GoogleNet, so that the network can learn the characteristics of artifacts. The network in the article contains 3 convolutional layers and 8 inception modules, each inception module contains 3 convolution kernels of different sizes, with sizes of 1×1, 3×3, 5×5, to improve feature extraction accuracy. The network is firstly a two-layer convolutional layer used to initially extract the artifact features of the input image, and then eight inception modules are used to process the artifact features more finely. The last layer of the network convolutional layer fuses all feature maps to form The artifact distribution map is output, and the sparsely reconstructed CT image is used to subtract the artifact distribution map to obtain the final corrected image. However, the effect of using this method to obtain reconstructed images still needs to be further improved to restore more complete and clearer image details.
发明内容SUMMARY OF THE INVENTION
本发明的目的是克服上述现有技术的缺陷,提供一种利用多尺度特征感知深度网络重建低剂量图像的方法,对由稀疏采样的投影数据重建得到的低剂量图像进一步进行处理,以获得高质量的图像,所获得的图像接近使用完整采样投影数据进行重建的效果。The purpose of the present invention is to overcome the above-mentioned defects of the prior art, and provide a method for reconstructing a low-dose image using a multi-scale feature-aware depth network, and further process the low-dose image reconstructed from sparsely sampled projection data to obtain high quality image, the obtained image is close to the effect of reconstruction using fully sampled projection data.
根据本发明的一方面,提供一种利用多尺度特征感知深度网络重建低剂量图像的方法。该方法包括以下步骤:According to an aspect of the present invention, a method for reconstructing a low-dose image using a multi-scale feature-aware deep network is provided. The method includes the following steps:
构建以低剂量图像作为输入、以正常剂量图像作为输出的端到端网络模型,该网络模型包括多个级联的多尺寸特征感知模块且该网络模型的输入端和输出端具有连接,每个多尺寸特征感知模块包含用于对不同尺度的输入特征图进行特征提取的主干结构、一级支路和二级支路,其中,二级支路将处理后的特征图尺寸调整到与一级支路一致并传递至一级支路中,以指导一级支路的特征提取;一级支路将处理后的特征图尺寸调整到与主干结构一致并传递至主干结构中,以指导主干结构的特征提取;将待重建的低剂量图像输入至经训练的所述网络模型,获得重建图像。Build an end-to-end network model with low-dose images as input and normal-dose images as output, the network model includes multiple cascaded multi-scale feature perception modules, and the input and output of the network model have connections, each The multi-scale feature perception module includes a backbone structure, a first-level branch and a second-level branch for feature extraction of input feature maps of different scales, wherein the second-level branch adjusts the processed feature map size to the same size as the first-level branch. The branch is consistent and passed to the first-level branch to guide the feature extraction of the first-level branch; the first-level branch adjusts the size of the processed feature map to be consistent with the backbone structure and transmits it to the backbone structure to guide the backbone structure feature extraction; input the low-dose image to be reconstructed into the trained network model to obtain a reconstructed image.
与现有技术相比,本发明的优点在于,利用多尺寸特征感知模块将特征图处理成不同尺度,然后在不同尺度上对特征信息进行学习,使得提取的特征信息更为全面。并且,不同支路的处理后的特征信息逐层由下至上进行传递,可以对上层支路提取的特征信息进行校正,从而提升了特征学习的准确性,使得重建图像中伪影校正更为精确,人体组织结构信息更为完整。Compared with the prior art, the present invention has the advantage that the feature map is processed into different scales by using a multi-scale feature perception module, and then the feature information is learned on different scales, so that the extracted feature information is more comprehensive. In addition, the processed feature information of different branches is transmitted layer by layer from bottom to top, and the feature information extracted by the upper branch can be corrected, thereby improving the accuracy of feature learning and making the artifact correction in the reconstructed image more accurate. , the human tissue structure information is more complete.
通过以下参照附图对本发明的示例性实施例的详细描述,本发明的其它特征及其优点将会变得清楚。Other features and advantages of the present invention will become apparent from the following detailed description of exemplary embodiments of the present invention with reference to the accompanying drawings.
被结合在说明书中并构成说明书的一部分的附图示出了本发明的实施例,并且连同其说明一起用于解释本发明的原理。The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the principles of the invention.
图1是根据本发明一个实施例的利用多尺度特征感知深度网络重建低剂量图像的方法的流程图;1 is a flowchart of a method for reconstructing a low-dose image using a multi-scale feature-aware deep network according to an embodiment of the present invention;
图2是根据本发明一个实施例的多尺度特征感知模块的结构图;2 is a structural diagram of a multi-scale feature perception module according to an embodiment of the present invention;
图3是根据本发明一个实施例的深度网络模型整体框架图;3 is an overall framework diagram of a deep network model according to an embodiment of the present invention;
图4是根据本发明一个实施例的多尺寸特征感知模块中残差块示意图;4 is a schematic diagram of a residual block in a multi-scale feature perception module according to an embodiment of the present invention;
图5是根据本发明一个实施例的完整采样的正常剂量口腔CT图像示意图;FIG. 5 is a schematic diagram of a fully sampled normal dose oral CT image according to an embodiment of the present invention;
图6是根据本发明一个实施例的稀疏采样的低剂量口腔CT图像示意图;6 is a schematic diagram of a sparsely sampled low-dose oral CT image according to an embodiment of the present invention;
图7是根据本发明一个实施例的重建后的口腔CT图像的效果图。FIG. 7 is an effect diagram of a reconstructed oral CT image according to an embodiment of the present invention.
现在将参照附图来详细描述本发明的各种示例性实施例。应注意到:除非另外具体说明,否则在这些实施例中阐述的部件和步骤的相对布置、数字表达式和数值不限制本发明的范围。Various exemplary embodiments of the present invention will now be described in detail with reference to the accompanying drawings. It should be noted that the relative arrangement of components and steps, the numerical expressions and numerical values set forth in these embodiments do not limit the scope of the invention unless specifically stated otherwise.
以下对至少一个示例性实施例的描述实际上仅仅是说明性的,决不作为对本发明及其应用或使用的任何限制。The following description of at least one exemplary embodiment is merely illustrative in nature and is in no way intended to limit the invention, its application, or uses.
对于相关领域普通技术人员已知的技术、方法和设备可能不作详细讨论,但在适当情况下,所述技术、方法和设备应当被视为说明书的一部分。Techniques, methods, and apparatus known to those of ordinary skill in the relevant art may not be discussed in detail, but where appropriate, such techniques, methods, and apparatus should be considered part of the specification.
在这里示出和讨论的所有例子中,任何具体值应被解释为仅仅是示例性的,而不是作为限制。因此,示例性实施例的其它例子可以具有不同的值。In all examples shown and discussed herein, any specific values should be construed as illustrative only and not limiting. Accordingly, other instances of the exemplary embodiment may have different values.
应注意到:相似的标号和字母在下面的附图中表示类似项,因此,一旦某一项在一个附图中被定义,则在随后的附图中不需要对其进行进一步讨论。It should be noted that like numerals and letters refer to like items in the following figures, so once an item is defined in one figure, it does not require further discussion in subsequent figures.
本发明提供利用多尺度特征感知深度网络来对稀疏采样的低剂量图 进行重建的技术方案,可应用于多种类型的图像数据重建,例如CT图像、MRI图像、PET图像和3D图像等。为清楚理解本发明的思想,在下文的描述中,将以口腔CT图像为例进行说明。The present invention provides a technical solution for reconstructing sparsely sampled low-dose maps using a multi-scale feature-aware deep network, which can be applied to various types of image data reconstruction, such as CT images, MRI images, PET images, and 3D images. In order to clearly understand the idea of the present invention, in the following description, an oral CT image will be used as an example for illustration.
简言之,本发明提供利用包含多尺寸特征感知模块(或称多尺度特征感知模块)的深度网络模型来对稀疏采样的低剂量口腔CT图像中的伪影进行校正,以提升图像质量。该深度网络模型是一个端到端网络,主体由多个多尺度特征感知模块和一条连接输入端和输出端的长跳跃连接构成。例如,多尺寸特征感知模块中按层级分为一条主干结构和两条支路。使用两条支路分别将输入的特征图缩小到不同的尺寸,从而在不同尺寸上对伪影的特征进行提取、学习。其中,最低层次的二级支路处理后的特征图尺寸会调整到与上层的一级支路一致,然后输入到一级支路中,用以校正和指导一级支路特征的提取和学习。模块主干结构在输入的特征图尺寸上进行学习,同样,一级支路处理后的特征图也会调整尺寸大小与主干特征图一致,然后传递到主干中,使主干对特征图信息进行更为准确的提取和学习。多个多尺寸特征感知模块之间可以连接使用来促使深度网络模型生成更高质量的CT图像。In short, the present invention provides a deep network model including a multi-scale feature perception module (or multi-scale feature perception module) to correct artifacts in sparsely sampled low-dose oral CT images to improve image quality. The deep network model is an end-to-end network, and the main body consists of multiple multi-scale feature perception modules and a long skip connection connecting the input and output. For example, the multi-scale feature perception module is divided into a backbone structure and two branches according to the level. Two branches are used to reduce the input feature map to different sizes, so as to extract and learn the features of artifacts on different sizes. Among them, the size of the feature map processed by the second-level branch of the lowest level will be adjusted to be consistent with the first-level branch of the upper layer, and then input into the first-level branch to correct and guide the extraction and learning of the features of the first-level branch. . The module backbone structure learns on the input feature map size. Similarly, the feature map processed by the first-level branch will also adjust the size to be consistent with the backbone feature map, and then pass it to the backbone to make the feature map information more refined by the backbone. Accurate extraction and learning. Multiple multi-scale feature perception modules can be connected to facilitate the deep network model to generate higher quality CT images.
具体地,参见图1所示,本发明实施例提供的利用多尺度特征感知深度网络重建低剂量图像的方法包括以下步骤:Specifically, as shown in FIG. 1 , the method for reconstructing a low-dose image using a multi-scale feature-aware deep network provided by an embodiment of the present invention includes the following steps:
步骤S110,构建多尺度特征感知模块,以对不同尺度的特征图进行特征提取。In step S110, a multi-scale feature perception module is constructed to perform feature extraction on feature maps of different scales.
具体地,参见图2所示,多尺寸特征感知模块按层级分为一条主干和两条支路,从上至下分别称为主干结构(或简称主干)、一级支路和二级支路。通过一次和两次逆shuffling操作分别将输入特征图尺寸缩小到1/2和1/4,特征图通道数分别提升到原来的4倍和16倍,然后将尺寸缩小后的特征图分别输入到两级支路中。Specifically, as shown in FIG. 2 , the multi-scale feature sensing module is divided into a backbone and two branches according to the level, which are respectively called the backbone structure (or backbone for short), the first-level branch and the second-level branch from top to bottom. . Through one and two inverse shuffling operations, the input feature map size is reduced to 1/2 and 1/4, respectively, and the number of feature map channels is increased to 4 times and 16 times, respectively, and then the reduced size feature maps are input to In the two-level branch.
在最下级支路中(即二级支路),先用一层卷积层将部分通道上的特征信息融合,以降低通道数,提升计算效率,然后使用一个残差块进一步学习特征。例如,参见图3所示,残差块中包含3层卷积层,每层卷积层后都利用激活函数LeakyReLU进行激活操作。残差块中第3层卷积层的处 理结果与残差块的输入进行融合,再通过激活函数LeakyReLU进行输出,从而实现低层次特征信息到高层次的传递,提升了网络训练的稳定性。残差块输出结果经过卷积层处理后,然后使用shuffling操作汇聚不同通道上的特征信息,将通道数降到1/4并将特征图尺寸扩大2倍,然后将特征信息传递到上级支路中。In the lowest-level branch (ie, the second-level branch), a layer of convolutional layer is used to fuse the feature information on some channels to reduce the number of channels and improve computational efficiency, and then use a residual block to further learn features. For example, as shown in Figure 3, the residual block contains 3 convolutional layers, and the activation function LeakyReLU is used for activation after each convolutional layer. The processing result of the third convolution layer in the residual block is fused with the input of the residual block, and then output through the activation function LeakyReLU, so as to realize the transfer of low-level feature information to high-level, and improve the stability of network training. After the output of the residual block is processed by the convolution layer, the shuffling operation is used to aggregate the feature information on different channels, the number of channels is reduced to 1/4 and the size of the feature map is enlarged by 2 times, and then the feature information is passed to the upper branch. middle.
在一级支路中,同样先用一层卷积层将部分通道上的特征信息融合,降低通道数。与二级支路不同的是,该一级支路先级联了来自二级支路的特征信息并经过卷积层融合信息后,再输入到残差块中进行处理,二级支路的特征信息用以对该一级支路的特征提取、学习进行指导和校正。残差块输出结果同样使用卷积层处理,然后通过shuffling操作汇聚不同通道上的特征信息,将通道数降到1/4并扩大特征图尺寸到原先的2倍,然后将特征信息传递到主干之中。In the first-level branch, a convolutional layer is also used to fuse the feature information on some channels to reduce the number of channels. Different from the secondary branch, the primary branch first cascades the feature information from the secondary branch and fuses the information through the convolution layer, and then inputs it into the residual block for processing. The feature information is used to guide and correct the feature extraction and learning of the first-level branch. The output of the residual block is also processed by the convolution layer, and then the feature information on different channels is aggregated through the shuffling operation, the number of channels is reduced to 1/4 and the size of the feature map is enlarged to 2 times the original size, and then the feature information is passed to the backbone. among.
主干结构与其它支路不同,根据下层支路特征信息的级联位置可以将主干分为两个部分,两个部分各自都有着3层卷积层,前一部分的卷积层后都跟着激活函数ReLU,后一部分则跟着激活函数LeakyReLU。前一部分提取、学习的特征信息与下层支路的特征信息融合后进行校正,然后再输入到后一部分进行后续处理。处理完后的特征信息与多尺寸特征感知模块的输入融合,得到最终结果,再输出到该模块之外。The backbone structure is different from other branches. The backbone can be divided into two parts according to the cascading position of the feature information of the lower branch. Each of the two parts has 3 layers of convolutional layers, and the convolutional layers of the former part are followed by activation functions. ReLU, the latter part is followed by the activation function LeakyReLU. The feature information extracted and learned from the former part is fused with the feature information of the lower branch for correction, and then input to the latter part for subsequent processing. The processed feature information is fused with the input of the multi-scale feature perception module to obtain the final result, which is then output outside the module.
在该步骤中,构建多尺寸特征感知模块中,用于对不同尺寸的特征图进行学习,能够提取更为准确的特征信息。In this step, a multi-size feature perception module is constructed for learning feature maps of different sizes, which can extract more accurate feature information.
需要说明的是,本领域技术人员也可对上述的多尺寸特征感知模块进行适当变型或改变,例如,采用更多的卷积层、一级支路和二级支路的残差块采用相同或不同的结构、采用其他的激活函数、或者包含更多级的支路等。此外,图2中标识的级联用于指示接收下级支路特征信息的位置,为便于理解网络的整体架构,也称为级联层。It should be noted that those skilled in the art can also make appropriate modifications or changes to the above-mentioned multi-scale feature perception module, for example, the residual blocks that use more convolution layers, first-level branches and second-level branches use the same Or a different structure, use other activation functions, or include more levels of branches, etc. In addition, the cascade identified in FIG. 2 is used to indicate the location where the feature information of the subordinate branch is received, and is also called a cascade layer for the convenience of understanding the overall architecture of the network.
步骤S120,构建包含多个多尺寸特征感知模块的端到端网络模型。Step S120, constructing an end-to-end network model including multiple multi-scale feature perception modules.
例如,参见图4所示的网络模型的整体框架,该网络模型包括两层卷积层(卷积核尺寸可设置为3×3)、两个多尺寸特征感知模块和一个连接网络模型输入端和输出端的长跳跃连接构成。For example, referring to the overall framework of the network model shown in Figure 4, the network model includes two convolution layers (the size of the convolution kernel can be set to 3 × 3), two multi-scale feature perception modules and a connection network model input and long jump connections at the output.
由于低剂量口腔CT图像中的伪影、噪声可被视为加性干扰因素,因此加入长跳跃连接使得网络模型针对于残差图像(即伪影、噪声)的特征进行学习。网络模型中第一层卷积层用于对输入图像进行初步的特征提取,生成例如有64个通道的特征图,然后输入到多尺寸特征感知模块之中。两个多尺寸特征感知模块用于对图像特征信息进行更为精确的提取和学习。最后一层卷积层将多尺寸特征感知模块输出的特征图中不同通道的信息融合到一起,得到一张单通道图像,将该单通道图像与输入图像融合得到网络模型最终输出的干净的伪影校正图像。Since artifacts and noise in low-dose oral CT images can be regarded as additive interference factors, adding long skip connections enables the network model to learn the features of residual images (ie, artifacts, noise). The first convolutional layer in the network model is used to perform preliminary feature extraction on the input image, generate a feature map with 64 channels, for example, and then input it into the multi-scale feature perception module. Two multi-scale feature perception modules are used for more accurate extraction and learning of image feature information. The last layer of convolutional layer fuses the information of different channels in the feature map output by the multi-scale feature perception module to obtain a single-channel image, and fuses the single-channel image with the input image to obtain the final output of the network model. Shadow corrected image.
需要说明的是,通过将多个多尺寸特征感知模块(不限于两个)进行连接,可以调整网络模型的性能,使网络模型可以应用到不同任务之中,从而提升泛化能力。It should be noted that by connecting multiple multi-scale feature perception modules (not limited to two), the performance of the network model can be adjusted, so that the network model can be applied to different tasks, thereby improving the generalization ability.
步骤S130,设计网络模型的损失函数。Step S130, designing the loss function of the network model.
假设低剂量口腔CT图像数据集为D1={x 1,x 2,…,x n},正常剂量CT图像数据集为D2={y 1,y 2,…,y n},其中,n是图像样本的总数。优选地,网络模型的损失函数分为重建损失和感知损失。 Assume that the low-dose oral CT image dataset is D1={x 1 ,x 2 ,...,x n }, and the normal-dose CT image dataset is D2={y 1 ,y 2 ,...,y n }, where n is The total number of image samples. Preferably, the loss function of the network model is divided into reconstruction loss and perceptual loss.
在一个实施例中,重建损失采用L1loss计算生成图像和正常剂量图像之间的差距,可表示为:In one embodiment, the reconstruction loss uses L1loss to calculate the difference between the generated image and the normal dose image, which can be expressed as:
其中Net代表本发明的网络模型,Net(x i)代表网络模型输出结果。 Wherein Net represents the network model of the present invention, and Net( xi ) represents the output result of the network model.
在一个实施例中,感知损失指的是使用网络模型VGG19来生成图像和正常剂量图像的高层特征,然后计算两者之间的L2loss,可表示为:In one embodiment, the perceptual loss refers to using the network model VGG19 to generate high-level features of the image and the normal dose image, and then calculating the L2loss between the two, which can be expressed as:
其中VGG 19(Net(x i))和VGG 19(y i)分别代表生成图像和正常剂量图像输入到VGG19网络后的输出结果。 where VGG 19 (Net(x i )) and VGG 19 (y i ) represent the output results of the generated image and normal dose image input to the VGG19 network, respectively.
训练网络模型的目的是以尽可能地减小上述两种损失函数的值为优化目标,对网络模型的参数进行更新。The purpose of training the network model is to minimize the value of the above two loss functions as the optimization goal, and update the parameters of the network model.
需说明的是,本领域技术人员也可对上述损失函数进行适当的变型或修改。例如,不必须包括感知损失或者采用VGG16网络替代VGG19网络。It should be noted that those skilled in the art can also make appropriate variations or modifications to the above loss function. For example, it is not necessary to include perceptual loss or replace VGG19 network with VGG16 network.
步骤S140,以优化设定的损失函数为目标训练网络模型。Step S140, train the network model with the optimized loss function as the target.
使用配对的稀疏采样的低剂量口腔CT图像和完整采样的正常剂量口腔CT图像构建训练数据集和测试集,使用Adam算法在训练过程中迭代优化网络模型的各层参数(如权重、偏置等)。训练后,可进一步使用测试集测试网络模型的性能。Use paired sparsely sampled low-dose oral CT images and fully sampled normal-dose oral CT images to construct training datasets and test sets, and use the Adam algorithm to iteratively optimize the parameters of each layer of the network model (such as weights, biases, etc.) during the training process. ). After training, the performance of the network model can be further tested using the test set.
步骤S150,将低剂量图像输入至经训练的网络模型,获得重建后的图像。Step S150, the low-dose image is input into the trained network model to obtain a reconstructed image.
利用训练好的网络模型,可实现低剂量口腔CT图像到正常剂量口腔CT图像之间的映射,即将稀疏采样的低剂量图像输入至经训练的网络模型,即可获得重建后的图像。Using the trained network model, the mapping between the low-dose oral CT image and the normal-dose oral CT image can be realized, that is, the sparsely sampled low-dose image is input into the trained network model, and the reconstructed image can be obtained.
为进一步验证本发明的效果,进行了实验,实验结果参见图5至图7所示,其中图5是完整采样的正常剂量口腔CT图像,图6是稀疏采样的低剂量口腔CT图像,图7是重建后的口腔CT图像的效果图。可以看出,本发明获得的重建图像细节更清楚、结构更完整。此外,CT重建图像质量评估参数见下表1。In order to further verify the effect of the present invention, experiments were carried out, and the experimental results are shown in Fig. 5 to Fig. 7, wherein Fig. 5 is a normal-dose oral CT image of complete sampling, Fig. 6 is a sparsely sampled low-dose oral CT image, Fig. 7 is the rendering of the reconstructed oral CT image. It can be seen that the reconstructed image obtained by the present invention has clearer details and a more complete structure. In addition, the CT reconstruction image quality evaluation parameters are shown in Table 1 below.
表1.CT重建图像质量评估参数表(PSNR,SSIM,NMSE)Table 1. CT reconstruction image quality evaluation parameter table (PSNR, SSIM, NMSE)
由表1可知,PSNR(峰值信噪比)和SSIM(结构相似性)两项指标上,本发明的处理结果有着明显的提升,可见本发明能有效提高图像质量,且在NMSE(归一化均方误差)指标上分值较低,表明本发明的处理减小了与正常剂量CT图像之间的误差,结果更为准确,接近于正常剂量CT图像。It can be seen from Table 1 that the processing results of the present invention are significantly improved in terms of PSNR (peak signal-to-noise ratio) and SSIM (structural similarity). The lower score on the mean square error) index indicates that the processing of the present invention reduces the error with the normal dose CT image, and the result is more accurate, which is close to the normal dose CT image.
综上所述,本发明的多尺寸特征感知模块将特征图处理成不同尺度,然后在不同尺度上对特征信息进行学习、使得提取的特征信息更为全面。并且,不同支路的处理后的特征信息逐层由下往上进行传递,下层支路提取的特征信息与上层支路特征信息融合后,可以对上层支路提取的特征信息进行校正,提升了特征学习的准确性,从而使得网络模型输出的图像中伪影校正更为精确,人体组织结构信息更为完整。此外,多尺寸特征感知 模块中应用shuffling和逆shuffling操作,在减小和增大特征图尺寸时相应地增大和减小特征图通道数,从而更好地保留像素信息,避免信息丢失。To sum up, the multi-scale feature perception module of the present invention processes the feature map into different scales, and then learns the feature information on different scales, so that the extracted feature information is more comprehensive. In addition, the processed feature information of different branches is transmitted from bottom to top layer by layer. After the feature information extracted by the lower layer branch is fused with the feature information of the upper layer branch, the feature information extracted by the upper layer branch can be corrected, which improves the performance of the upper layer branch. The accuracy of feature learning makes the artifact correction in the images output by the network model more accurate, and the human tissue structure information is more complete. In addition, shuffling and inverse shuffling operations are applied in the multi-size feature perception module, and the number of feature map channels is correspondingly increased and decreased when reducing and increasing the size of the feature map, so as to better preserve pixel information and avoid information loss.
本发明可以是系统、方法和/或计算机程序产品。计算机程序产品可以包括计算机可读存储介质,其上载有用于使处理器实现本发明的各个方面的计算机可读程序指令。The present invention may be a system, method and/or computer program product. The computer program product may include a computer-readable storage medium having computer-readable program instructions loaded thereon for causing a processor to implement various aspects of the present invention.
计算机可读存储介质可以是可以保持和存储由指令执行设备使用的指令的有形设备。计算机可读存储介质例如可以是――但不限于――电存储设备、磁存储设备、光存储设备、电磁存储设备、半导体存储设备或者上述的任意合适的组合。计算机可读存储介质的更具体的例子(非穷举的列表)包括:便携式计算机盘、硬盘、随机存取存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、静态随机存取存储器(SRAM)、便携式压缩盘只读存储器(CD-ROM)、数字多功能盘(DVD)、记忆棒、软盘、机械编码设备、例如其上存储有指令的打孔卡或凹槽内凸起结构、以及上述的任意合适的组合。这里所使用的计算机可读存储介质不被解释为瞬时信号本身,诸如无线电波或者其他自由传播的电磁波、通过波导或其他传输媒介传播的电磁波(例如,通过光纤电缆的光脉冲)、或者通过电线传输的电信号。A computer-readable storage medium may be a tangible device that can hold and store instructions for use by the instruction execution device. The computer-readable storage medium may be, for example, but not limited to, an electrical storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. More specific examples (non-exhaustive list) of computer readable storage media include: portable computer disks, hard disks, random access memory (RAM), read only memory (ROM), erasable programmable read only memory (EPROM) or flash memory), static random access memory (SRAM), portable compact disk read only memory (CD-ROM), digital versatile disk (DVD), memory sticks, floppy disks, mechanically coded devices, such as printers with instructions stored thereon Hole cards or raised structures in grooves, and any suitable combination of the above. Computer-readable storage media, as used herein, are not to be construed as transient signals per se, such as radio waves or other freely propagating electromagnetic waves, electromagnetic waves propagating through waveguides or other transmission media (eg, light pulses through fiber optic cables), or through electrical wires transmitted electrical signals.
这里所描述的计算机可读程序指令可以从计算机可读存储介质下载到各个计算/处理设备,或者通过网络、例如因特网、局域网、广域网和/或无线网下载到外部计算机或外部存储设备。网络可以包括铜传输电缆、光纤传输、无线传输、路由器、防火墙、交换机、网关计算机和/或边缘服务器。每个计算/处理设备中的网络适配卡或者网络接口从网络接收计算机可读程序指令,并转发该计算机可读程序指令,以供存储在各个计算/处理设备中的计算机可读存储介质中。The computer readable program instructions described herein may be downloaded to various computing/processing devices from a computer readable storage medium, or to an external computer or external storage device over a network such as the Internet, a local area network, a wide area network, and/or a wireless network. The network may include copper transmission cables, fiber optic transmission, wireless transmission, routers, firewalls, switches, gateway computers, and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer-readable program instructions from a network and forwards the computer-readable program instructions for storage in a computer-readable storage medium in each computing/processing device .
用于执行本发明操作的计算机程序指令可以是汇编指令、指令集架构(ISA)指令、机器指令、机器相关指令、微代码、固件指令、状态设置数据、或者以一种或多种编程语言的任意组合编写的源代码或目标代码,所述编程语言包括面向对象的编程语言—诸如Smalltalk、C++等,以及常规 的过程式编程语言—诸如“C”语言或类似的编程语言。计算机可读程序指令可以完全地在用户计算机上执行、部分地在用户计算机上执行、作为一个独立的软件包执行、部分在用户计算机上部分在远程计算机上执行、或者完全在远程计算机或服务器上执行。在涉及远程计算机的情形中,远程计算机可以通过任意种类的网络—包括局域网(LAN)或广域网(WAN)—连接到用户计算机,或者,可以连接到外部计算机(例如利用因特网服务提供商来通过因特网连接)。在一些实施例中,通过利用计算机可读程序指令的状态信息来个性化定制电子电路,例如可编程逻辑电路、现场可编程门阵列(FPGA)或可编程逻辑阵列(PLA),该电子电路可以执行计算机可读程序指令,从而实现本发明的各个方面。The computer program instructions for carrying out the operations of the present invention may be assembly instructions, instruction set architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state setting data, or instructions in one or more programming languages. Source or object code, written in any combination, including object-oriented programming languages, such as Smalltalk, C++, etc., and conventional procedural programming languages, such as the "C" language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the user's computer and partly on a remote computer, or entirely on the remote computer or server implement. In the case of a remote computer, the remote computer may be connected to the user's computer through any kind of network, including a local area network (LAN) or a wide area network (WAN), or may be connected to an external computer (eg, using an Internet service provider through the Internet connect). In some embodiments, custom electronic circuits, such as programmable logic circuits, field programmable gate arrays (FPGAs), or programmable logic arrays (PLAs), can be personalized by utilizing state information of computer readable program instructions. Computer readable program instructions are executed to implement various aspects of the present invention.
这里参照根据本发明实施例的方法、装置(系统)和计算机程序产品的流程图和/或框图描述了本发明的各个方面。应当理解,流程图和/或框图的每个方框以及流程图和/或框图中各方框的组合,都可以由计算机可读程序指令实现。Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.
这些计算机可读程序指令可以提供给通用计算机、专用计算机或其它可编程数据处理装置的处理器,从而生产出一种机器,使得这些指令在通过计算机或其它可编程数据处理装置的处理器执行时,产生了实现流程图和/或框图中的一个或多个方框中规定的功能/动作的装置。也可以把这些计算机可读程序指令存储在计算机可读存储介质中,这些指令使得计算机、可编程数据处理装置和/或其他设备以特定方式工作,从而,存储有指令的计算机可读介质则包括一个制造品,其包括实现流程图和/或框图中的一个或多个方框中规定的功能/动作的各个方面的指令。These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer or other programmable data processing apparatus to produce a machine that causes the instructions when executed by the processor of the computer or other programmable data processing apparatus , resulting in means for implementing the functions/acts specified in one or more blocks of the flowchart and/or block diagrams. These computer readable program instructions can also be stored in a computer readable storage medium, these instructions cause a computer, programmable data processing apparatus and/or other equipment to operate in a specific manner, so that the computer readable medium on which the instructions are stored includes An article of manufacture comprising instructions for implementing various aspects of the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
也可以把计算机可读程序指令加载到计算机、其它可编程数据处理装置、或其它设备上,使得在计算机、其它可编程数据处理装置或其它设备上执行一系列操作步骤,以产生计算机实现的过程,从而使得在计算机、其它可编程数据处理装置、或其它设备上执行的指令实现流程图和/或框图中的一个或多个方框中规定的功能/动作。Computer readable program instructions can also be loaded onto a computer, other programmable data processing apparatus, or other equipment to cause a series of operational steps to be performed on the computer, other programmable data processing apparatus, or other equipment to produce a computer-implemented process , thereby causing instructions executing on a computer, other programmable data processing apparatus, or other device to implement the functions/acts specified in one or more blocks of the flowcharts and/or block diagrams.
附图中的流程图和框图显示了根据本发明的多个实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程 图或框图中的每个方框可以代表一个模块、程序段或指令的一部分,所述模块、程序段或指令的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个连续的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或动作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。对于本领域技术人员来说公知的是,通过硬件方式实现、通过软件方式实现以及通过软件和硬件结合的方式实现都是等价的。The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more functions for implementing the specified logical function(s) executable instructions. In some alternative implementations, the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It is also noted that each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations, can be implemented in dedicated hardware-based systems that perform the specified functions or actions , or can be implemented in a combination of dedicated hardware and computer instructions. It is well known to those skilled in the art that implementation in hardware, implementation in software, and implementation in a combination of software and hardware are all equivalent.
以上已经描述了本发明的各实施例,上述说明是示例性的,并非穷尽性的,并且也不限于所披露的各实施例。在不偏离所说明的各实施例的范围和精神的情况下,对于本技术领域的普通技术人员来说许多修改和变更都是显而易见的。本文中所用术语的选择,旨在最好地解释各实施例的原理、实际应用或对市场中的技术改进,或者使本技术领域的其它普通技术人员能理解本文披露的各实施例。本发明的范围由所附权利要求来限定。Various embodiments of the present invention have been described above, and the foregoing descriptions are exemplary, not exhaustive, and not limiting of the disclosed embodiments. Numerous modifications and variations will be apparent to those of ordinary skill in the art without departing from the scope and spirit of the described embodiments. The terminology used herein was chosen to best explain the principles of the embodiments, the practical application or technical improvement in the marketplace, or to enable others of ordinary skill in the art to understand the embodiments disclosed herein. The scope of the invention is defined by the appended claims.
Claims (10)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/CN2020/107822 WO2022027595A1 (en) | 2020-08-07 | 2020-08-07 | Method for reconstructing low-dose image by using multiscale feature sensing deep network |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/CN2020/107822 WO2022027595A1 (en) | 2020-08-07 | 2020-08-07 | Method for reconstructing low-dose image by using multiscale feature sensing deep network |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2022027595A1 true WO2022027595A1 (en) | 2022-02-10 |
Family
ID=80118955
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2020/107822 Ceased WO2022027595A1 (en) | 2020-08-07 | 2020-08-07 | Method for reconstructing low-dose image by using multiscale feature sensing deep network |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2022027595A1 (en) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN114821682A (en) * | 2022-06-30 | 2022-07-29 | 广州脉泽科技有限公司 | Multi-sample mixed palm vein identification method based on deep learning algorithm |
| CN116737925A (en) * | 2023-05-29 | 2023-09-12 | 北京龙智数科科技服务有限公司 | Model bypass tuning method and device |
| CN120107110A (en) * | 2025-05-09 | 2025-06-06 | 西安理工大学 | Low-dose CT image denoising method and system, and electronic equipment |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US8818065B2 (en) * | 2011-07-01 | 2014-08-26 | Carestream Health, Inc. | Methods and apparatus for scatter correction for CBCT system and cone-beam image reconstruction |
| CN111179366A (en) * | 2019-12-18 | 2020-05-19 | 深圳先进技术研究院 | Low-dose image reconstruction method and system based on anatomical difference prior |
| CN111325695A (en) * | 2020-02-29 | 2020-06-23 | 深圳先进技术研究院 | Low-dose image enhancement method, system and storage medium based on multiple dose levels |
| CN111340744A (en) * | 2020-02-20 | 2020-06-26 | 合肥工业大学 | Low-quality image downsampling method and system based on attention dual-stream deep network |
| CN111489406A (en) * | 2020-03-26 | 2020-08-04 | 深圳先进技术研究院 | Training and generating method, device and storage medium for generating high-energy CT image model |
-
2020
- 2020-08-07 WO PCT/CN2020/107822 patent/WO2022027595A1/en not_active Ceased
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US8818065B2 (en) * | 2011-07-01 | 2014-08-26 | Carestream Health, Inc. | Methods and apparatus for scatter correction for CBCT system and cone-beam image reconstruction |
| CN111179366A (en) * | 2019-12-18 | 2020-05-19 | 深圳先进技术研究院 | Low-dose image reconstruction method and system based on anatomical difference prior |
| CN111340744A (en) * | 2020-02-20 | 2020-06-26 | 合肥工业大学 | Low-quality image downsampling method and system based on attention dual-stream deep network |
| CN111325695A (en) * | 2020-02-29 | 2020-06-23 | 深圳先进技术研究院 | Low-dose image enhancement method, system and storage medium based on multiple dose levels |
| CN111489406A (en) * | 2020-03-26 | 2020-08-04 | 深圳先进技术研究院 | Training and generating method, device and storage medium for generating high-energy CT image model |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN114821682A (en) * | 2022-06-30 | 2022-07-29 | 广州脉泽科技有限公司 | Multi-sample mixed palm vein identification method based on deep learning algorithm |
| CN116737925A (en) * | 2023-05-29 | 2023-09-12 | 北京龙智数科科技服务有限公司 | Model bypass tuning method and device |
| CN120107110A (en) * | 2025-05-09 | 2025-06-06 | 西安理工大学 | Low-dose CT image denoising method and system, and electronic equipment |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Zhang et al. | TransCT: dual-path transformer for low dose computed tomography | |
| CN109035169B (en) | A deep network training method for unsupervised/semi-supervised CT image reconstruction | |
| CN111179366B (en) | Low-dose image reconstruction method and system based on anatomical difference prior | |
| CN109949276B (en) | Lymph node detection method for improving SegNet segmentation network | |
| CN109949235A (en) | A denoising method for chest X-ray film based on deep convolutional neural network | |
| CN111340903B (en) | Method and system for generating synthetic PET-CT images based on non-attenuation corrected PET images | |
| US10726948B2 (en) | Medical imaging device- and display-invariant segmentation and measurement | |
| US11915401B2 (en) | Apriori guidance network for multitask medical image synthesis | |
| WO2022027595A1 (en) | Method for reconstructing low-dose image by using multiscale feature sensing deep network | |
| CN110009613A (en) | Low-dose CT imaging method, device and system based on deep dense network | |
| CN111899315A (en) | Method for reconstructing low-dose image by using multi-scale feature perception depth network | |
| WO2022246677A1 (en) | Method for reconstructing enhanced ct image | |
| Takam et al. | Spark architecture for deep learning-based dose optimization in medical imaging | |
| WO2021168920A1 (en) | Low-dose image enhancement method and system based on multiple dose levels, and computer device, and storage medium | |
| WO2022094779A1 (en) | Deep learning framework and method for generating ct image from pet image | |
| WO2021159236A1 (en) | Method and system for generating composite pet-ct image on basis of non-attenuation-corrected pet image | |
| CN117541481B (en) | A low-dose CT image restoration method, system and storage medium | |
| Li et al. | MIS-Net: A deep learning-based multi-class segmentation model for CT images | |
| CN115049554B (en) | Low-dose CT denoising method based on Swin Transformer | |
| CN113053496B (en) | A Deep Learning Method for Low Dose Estimation in Medical Images | |
| WO2021051049A1 (en) | Few-view ct image reconstruction system | |
| Mangalagiri et al. | Toward generating synthetic CT volumes using a 3D-conditional generative adversarial network | |
| CN113129296B (en) | Image processing method based on transform domain de-noising autoencoder as prior | |
| Xia et al. | Dynamic controllable residual generative adversarial network for low-dose computed tomography imaging | |
| CN113469915A (en) | PET reconstruction method based on denoising and scoring matching network |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20948255 Country of ref document: EP Kind code of ref document: A1 |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 30.06.2023) |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 20948255 Country of ref document: EP Kind code of ref document: A1 |