WO2025035380A1 - Pet imaging method and apparatus based on optical flow registration, and device and storage medium - Google Patents
Pet imaging method and apparatus based on optical flow registration, and device and storage medium Download PDFInfo
- Publication number
- WO2025035380A1 WO2025035380A1 PCT/CN2023/113053 CN2023113053W WO2025035380A1 WO 2025035380 A1 WO2025035380 A1 WO 2025035380A1 CN 2023113053 W CN2023113053 W CN 2023113053W WO 2025035380 A1 WO2025035380 A1 WO 2025035380A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- feature map
- pet
- optical flow
- multimodal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
Definitions
- the present application belongs to the field of medical imaging technology, and in particular relates to a PET imaging method, device, equipment and storage medium based on optical flow registration.
- PET Pulsitron Emission Tomography
- a labeled radioactive drug usually a glucose molecule with a positron-emitting radioactive isotope (fluorine-18 labeled glucose, FDG).
- FDG positron-18 labeled glucose
- the existing PET imaging technology still has the following shortcomings:
- Slow image reconstruction speed The amount of data collected by full sampling is large, and complex image reconstruction algorithms are required, resulting in slow image reconstruction speed and prolonging the diagnosis process.
- Motion artifacts The long scanning process is easily affected by the patient's movement, resulting in image artifacts, thereby reducing image quality and affecting the doctor's accurate judgment of the patient's condition.
- the present application provides a PET imaging method, apparatus, device and storage medium based on optical flow registration, aiming to solve at least one of the above-mentioned technical problems in the prior art to a certain extent.
- a PET imaging method based on optical flow registration comprising:
- PET images and MR images of the irradiated area were collected separately;
- the first multimodal enhanced feature map and the second multimodal enhanced feature map are fused by a cross-modal feature fusion module using a multi-head cross-attention mechanism to generate a multimodal fusion feature map, and a PET reconstructed image is generated according to the multimodal fusion feature map.
- the technical solution adopted by the embodiment of the present application also includes: before obtaining the optical flow information of the PET image and the MR image through the optical flow registration network, it also includes:
- the Canny operator is used to extract contour information from the PET image and the MR image respectively.
- the technical solution adopted by the embodiment of the present application also includes: the use of the Canny operator to extract contour information from the PET image and the MR image respectively is specifically as follows:
- the PET image and the MR image are subjected to image corrosion and dilation processing by using morphological image processing technology, small objects and thin lines in the PET image and the MR image are eliminated, the basic contour shape of the radiation part is retained, and finally a binary mask composed of 0 and 1 is generated to represent the contour information of the PET image and the MR image respectively.
- the optical flow registration network is a
- the U-net model includes an encoder and a decoder.
- the optical flow information of the PET image and the MR image is obtained through the optical flow registration network.
- the first feature map and the second feature map of the PET image and the corrected MR image are respectively extracted as follows:
- the contour information of the PET image and the MR image is input into the optical flow registration network, the encoder part of the optical flow registration network uses three 1x1 convolution layers to extract feature maps, and uses a 3 ⁇ 3 convolution layer to downsample the extracted feature maps; the decoder part of the optical flow registration network uses upsampling and 1x1 convolution layers to expand and compress the extracted feature maps to restore the size of the optical flow map, and generates a weight mask with a channel number of 64, generates the weight mask by applying a Sigmoid activation function, and outputs a first feature map F pet and a second feature map F mr of the PET image and the corrected MR image.
- the technical solution adopted in the embodiment of the present application also includes: the spatial channel feature enhancement module performs feature enhancement processing on the first feature map and the second feature map in the spatial dimension and the channel dimension respectively, and fuses the feature enhancement processing results in the spatial dimension and the channel dimension to generate the first multimodal enhanced feature map of the PET image and the second multimodal enhanced feature map of the corrected MR image.
- the spatial channel feature enhancement module performs feature enhancement processing on the first feature map and the second feature map in the spatial dimension and the channel dimension respectively, and fuses the feature enhancement processing results in the spatial dimension and the channel dimension to generate the first multimodal enhanced feature map of the PET image and the second multimodal enhanced feature map of the corrected MR image.
- the spatial channel feature enhancement module includes a spatial enhancement module and a channel enhancement module.
- the spatial enhancement module compresses the first feature map F pet and the second feature map F mr through two convolutional layers to obtain two matrices S1 and S2.
- the matrices S1 and S2 are connected in the spatial dimension to obtain a single-dimensional matrix S containing attention information of the PET and MR image streams.
- the first feature map F pet and the second feature map F mr are activated by a Sigmoid function and element-by-element operations are performed to obtain the first spatial enhancement feature map F s-pet and the second spatial enhancement feature map F s-mr of the PET image and the corrected MR image in the spatial dimension respectively.
- the channel enhancement module uses global average pooling to compress the first feature map F pet and the second feature map F mr , and creates a channel descriptor, and calculates the encoding of the PET image and the MR image through the channel descriptor Vectors C1 and C2, connecting the encoding vectors C1 and C2, and creating a fused embedding vector through a fully connected layer, and then activating the embedding vector with a Sigmoid function and performing element-by-element operations to obtain the first channel enhanced feature map Fc -pet and the second channel enhanced feature map Fc -mr of the PET image and the corrected MR image respectively;
- the first spatial enhancement feature map and the first channel enhancement feature map of the PET image and the second spatial enhancement feature map and the second channel enhancement feature map of the corrected MR image are fused respectively to generate a first multimodal enhancement feature map F sc-pet of the PET image and a second multimodal enhancement feature map F sc-mr of the corrected MR image.
- the technical solution adopted in the embodiment of the present application also includes: after the cross-modal feature fusion module adopts a multi-head cross attention mechanism to fuse the first multimodal enhanced feature map and the second multimodal enhanced feature map, the multimodal fusion feature map is generated specifically as follows:
- the cross-modal feature fusion module adopts a multi-head cross attention mechanism and a Transformer architecture to perform a linear projection operation on the first multimodal enhanced feature map F sc-pet and the second multimodal enhanced feature map F sc-mr to generate a query matrix Q, a key matrix K and a value matrix V for PET images and MR images, respectively, and perform an attention operation between the vectorized features of the two modalities:
- ⁇ represents the sigmoid function.
- K pet , V pet represent the query matrix Q, key matrix K and value matrix V of PET images respectively.
- K mr , V mr represent the query matrix Q, key matrix K and value matrix V of MR images respectively;
- the attention maps of the two modalities are spliced in the channel dimension to generate a multimodal fusion feature map.
- the technical solution adopted by the embodiment of the present application also includes: generating a PET reconstructed image according to the multimodal fusion feature map is specifically:
- the multimodal fusion feature map is input into a decoder through a jump connection, and a PET reconstructed image is output through the decoder.
- a PET imaging device based on optical flow registration comprising:
- Image acquisition module used to respectively acquire PET images and MR images of the radiation site
- An optical flow registration module used for acquiring optical flow information of the PET image and the MR image through an optical flow registration network, and extracting a first feature map and a second feature map of the PET image and the corrected MR image respectively after correcting the position of the MR image according to the optical flow information;
- a feature enhancement module is used to perform feature enhancement processing of the first feature map and the second feature map in the spatial dimension and the channel dimension respectively through the spatial channel feature enhancement module, and fuse the feature enhancement processing results in the spatial dimension and the channel dimension to generate a first multimodal enhanced feature map of the PET image and a second multimodal enhanced feature map of the corrected MR image;
- Feature fusion module used to fuse the first multimodal enhanced feature map and the second multimodal enhanced feature map through a cross-modal feature fusion module using a multi-head cross-attention mechanism to generate a multimodal fusion feature map, and generate a PET reconstructed image based on the multimodal fusion feature map.
- a device includes a processor and a memory coupled to the processor, wherein:
- the memory stores program instructions for implementing the PET imaging method based on optical flow registration
- the processor is used to execute the program instructions stored in the memory to control a PET imaging method based on optical flow registration.
- a storage medium storing program instructions executable by a processor, wherein the program instructions are used to execute the PET imaging method based on optical flow registration.
- the beneficial effects produced by the embodiments of the present application are: the PET imaging method, device, equipment and storage medium based on optical flow registration in the embodiments of the present application.
- FIG1 is a flow chart of a PET imaging method based on optical flow registration according to an embodiment of the present application
- FIG2 is a schematic structural diagram of a PET imaging device based on optical flow registration according to an embodiment of the present application
- FIG3 is a schematic diagram of the device structure of an embodiment of the present application.
- FIG. 4 is a schematic diagram of the structure of a storage medium according to an embodiment of the present application.
- first”, “second” and “third” in this application are only used for descriptive purposes and cannot be understood as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Therefore, the features defined as “first”, “second” and “third” may explicitly or implicitly include at least one of the features.
- the meaning of “multiple” is at least two, such as two, three, etc., unless otherwise clearly and specifically defined. All directional indications in the embodiments of the present application (such as up, down, left, right, front, back%) are only used to explain the relative position relationship, movement, etc. between the components in a certain specific posture (as shown in the accompanying drawings). If the specific posture changes, the directional indication will also change accordingly.
- FIG1 is a flow chart of a PET imaging method based on optical flow registration according to an embodiment of the present application.
- the PET imaging method based on optical flow registration according to an embodiment of the present application comprises the following steps:
- the lower and upper thresholds of contour information extraction are set to 20 and 50, respectively.
- the contour information extraction process is as follows: First, morphological image processing technology is used to perform image erosion and dilation on the PET image and the MR image to eliminate small objects and thin lines in the PET image and the MR image, while retaining the basic contour shape of the radiation site. During the processing, a circular kernel of size 5 ⁇ 5 is used to ensure isotropic erosion and dilation processes to maintain the integrity of the contour shape and prevent unnecessary distortion. Finally, a binary mask consisting of 0 and 1 is generated to represent the contour information I pet and I mr of the PET image and the MR image, respectively.
- the optical flow registration network calculates the optical flow information according to the contour information of the PET image and the MR image, corrects the position of the MR image according to the optical flow information, and extracts the first feature map and the second feature map of the PET image and the corrected MR image respectively;
- the optical flow registration network is a U-net-based model that uses the PET image as a reference and uses the inverse image transformation to correct the position of the MR image to achieve the effect of PET/MR dual-modality registration and alignment.
- the optical flow registration network consists of two parts: an encoder and a decoder.
- the encoder part uses three
- the 1x1 convolution layer extracts the feature map, and the 3 ⁇ 3 convolution layer is used to downsample the extracted feature map by 2 times, thereby gradually reducing the size of the extracted feature map from 256 ⁇ 256 to 32 ⁇ 32.
- the decoder part uses upsampling and 1x1 convolution layers to expand and compress the extracted feature map to restore the size of the optical flow map, and finally generates a weight mask with 64 channels. Finally, the weight mask is generated by applying the Sigmoid activation function, and the first feature map F pet and the second feature map F mr of the PET image and the corrected MR image are output.
- the optical flow registration network parameter settings are shown in Table 1 below:
- S130 inputting the first feature map and the second feature map into a spatial channel feature enhancement module, and performing feature enhancement processing on the first feature map and the second feature map in the spatial dimension and the channel dimension respectively by the spatial channel feature enhancement module, so as to obtain a first spatial enhancement feature map in the spatial dimension and a first channel enhancement feature map in the channel dimension of the PET image, and a second spatial enhancement feature map in the spatial dimension and a second channel enhancement feature map in the channel dimension of the corrected MR image respectively;
- the spatial channel feature enhancement module includes a spatial enhancement module and a channel enhancement module.
- the spatial enhancement module can improve the network's attention to important areas such as high-density noise.
- the first feature map F pet and the second feature map F mr of the input are compressed by two convolutional layers K1 and K2 to obtain two matrices S1 and S2.
- the matrices S1 and S2 are connected in the spatial dimension to obtain a single-dimensional matrix S containing the attention information of the PET and MR image streams.
- the first spatial enhancement feature map F s-pet and the second spatial enhancement feature map F s-mr of the PET image and the corrected MR image in the spatial dimension are finally obtained by activating the input feature map with a Sigmoid function and performing element-by-element operations.
- the parameter settings of the spatial enhancement module are shown in Table 3 below:
- the channel enhancement module global average pooling is used to compress each H ⁇ W input feature map into a 1 ⁇ 1 value, and a channel descriptor is created.
- the encoding vectors C1 and C2 of the PET image and MR image are calculated through the channel descriptor, and then the encoding vectors C1 and C2 are connected, and a fused embedding vector is created through a fully connected layer.
- the first channel enhanced feature map Fc -pet and the second channel enhanced feature map Fc -mr of the PET image and the corrected MR image are obtained by activating the embedding vector with a Sigmoid function and performing element-by-element operations.
- the channel enhancement module parameter settings are shown in Table 4 below:
- S140 Fusing the first spatial enhancement feature map and the first channel enhancement feature map of the PET image and the second spatial enhancement feature map and the second channel enhancement feature map of the corrected MR image, respectively. generating a first multimodal enhanced feature map of the PET image and a second multimodal enhanced feature map of the corrected MR image;
- a first multimodal enhancement feature map F sc-pet and a second multimodal enhancement feature map F sc-mr combining the two modal information of PET and MR are generated, thereby comprehensively considering the two modal information and making the feature enhancement process more effective.
- the first multimodal enhancement feature map F sc-pet and the second multimodal enhancement feature map F sc-mr can be defined by the following formula:
- F sc-pet F pet +F s-pet +F c-pet (1)
- F sc-mr F mr +F s-mr +F c-mr (2)
- the cross-modal feature fusion module adopts a multi-head cross-attention mechanism (CM-FFM) and a Transformer architecture to obtain the query matrix Q (queries), key matrix K (keys) and value matrix V (values) by calculating the dot product between the PET and MR images, thereby estimating the correlation between the multimodal enhanced feature maps, and calculating the cross-attention weights based on the softmax function through scaling and normalization.
- CM-FFM multi-head cross-attention mechanism
- the first multimodal enhanced feature map F sc-pet and the second multimodal enhanced feature map F sc-mr are first linearly projected to generate the query matrix Q, key matrix K and value matrix V of the PET image and the MR image, and then the attention operation is performed between the vectorized features of the two modalities.
- the calculation formula is as follows:
- ⁇ represents the sigmoid function, which is used to map the input to the range of 0 to 1, ensuring that the attention weight is within a reasonable range to represent the correlation between the two modalities and prevent the weight from deviating too much from the normal range.
- K pet , V pet represent the query matrix Q, key matrix K and value matrix V of PET images respectively.
- K pet and V pet are both obtained by projecting the first multimodal enhanced feature map F sc-pet , and represent information extracted from the PET image.
- Kmr , Vmr represent the query matrix Q, key matrix K and value matrix V of MR images respectively.
- K mr and V mr are both obtained by projecting the second multimodal enhanced feature map F sc-mr , and represent the information extracted from the MR image.
- the attention maps of the two modalities are spliced in the channel dimension to generate a multi-modal fusion feature map, thereby providing rich and favorable image details for PET image denoising.
- the multi-head cross attention mechanism parameter settings are shown in Table 5 below:
- S160 input the multimodal fusion feature map into the decoder through the skip connection, and output the denoised PET reconstructed image through the decoder;
- the embodiment of the present application selects a bilinear upsampling method to reduce The magnification factor is 2, and a convolutional layer is used at each stage, which helps to restore the multi-level feature map to the original image resolution and generate a relatively smooth feature map.
- the embodiment of the present application introduces batch normalization in each convolutional layer during the downsampling and upsampling of the image, and introduces jump connections at the encoding and decoding ends to retain the most favorable image details from the two modalities.
- CT images can also be used as a substitute for MR images to denoise PET images.
- Table 6 the decoder parameter settings are shown in Table 6 below:
- the PET imaging method based on optical flow registration in the embodiment of the present application utilizes the contour information of the PET image and the MR image, and aligns the PET image and the MR image through the optical flow correction network, so that the multimodal data can be better aligned, effectively reducing the image distortion and artifacts caused by the registration problem, and improving the clarity and accuracy of the image;
- the spatial channel feature enhancement module uses matrices and vectors to represent spatial attention and channel attention, which is calculated not only based on the information of the current modality, but also combined with the influence of other modalities to ensure the effectiveness of feature extraction at each downsampling stage, which can better capture the important features in the image and improve the accuracy and robustness of feature extraction.
- FIG2 is a schematic diagram of the structure of a PET imaging device based on optical flow registration according to an embodiment of the present application.
- the PET imaging device 40 based on optical flow registration according to an embodiment of the present application comprises:
- Image acquisition module 41 used to respectively acquire PET images and MR images of the radiation site;
- An optical flow registration module 42 is used to obtain optical flow information of the PET image and the MR image through an optical flow registration network, and after correcting the position of the MR image according to the optical flow information, extract a first feature map and a second feature map of the PET image and the corrected MR image respectively;
- the feature enhancement module 43 is used to perform feature enhancement processing of the first feature map and the second feature map in the spatial dimension and the channel dimension respectively through the spatial channel feature enhancement module, and fuse the feature enhancement processing results in the spatial dimension and the channel dimension to generate a first multimodal enhanced feature map of the PET image and a second multimodal enhanced feature map of the corrected MR image;
- Feature fusion module 44 used to fuse the first multimodal enhanced feature map and the second multimodal enhanced feature map through a cross-modal feature fusion module using a multi-head cross-attention mechanism to generate a multimodal fusion feature map, and generate a PET reconstructed image according to the multimodal fusion feature map.
- FIG3 is a schematic diagram of the device structure of an embodiment of the present application.
- the device 50 includes:
- a memory 51 storing executable program instructions
- a processor 52 connected to the memory 51;
- the processor 52 is used to call the executable program instructions stored in the memory 51 and perform the following steps: respectively collect PET images and MR images of the radiation site; obtain the optical flow information of the PET image and the MR image through the optical flow registration network, correct the position of the MR image according to the optical flow information, and respectively extract the first feature map and the second feature map of the PET image and the corrected MR image; respectively perform feature enhancement processing of the spatial dimension and the channel dimension on the first feature map and the second feature map through the spatial channel feature enhancement module, and fuse the feature enhancement processing results of the spatial dimension and the channel dimension to generate The first multimodal enhanced feature map of the PET image and the second multimodal enhanced feature map of the corrected MR image; after fusing the first multimodal enhanced feature map and the second multimodal enhanced feature map through a cross-modal feature fusion module using a multi-head cross-attention mechanism, a multimodal fusion feature map is generated, and a PET reconstructed image is generated according to the multimodal fusion feature map.
- the processor 52 may also be referred to as a CPU (Central Processing Unit).
- the processor 52 may be an integrated circuit chip having signal processing capabilities.
- the processor 52 may also be a general-purpose processor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
- DSP digital signal processor
- ASIC application-specific integrated circuit
- FPGA field-programmable gate array
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor, etc.
- FIG. 4 is a schematic diagram of the structure of the storage medium of the embodiment of the present application.
- the storage medium of the embodiment of the present application stores a program instruction 61 that can implement the following steps: respectively collect PET images and MR images of the radiation site; obtain the optical flow information of the PET image and the MR image through the optical flow registration network, and extract the first feature map and the second feature map of the PET image and the corrected MR image respectively after correcting the position of the MR image according to the optical flow information; perform feature enhancement processing of the spatial dimension and the channel dimension on the first feature map and the second feature map respectively through the spatial channel feature enhancement module, and fuse the feature enhancement processing results of the spatial dimension and the channel dimension to generate the first multimodal enhancement feature map of the PET image and the second multimodal enhancement feature map of the corrected MR image; after fusing the first multimodal enhancement feature map and the second multimodal enhancement feature map through the cross-modal feature fusion module using a multi-head cross attention mechanism, a multimodal fusion feature map is generated, and a PET reconstructed image is
- the program instructions 61 may be stored in the above storage medium in the form of a software product, including several instructions for enabling a device (which may be a personal computer, a server, or a network device, etc.) or a processor to execute the entire method of each embodiment of the present application.
- the aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (ROM), random access memory (RAM), disk or optical disk and other media that can store program instructions, or terminal devices such as computers, servers, mobile phones, tablets, etc.
- the server can be an independent server, or it can be a cloud server that provides basic cloud computing services such as cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, middleware services, domain name services, security services, content delivery networks (CDN), and big data and artificial intelligence platforms.
- cloud computing services such as cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, middleware services, domain name services, security services, content delivery networks (CDN), and big data and artificial intelligence platforms.
- each functional unit in each embodiment of the present application can be integrated into one processing unit, or each unit can exist physically separately, or two or more units can be integrated into one unit.
- the above integrated unit can be implemented in the form of hardware or in the form of software functional units. The above is only an implementation method of the present application, and does not limit the patent scope of the present application. Any equivalent structure or equivalent process transformation made by using the description and drawings of this application, or directly or indirectly used in other related technical fields, is also included in the patent protection scope of the present application.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Graphics (AREA)
- Computer Hardware Design (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Image Processing (AREA)
Abstract
Description
本申请属于医学成像技术领域,特别涉及一种基于光流配准的PET成像方法、装置、设备以及存储介质。The present application belongs to the field of medical imaging technology, and in particular relates to a PET imaging method, device, equipment and storage medium based on optical flow registration.
PET(Positron Emission Tomography,正电子发射断层成像)是一种重要的医学成像技术,用于获取生物体内部的功能和代谢信息。它通过检测放射性同位素标记的生物分子在体内的分布和浓度来揭示组织和器官的代谢活动。在PET成像过程中,患者会接受一种被标记的放射性药物,通常是带有正电子放射性同位素的葡萄糖分子(氟-18标记的葡萄糖,FDG)。这些药物被注射到患者体内后,它们会被身体的组织和器官代谢吸收。正电子发生衰变后会释放出两个高能正电子,这些正电子与体内的电子相遇并发生湮灭反应。在湮灭过程中,产生了两个相对方向的伽玛光子,这些光子可以被探测器捕获。PET (Positron Emission Tomography) is an important medical imaging technology used to obtain functional and metabolic information inside organisms. It reveals the metabolic activities of tissues and organs by detecting the distribution and concentration of radioactive isotope-labeled biological molecules in the body. During PET imaging, patients receive a labeled radioactive drug, usually a glucose molecule with a positron-emitting radioactive isotope (fluorine-18 labeled glucose, FDG). After these drugs are injected into the patient's body, they are metabolized and absorbed by the body's tissues and organs. When the positron decays, two high-energy positrons are released, which meet and annihilate with electrons in the body. During the annihilation process, two gamma photons in opposite directions are produced, which can be captured by the detector.
由于PET扫描过程中放射性同位素的使用以及与其相关的辐射剂量是患者和医护人员面临的潜在风险,PET辐射剂量问题越来越受到人们的重视。然而,随着辐射剂量和扫描时间的降低,更多的噪声会出现在成像过程中,从而导致PET成像质量较差。The use of radioisotopes in PET scanning and the associated radiation dose are potential risks to patients and medical staff, and the issue of PET radiation dose has received increasing attention. However, as the radiation dose and scanning time decrease, more noise will appear in the imaging process, resulting in poor PET image quality.
为了解决上述问题,Yuru He等人于2021年在IEEE Access期刊上发表了题为《Dynamic PET Image Denoising With Deep Learning-Based Joint Filtering》的文章。该研究基于空间变体线性表示模型(SVLRM)的结构,对网络框架进行了改进。研究采用了具有12个卷积层的CNN对PET噪声图像进行去噪。在网络输入端,该文章将PET和MR(Magnetic Resonance,核磁共振)图像进行融合, 通过通道维度的拼接将两个模态的信息结合起来。随后,通过共享权重的网络模型处理融合后的输入,并输出最终的去噪PET图像。然而,由于PET和MR模态之间可能存在不配准问题,该文章在多模态融合过程中没有有效解决这个问题;另外,信息融合的过程也存在改进的空间。In order to solve the above problems, Yuru He et al. published an article titled "Dynamic PET Image Denoising With Deep Learning-Based Joint Filtering" in the IEEE Access journal in 2021. The study improved the network framework based on the structure of the spatial variant linear representation model (SVLRM). The study used a CNN with 12 convolutional layers to denoise PET noisy images. At the network input, the article fused PET and MR (Magnetic Resonance) images. The information of the two modalities is combined by splicing in the channel dimension. Subsequently, the fused input is processed by a network model with shared weights, and the final denoised PET image is output. However, due to the possible misalignment problem between the PET and MR modalities, this article did not effectively solve this problem in the multimodal fusion process; in addition, there is room for improvement in the information fusion process.
同时,现有的PET成像技术还存在以下不足:At the same time, the existing PET imaging technology still has the following shortcomings:
1、PET扫描时间长:PET扫描过程中,需要对患者进行连续的图像采集,导致整个扫描过程变得相对较长,会给患者带来不便,并且可能增加扫描过程中的不适感。1. Long PET scanning time: During the PET scanning process, continuous image acquisition of the patient is required, which makes the entire scanning process relatively long, causing inconvenience to the patient and may increase discomfort during the scanning process.
2、图像重建速度慢:全采样所采集的数据量较大,需要进行复杂的图像重建算法,导致图像重建速度较慢,延长了诊断过程的时间。2. Slow image reconstruction speed: The amount of data collected by full sampling is large, and complex image reconstruction algorithms are required, resulting in slow image reconstruction speed and prolonging the diagnosis process.
3、运动伪影:长时间的扫描过程容易受到患者运动的影响,导致图像出现伪影,从而降低图像质量,影响医生对患者病情的准确判断。3. Motion artifacts: The long scanning process is easily affected by the patient's movement, resulting in image artifacts, thereby reducing image quality and affecting the doctor's accurate judgment of the patient's condition.
4、缺乏多模态数据使用及配准:大部分算法仅基于单一的PET模态进行去噪,忽略了利用MR先验知识引导PET图像去噪的可能性,限制了算法在利用多模态信息方面的能力。4. Lack of multimodal data usage and registration: Most algorithms perform denoising based on only a single PET modality, ignoring the possibility of using MR prior knowledge to guide PET image denoising, which limits the algorithm's ability to utilize multimodal information.
5、多模态融合过程处理简单:现有算法在多模态融合中通常只是简单地将多模态数据在通道维度上拼接起来,缺乏对多模态特征有效交互的深入探究。这种简单的拼接方法可能无法充分利用多模态数据之间的相关性和互补性,导致融合结果不尽如人意。5. Simple processing of multimodal fusion process: Existing algorithms usually simply splice multimodal data in the channel dimension in multimodal fusion, lacking in-depth exploration of the effective interaction of multimodal features. This simple splicing method may not fully utilize the correlation and complementarity between multimodal data, resulting in unsatisfactory fusion results.
发明内容Summary of the invention
本申请提供了一种基于光流配准的PET成像方法、装置、设备以及存储介质,旨在至少在一定程度上解决现有技术中的上述技术问题之一。 The present application provides a PET imaging method, apparatus, device and storage medium based on optical flow registration, aiming to solve at least one of the above-mentioned technical problems in the prior art to a certain extent.
为了解决上述问题,本申请提供了如下技术方案:In order to solve the above problems, this application provides the following technical solutions:
一种基于光流配准的PET成像方法,包括:A PET imaging method based on optical flow registration, comprising:
分别采集放射部位的PET图像和MR图像;PET images and MR images of the irradiated area were collected separately;
通过光流配准网络获取所述PET图像和MR图像的光流信息,根据所述光流信息校正MR图像的位置后,分别提取所述PET图像和校正后MR图像的第一特征图和第二特征图;Acquire optical flow information of the PET image and the MR image through an optical flow registration network, correct the position of the MR image according to the optical flow information, and then extract a first feature map and a second feature map of the PET image and the corrected MR image respectively;
通过空间通道特征增强模块对所述第一特征图和第二特征图分别进行空间维度和通道维度的特征增强处理,并将空间维度和通道维度的特征增强处理结果进行融合,生成所述PET图像的第一多模态增强特征图和校正后MR图像的第二多模态增强特征图;Performing feature enhancement processing of the spatial dimension and the channel dimension on the first feature map and the second feature map respectively by a spatial channel feature enhancement module, and fusing the feature enhancement processing results of the spatial dimension and the channel dimension to generate a first multimodal enhanced feature map of the PET image and a second multimodal enhanced feature map of the corrected MR image;
通过跨模态特征融合模块采用多头交叉注意力机制对所述第一多模态增强特征图和第二多模态增强特征图进行融合后,生成多模态融合特征图,根据所述多模态融合特征图生成PET重建图像。The first multimodal enhanced feature map and the second multimodal enhanced feature map are fused by a cross-modal feature fusion module using a multi-head cross-attention mechanism to generate a multimodal fusion feature map, and a PET reconstructed image is generated according to the multimodal fusion feature map.
本申请实施例采取的技术方案还包括:所述通过光流配准网络获取所述PET图像和MR图像的光流信息之前,还包括:The technical solution adopted by the embodiment of the present application also includes: before obtaining the optical flow information of the PET image and the MR image through the optical flow registration network, it also includes:
使用Canny算子分别对所述PET图像和MR图像进行轮廓信息提取。The Canny operator is used to extract contour information from the PET image and the MR image respectively.
本申请实施例采取的技术方案还包括:所述使用Canny算子分别对PET图像和MR图像进行轮廓信息提取具体为:The technical solution adopted by the embodiment of the present application also includes: the use of the Canny operator to extract contour information from the PET image and the MR image respectively is specifically as follows:
采用形态学图像处理技术对所述PET图像和MR图像进行图像腐蚀和膨胀处理,消除所述PET图像和MR图像中的小物体和细线,保留放射部位的基本轮廓形状,最后生成一个由0和1组成的二进制掩膜,分别用于表示PET图像和MR图像的轮廓信息。The PET image and the MR image are subjected to image corrosion and dilation processing by using morphological image processing technology, small objects and thin lines in the PET image and the MR image are eliminated, the basic contour shape of the radiation part is retained, and finally a binary mask composed of 0 and 1 is generated to represent the contour information of the PET image and the MR image respectively.
本申请实施例采取的技术方案还包括:所述光流配准网络是一个基于 U-net的模型,包括编码器和解码器两部分,所述通过光流配准网络获取所述PET图像和MR图像的光流信息,根据所述光流信息校正MR图像的位置后,分别提取所述PET图像和校正后MR图像的第一特征图和第二特征图具体为:The technical solution adopted by the embodiment of the present application also includes: the optical flow registration network is a The U-net model includes an encoder and a decoder. The optical flow information of the PET image and the MR image is obtained through the optical flow registration network. After the position of the MR image is corrected according to the optical flow information, the first feature map and the second feature map of the PET image and the corrected MR image are respectively extracted as follows:
将所述PET图像和MR图像的轮廓信息输入光流配准网络,所述光流配准网络的编码器部分使用三个1x1的卷积层提取特征图,并使用3×3卷积层对提取的特征图进行下采样;所述光流配准网络的解码器部分使用上采样和1x1卷积层对提取的特征图进行扩张和压缩,以恢复光流图的大小,并生成通道数为64的权重掩码,通过应用Sigmoid激活函数生成权重掩码,并输出PET图像和校正后MR图像的第一特征图Fpet和第二特征图Fmr。The contour information of the PET image and the MR image is input into the optical flow registration network, the encoder part of the optical flow registration network uses three 1x1 convolution layers to extract feature maps, and uses a 3×3 convolution layer to downsample the extracted feature maps; the decoder part of the optical flow registration network uses upsampling and 1x1 convolution layers to expand and compress the extracted feature maps to restore the size of the optical flow map, and generates a weight mask with a channel number of 64, generates the weight mask by applying a Sigmoid activation function, and outputs a first feature map F pet and a second feature map F mr of the PET image and the corrected MR image.
本申请实施例采取的技术方案还包括:所述通过空间通道特征增强模块对所述第一特征图和第二特征图分别进行空间维度和通道维度的特征增强处理,并将空间维度和通道维度的特征增强处理结果进行融合,生成所述PET图像的第一多模态增强特征图和校正后MR图像的第二多模态增强特征图具体为:The technical solution adopted in the embodiment of the present application also includes: the spatial channel feature enhancement module performs feature enhancement processing on the first feature map and the second feature map in the spatial dimension and the channel dimension respectively, and fuses the feature enhancement processing results in the spatial dimension and the channel dimension to generate the first multimodal enhanced feature map of the PET image and the second multimodal enhanced feature map of the corrected MR image. Specifically,
所述空间通道特征增强模块包括空间增强模块和通道增强模块,所述在空间增强模块通过两个卷积层和对所述第一特征图Fpet和第二特征图Fmr进行压缩,得到两个矩阵S1和S2,将矩阵S1和S2在空间维度上进行连接,得到一个单一维度且包含PET和MR图像流的注意力信息的矩阵S,并通过对第一特征图Fpet和第二特征图Fmr进行Sigmoid函数激活并进行逐元素操作,分别得到PET图像和校正后MR图像在空间维度的第一空间增强特征图Fs-pet和第二空间增强特征图Fs-mr;The spatial channel feature enhancement module includes a spatial enhancement module and a channel enhancement module. The spatial enhancement module compresses the first feature map F pet and the second feature map F mr through two convolutional layers to obtain two matrices S1 and S2. The matrices S1 and S2 are connected in the spatial dimension to obtain a single-dimensional matrix S containing attention information of the PET and MR image streams. The first feature map F pet and the second feature map F mr are activated by a Sigmoid function and element-by-element operations are performed to obtain the first spatial enhancement feature map F s-pet and the second spatial enhancement feature map F s-mr of the PET image and the corrected MR image in the spatial dimension respectively.
所述通道增强模块使用全局平均池化将第一特征图Fpet和第二特征图Fmr进行压缩,并创建通道描述符,通过通道描述符计算PET图像和MR图像的编码 向量C1和C2,将所述编码向量C1和C2进行连接,并通过全连接层创建融合的嵌入向量,然后对所述嵌入向量进行Sigmoid函数激活并进行逐元素操作,分别得到所述PET图像和校正后MR图像的第一通道增强特征图Fc-pet和第二通道增强特征图Fc-mr;The channel enhancement module uses global average pooling to compress the first feature map F pet and the second feature map F mr , and creates a channel descriptor, and calculates the encoding of the PET image and the MR image through the channel descriptor Vectors C1 and C2, connecting the encoding vectors C1 and C2, and creating a fused embedding vector through a fully connected layer, and then activating the embedding vector with a Sigmoid function and performing element-by-element operations to obtain the first channel enhanced feature map Fc -pet and the second channel enhanced feature map Fc -mr of the PET image and the corrected MR image respectively;
分别对所述PET图像的第一空间增强特征图和第一通道增强特征图以及校正后MR图像的第二空间增强特征图和第二通道增强特征图进行融合,生成所述PET图像的第一多模态增强特征图Fsc-pet和校正后MR图像的第二多模态增强特征图Fsc-mr。The first spatial enhancement feature map and the first channel enhancement feature map of the PET image and the second spatial enhancement feature map and the second channel enhancement feature map of the corrected MR image are fused respectively to generate a first multimodal enhancement feature map F sc-pet of the PET image and a second multimodal enhancement feature map F sc-mr of the corrected MR image.
本申请实施例采取的技术方案还包括:所述通过跨模态特征融合模块采用多头交叉注意力机制对所述第一多模态增强特征图和第二多模态增强特征图进行融合后,生成多模态融合特征图具体为:The technical solution adopted in the embodiment of the present application also includes: after the cross-modal feature fusion module adopts a multi-head cross attention mechanism to fuse the first multimodal enhanced feature map and the second multimodal enhanced feature map, the multimodal fusion feature map is generated specifically as follows:
所述跨模态特征融合模块采用多头交叉注意力机制和Transformer架构,将所述第一多模态增强特征图Fsc-pet和第二多模态增强特征图Fsc-mr进行线性投影操作,分别生成PET图像和MR图像的查询矩阵Q、键矩阵K和值矩阵V,并在两种模态的向量化特征之间进行注意力操作:
The cross-modal feature fusion module adopts a multi-head cross attention mechanism and a Transformer architecture to perform a linear projection operation on the first multimodal enhanced feature map F sc-pet and the second multimodal enhanced feature map F sc-mr to generate a query matrix Q, a key matrix K and a value matrix V for PET images and MR images, respectively, and perform an attention operation between the vectorized features of the two modalities:
上述公式中,σ表示sigmoid函数,表示特征维度的平方根,Kpet,Vpet分别表示PET图像的查询矩阵Q、键矩阵K和值矩阵V,Kmr,Vmr分别表示MR图像的查询矩阵Q、键矩阵K和值矩阵V; In the above formula, σ represents the sigmoid function. represents the square root of the feature dimension, K pet , V pet represent the query matrix Q, key matrix K and value matrix V of PET images respectively. K mr , V mr represent the query matrix Q, key matrix K and value matrix V of MR images respectively;
将两种模态的注意力操作结果通过平均池化操作进行压缩后,在通道维度上将两种模态的注意力图进行拼接,生成多模态融合特征图。After compressing the attention operation results of the two modalities through the average pooling operation, the attention maps of the two modalities are spliced in the channel dimension to generate a multimodal fusion feature map.
本申请实施例采取的技术方案还包括:所述根据所述多模态融合特征图生成PET重建图像具体为:The technical solution adopted by the embodiment of the present application also includes: generating a PET reconstructed image according to the multimodal fusion feature map is specifically:
通过跳跃连接将所述多模态融合特征图输入解码器,通过所述解码器输出PET重建图像。The multimodal fusion feature map is input into a decoder through a jump connection, and a PET reconstructed image is output through the decoder.
本申请实施例采取的另一技术方案为:一种基于光流配准的PET成像装置,包括:Another technical solution adopted by the embodiment of the present application is: a PET imaging device based on optical flow registration, comprising:
图像采集模块:用于分别采集放射部位的PET图像和MR图像;Image acquisition module: used to respectively acquire PET images and MR images of the radiation site;
光流配准模块:用于通过光流配准网络获取所述PET图像和MR图像的光流信息,根据所述光流信息校正MR图像的位置后,分别提取所述PET图像和校正后MR图像的第一特征图和第二特征图;An optical flow registration module: used for acquiring optical flow information of the PET image and the MR image through an optical flow registration network, and extracting a first feature map and a second feature map of the PET image and the corrected MR image respectively after correcting the position of the MR image according to the optical flow information;
特征增强模块:用于通过空间通道特征增强模块对所述第一特征图和第二特征图分别进行空间维度和通道维度的特征增强处理,并将空间维度和通道维度的特征增强处理结果进行融合,生成所述PET图像的第一多模态增强特征图和校正后MR图像的第二多模态增强特征图;A feature enhancement module is used to perform feature enhancement processing of the first feature map and the second feature map in the spatial dimension and the channel dimension respectively through the spatial channel feature enhancement module, and fuse the feature enhancement processing results in the spatial dimension and the channel dimension to generate a first multimodal enhanced feature map of the PET image and a second multimodal enhanced feature map of the corrected MR image;
特征融合模块:用于通过跨模态特征融合模块采用多头交叉注意力机制对所述第一多模态增强特征图和第二多模态增强特征图进行融合后,生成多模态融合特征图,根据所述多模态融合特征图生成PET重建图像。Feature fusion module: used to fuse the first multimodal enhanced feature map and the second multimodal enhanced feature map through a cross-modal feature fusion module using a multi-head cross-attention mechanism to generate a multimodal fusion feature map, and generate a PET reconstructed image based on the multimodal fusion feature map.
本申请实施例采取的又一技术方案为:一种设备,所述设备包括处理器、与所述处理器耦接的存储器,其中,Another technical solution adopted by the embodiment of the present application is: a device, the device includes a processor and a memory coupled to the processor, wherein:
所述存储器存储有用于实现所述基于光流配准的PET成像方法的程序指令; The memory stores program instructions for implementing the PET imaging method based on optical flow registration;
所述处理器用于执行所述存储器存储的所述程序指令以控制基于光流配准的PET成像方法。The processor is used to execute the program instructions stored in the memory to control a PET imaging method based on optical flow registration.
本申请实施例采取的又一技术方案为:一种存储介质,存储有处理器可运行的程序指令,所述程序指令用于执行所述基于光流配准的PET成像方法。Another technical solution adopted by the embodiment of the present application is: a storage medium storing program instructions executable by a processor, wherein the program instructions are used to execute the PET imaging method based on optical flow registration.
相对于现有技术,本申请实施例产生的有益效果在于:本申请实施例的基于光流配准的PET成像方法、装置、设备以及存储介质。Compared with the prior art, the beneficial effects produced by the embodiments of the present application are: the PET imaging method, device, equipment and storage medium based on optical flow registration in the embodiments of the present application.
图1是本申请实施例的基于光流配准的PET成像方法的流程图;FIG1 is a flow chart of a PET imaging method based on optical flow registration according to an embodiment of the present application;
图2为本申请实施例的基于光流配准的PET成像装置结构示意图;FIG2 is a schematic structural diagram of a PET imaging device based on optical flow registration according to an embodiment of the present application;
图3为本申请实施例的设备结构示意图;FIG3 is a schematic diagram of the device structure of an embodiment of the present application;
图4为本申请实施例的存储介质的结构示意图。FIG. 4 is a schematic diagram of the structure of a storage medium according to an embodiment of the present application.
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅是本申请的一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The following will be combined with the drawings in the embodiments of the present application to clearly and completely describe the technical solutions in the embodiments of the present application. Obviously, the described embodiments are only part of the embodiments of the present application, not all of the embodiments. Based on the embodiments in the present application, all other embodiments obtained by ordinary technicians in this field without creative work are within the scope of protection of this application.
本申请中的术语“第一”、“第二”、“第三”仅用于描述目的,而不能理解为指示或暗示相对重要性或者隐含指明所指示的技术特征的数量。由此,限定有“第一”、“第二”、“第三”的特征可以明示或者隐含地包括至少一个该特征。本申请的描述中,“多个”的含义是至少两个,例如两个,三个等,除非另有明确具体的限定。本申请实施例中所有方向性指示(诸如上、下、左、右、前、后……)仅用于解释在某一特定姿态(如附图所示)下各部件之间的相对位置关系、运动情况等,如果该特定姿态发生改变时,则该方向性指示也相应地随之改变。此外,术语“包括”和“具有”以及它们任何变形,意图在于覆盖不 排他的包含。例如包含了一系列步骤或单元的过程、方法、系统、产品或设备没有限定于已列出的步骤或单元,而是可选地还包括没有列出的步骤或单元,或可选地还包括对于这些过程、方法、产品或设备固有的其它步骤或单元。The terms "first", "second" and "third" in this application are only used for descriptive purposes and cannot be understood as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Therefore, the features defined as "first", "second" and "third" may explicitly or implicitly include at least one of the features. In the description of this application, the meaning of "multiple" is at least two, such as two, three, etc., unless otherwise clearly and specifically defined. All directional indications in the embodiments of the present application (such as up, down, left, right, front, back...) are only used to explain the relative position relationship, movement, etc. between the components in a certain specific posture (as shown in the accompanying drawings). If the specific posture changes, the directional indication will also change accordingly. In addition, the terms "including" and "having" and any variations thereof are intended to cover Exclusive inclusion. For example, a process, method, system, product or apparatus comprising a series of steps or units is not limited to the listed steps or units, but may optionally include steps or units not listed, or may optionally include other steps or units inherent to the process, method, product or apparatus.
在本文中提及“实施例”意味着,结合实施例描述的特定特征、结构或特性可以包含在本申请的至少一个实施例中。在说明书中的各个位置出现该短语并不一定均是指相同的实施例,也不是与其它实施例互斥的独立的或备选的实施例。本领域技术人员显式地和隐式地理解的是,本文所描述的实施例可以与其它实施例相结合。Reference to "embodiments" herein means that a particular feature, structure, or characteristic described in conjunction with the embodiments may be included in at least one embodiment of the present application. The appearance of the phrase in various locations in the specification does not necessarily refer to the same embodiment, nor is it an independent or alternative embodiment that is mutually exclusive with other embodiments. It is explicitly and implicitly understood by those skilled in the art that the embodiments described herein may be combined with other embodiments.
请参阅图1,是本申请实施例的基于光流配准的PET成像方法的流程图。本申请实施例的基于光流配准的PET成像方法包括以下步骤:Please refer to FIG1 , which is a flow chart of a PET imaging method based on optical flow registration according to an embodiment of the present application. The PET imaging method based on optical flow registration according to an embodiment of the present application comprises the following steps:
S100:分别采集放射部位的PET图像和MR图像;S100: Acquire PET images and MR images of the irradiated part respectively;
S110:使用Canny算子分别对PET图像和MR图像进行轮廓信息提取;S110: Use the Canny operator to extract contour information from the PET image and the MR image respectively;
本步骤中,轮廓信息提取的下限和上限阈值分别设置为20和50,轮廓信息提取过程具体为:首先,采用形态学图像处理技术对PET图像和MR图像进行图像腐蚀和膨胀处理,消除PET图像和MR图像中的小物体和细线等,同时保留放射部位的基本轮廓形状。在处理过程中,使用一个大小为5×5的圆形核确保各向同性的腐蚀和膨胀过程,以保持轮廓形状的完整性并防止不必要的扭曲,最后生成一个由0和1组成的二进制掩膜,分别用于表示PET图像和MR图像的轮廓信息Ipet和Imr。In this step, the lower and upper thresholds of contour information extraction are set to 20 and 50, respectively. The contour information extraction process is as follows: First, morphological image processing technology is used to perform image erosion and dilation on the PET image and the MR image to eliminate small objects and thin lines in the PET image and the MR image, while retaining the basic contour shape of the radiation site. During the processing, a circular kernel of size 5×5 is used to ensure isotropic erosion and dilation processes to maintain the integrity of the contour shape and prevent unnecessary distortion. Finally, a binary mask consisting of 0 and 1 is generated to represent the contour information I pet and I mr of the PET image and the MR image, respectively.
S120:将PET图像和MR图像的轮廓信息输入光流配准网络,光流配准网络根据PET图像和MR图像的轮廓信息计算光流信息,根据光流信息校正MR图像的位置后,分别提取PET图像和校正后MR图像的第一特征图和第二特征图;S120: inputting the contour information of the PET image and the MR image into the optical flow registration network, the optical flow registration network calculates the optical flow information according to the contour information of the PET image and the MR image, corrects the position of the MR image according to the optical flow information, and extracts the first feature map and the second feature map of the PET image and the corrected MR image respectively;
本步骤中,光流配准网络是一个基于U-net的模型,其以PET图像作为参考,利用反向图像变换校正MR图像的位置,达到PET/MR双模态配准对齐的效果。具体的,光流配准网络包括编码器和解码器两部分,编码器部分使用三个 1x1的卷积层提取特征图,并使用3×3卷积层对提取的特征图进行2倍下采样,从而将提取的特征图尺寸从256×256逐渐减小到32×32。解码器部分使用上采样和1x1卷积层对提取的特征图进行扩张和压缩,以恢复光流图的大小,并最终生成通道数为64的权重掩码,最后,通过应用Sigmoid激活函数生成权重掩码,并输出PET图像和校正后MR图像的第一特征图Fpet和第二特征图Fmr。具体的,光流配准网络参数设置如下表1所示:In this step, the optical flow registration network is a U-net-based model that uses the PET image as a reference and uses the inverse image transformation to correct the position of the MR image to achieve the effect of PET/MR dual-modality registration and alignment. Specifically, the optical flow registration network consists of two parts: an encoder and a decoder. The encoder part uses three The 1x1 convolution layer extracts the feature map, and the 3×3 convolution layer is used to downsample the extracted feature map by 2 times, thereby gradually reducing the size of the extracted feature map from 256×256 to 32×32. The decoder part uses upsampling and 1x1 convolution layers to expand and compress the extracted feature map to restore the size of the optical flow map, and finally generates a weight mask with 64 channels. Finally, the weight mask is generated by applying the Sigmoid activation function, and the first feature map F pet and the second feature map F mr of the PET image and the corrected MR image are output. Specifically, the optical flow registration network parameter settings are shown in Table 1 below:
表1 PWC光流配准网络参数
Table 1 PWC optical flow registration network parameters
编码器参数如下表2所示:The encoder parameters are shown in Table 2 below:
表2编码器参数设置
Table 2 Encoder parameter settings
S130:将第一特征图和第二特征图输入空间通道特征增强模块,通过空间通道特征增强模块对第一特征图和第二特征图分别进行空间维度和通道维度的特征增强处理,分别得到PET图像在空间维度的第一空间增强特征图和在通道维度的第一通道增强特征图以及校正后MR图像在空间维度的第二空间增强特征图和在通道维度的第二通道增强特征图;S130: inputting the first feature map and the second feature map into a spatial channel feature enhancement module, and performing feature enhancement processing on the first feature map and the second feature map in the spatial dimension and the channel dimension respectively by the spatial channel feature enhancement module, so as to obtain a first spatial enhancement feature map in the spatial dimension and a first channel enhancement feature map in the channel dimension of the PET image, and a second spatial enhancement feature map in the spatial dimension and a second channel enhancement feature map in the channel dimension of the corrected MR image respectively;
本步骤中,空间通道特征增强模块(SC-FEM)包括空间增强模块和通道增强模块,空间增强模块可以提高网络对高密度噪声等重要区域的关注度。在 空间增强模块中,通过两个卷积层K1和K2对输入的第一特征图Fpet和第二特征图Fmr进行压缩,得到两个矩阵S1和S2,将矩阵S1和S2在空间维度上进行连接,得到一个单一维度且包含PET和MR图像流的注意力信息的矩阵S,通过对输入特征图进行Sigmoid函数激活并进行逐元素操作,最终分别得到PET图像和校正后MR图像在空间维度的第一空间增强特征图Fs-pet和第二空间增强特征图Fs-mr。空间增强模块参数设置如下表3所示:In this step, the spatial channel feature enhancement module (SC-FEM) includes a spatial enhancement module and a channel enhancement module. The spatial enhancement module can improve the network's attention to important areas such as high-density noise. In the spatial enhancement module, the first feature map F pet and the second feature map F mr of the input are compressed by two convolutional layers K1 and K2 to obtain two matrices S1 and S2. The matrices S1 and S2 are connected in the spatial dimension to obtain a single-dimensional matrix S containing the attention information of the PET and MR image streams. The first spatial enhancement feature map F s-pet and the second spatial enhancement feature map F s-mr of the PET image and the corrected MR image in the spatial dimension are finally obtained by activating the input feature map with a Sigmoid function and performing element-by-element operations. The parameter settings of the spatial enhancement module are shown in Table 3 below:
表3空间增强模块参数设置
Table 3. Parameter settings of spatial enhancement module
而在通道增强模块中,使用全局平均池化将每个H×W的输入特征图压缩为一个1×1的值,并创建通道描述符,通过通道描述符计算PET图像和MR图像的编码向量C1和C2,然后将编码向量C1和C2进行连接,并通过全连接层创建融合的嵌入向量,最后,通过对嵌入向量进行Sigmoid函数激活并进行逐元素操作,分别得到PET图像和校正后MR图像的第一通道增强特征图Fc-pet和第二通道增强特征图Fc-mr。具体的,通道增强模块参数设置如下表4所示:In the channel enhancement module, global average pooling is used to compress each H×W input feature map into a 1×1 value, and a channel descriptor is created. The encoding vectors C1 and C2 of the PET image and MR image are calculated through the channel descriptor, and then the encoding vectors C1 and C2 are connected, and a fused embedding vector is created through a fully connected layer. Finally, the first channel enhanced feature map Fc -pet and the second channel enhanced feature map Fc -mr of the PET image and the corrected MR image are obtained by activating the embedding vector with a Sigmoid function and performing element-by-element operations. Specifically, the channel enhancement module parameter settings are shown in Table 4 below:
表4通道增强模块参数设置
Table 4 Channel enhancement module parameter settings
S140:分别对PET图像的第一空间增强特征图和第一通道增强特征图以及校正后MR图像的第二空间增强特征图和第二通道增强特征图进行融合,分别 生成PET图像的第一多模态增强特征图和校正后MR图像的第二多模态增强特征图;S140: Fusing the first spatial enhancement feature map and the first channel enhancement feature map of the PET image and the second spatial enhancement feature map and the second channel enhancement feature map of the corrected MR image, respectively. generating a first multimodal enhanced feature map of the PET image and a second multimodal enhanced feature map of the corrected MR image;
本步骤中,通过对PET图像的第一空间增强特征图和第一通道增强特征图以及校正后MR图像的第二空间增强特征图和第二通道增强特征图进行融合,生成结合了PET和MR两种模态信息的第一多模态增强特征图Fsc-pet和第二多模态增强特征图Fsc-mr,从而综合考虑了两种模态信息,使特征增强过程更加有效。具体的,第一多模态增强特征图Fsc-pet和第二多模态增强特征图Fsc-mr可以通过以下公式定义:
Fsc-pet=Fpet+Fs-pet+Fc-pet (1)
Fsc-mr=Fmr+Fs-mr+Fc-mr (2)In this step, by fusing the first spatial enhancement feature map and the first channel enhancement feature map of the PET image and the second spatial enhancement feature map and the second channel enhancement feature map of the corrected MR image, a first multimodal enhancement feature map F sc-pet and a second multimodal enhancement feature map F sc-mr combining the two modal information of PET and MR are generated, thereby comprehensively considering the two modal information and making the feature enhancement process more effective. Specifically, the first multimodal enhancement feature map F sc-pet and the second multimodal enhancement feature map F sc-mr can be defined by the following formula:
F sc-pet =F pet +F s-pet +F c-pet (1)
F sc-mr =F mr +F s-mr +F c-mr (2)
S150:将第一多模态增强特征图和第二多模态增强特征图输入跨模态特征融合模块,跨模态特征融合模块采用多头交叉注意力机制对第一多模态增强特征图和第二多模态增强特征图进行融合后,生成多模态融合特征图;S150: inputting the first multimodal enhanced feature map and the second multimodal enhanced feature map into a cross-modal feature fusion module, and the cross-modal feature fusion module fuses the first multimodal enhanced feature map and the second multimodal enhanced feature map using a multi-head cross attention mechanism to generate a multimodal fusion feature map;
本步骤中,跨模态特征融合模块采用多头交叉注意力机制(CM-FFM)和Transformer架构,通过计算PET和MR图像之间的点积来获取查询矩阵Q(queries)、键矩阵K(keys)和值矩阵V(values),从而估计多模态增强特征图之间的关联性,通过缩放和归一化处理,基于softmax函数计算得到跨注意力权重。具体而言,将输入的第一多模态增强特征图Fsc-pet和第二多模态增强特征图Fsc-mr首先进行线性投影操作,生成PET图像和MR图像的查询矩阵Q、键矩阵K和值矩阵V,然后,在两种模态的向量化特征之间进行注意力操作,计算公式如下:
In this step, the cross-modal feature fusion module adopts a multi-head cross-attention mechanism (CM-FFM) and a Transformer architecture to obtain the query matrix Q (queries), key matrix K (keys) and value matrix V (values) by calculating the dot product between the PET and MR images, thereby estimating the correlation between the multimodal enhanced feature maps, and calculating the cross-attention weights based on the softmax function through scaling and normalization. Specifically, the first multimodal enhanced feature map F sc-pet and the second multimodal enhanced feature map F sc-mr are first linearly projected to generate the query matrix Q, key matrix K and value matrix V of the PET image and the MR image, and then the attention operation is performed between the vectorized features of the two modalities. The calculation formula is as follows:
上述公式中,σ表示sigmoid函数,用于将输入映射到0到1的范围内,确保注意力权重在合理的范围内,以表示两个模态之间的关联性,防止权重过于偏离正常范围。表示特征维度的平方根,用于缩放注意力权重,确保在计算点积时,注意力分数不会过大,从而维持模型的稳定性。Kpet,Vpet分别表示PET图像的查询矩阵Q、键矩阵K和值矩阵V,Kpet,Vpet均由第一多模态增强特征图Fsc-pet投影得到,表示由PET图像提取得到的信息。Kmr,Vmr分别表示MR图像的查询矩阵Q、键矩阵K和值矩阵V,Kmr,Vmr均由第二多模态增强特征图Fsc-mr投影得到,表示由MR图像提取得到的信息。In the above formula, σ represents the sigmoid function, which is used to map the input to the range of 0 to 1, ensuring that the attention weight is within a reasonable range to represent the correlation between the two modalities and prevent the weight from deviating too much from the normal range. Represents the square root of the feature dimension and is used to scale the attention weights to ensure that the attention scores are not too large when calculating the dot product, thereby maintaining the stability of the model. K pet , V pet represent the query matrix Q, key matrix K and value matrix V of PET images respectively. K pet and V pet are both obtained by projecting the first multimodal enhanced feature map F sc-pet , and represent information extracted from the PET image. Kmr , Vmr represent the query matrix Q, key matrix K and value matrix V of MR images respectively. K mr and V mr are both obtained by projecting the second multimodal enhanced feature map F sc-mr , and represent the information extracted from the MR image.
将两种模态的注意力操作结果通过平均池化操作进行压缩后,在通道维度上将两种模态的注意力图进行拼接,生成多模态融合特征图,从而为PET图像去噪提供了丰富的有利图像细节。具体的,多头交叉注意力机制参数设置如下表5所示:After compressing the attention operation results of the two modalities through the average pooling operation, the attention maps of the two modalities are spliced in the channel dimension to generate a multi-modal fusion feature map, thereby providing rich and favorable image details for PET image denoising. Specifically, the multi-head cross attention mechanism parameter settings are shown in Table 5 below:
表5多头交叉注意力机制参数设置
Table 5 Multi-head cross attention mechanism parameter settings
S160:通过跳跃连接将多模态融合特征图输入解码器,通过解码器输出去噪后的PET重建图像;S160: input the multimodal fusion feature map into the decoder through the skip connection, and output the denoised PET reconstructed image through the decoder;
本步骤中,在解码器部分,考虑到转置卷积在上采样过程中常常会产生棋盘状伪影,并涉及大量可训练参数,本申请实施例选择双线性上采样方法,缩 放因子为2,并在每个阶段使用卷积层,有助于将多层级特征图恢复到原始图像分辨率,并生成相对平滑的特征图。为了增强模型的泛化能力和改善激活函数的表达能力,本申请实施例在图像的下采样和上采样过程中,在每个卷积层中分别引入批归一化,并在编码和解码端引入跳跃连接,以保留来自两种模态的最有利的图像细节。可以理解,在本申请其他实施例中,还可以利用CT图像作为MR图像的替代对PET图像进行去噪。具体的,解码器参数设置如下表6所示:In this step, in the decoder part, considering that transposed convolution often produces chessboard artifacts during upsampling and involves a large number of trainable parameters, the embodiment of the present application selects a bilinear upsampling method to reduce The magnification factor is 2, and a convolutional layer is used at each stage, which helps to restore the multi-level feature map to the original image resolution and generate a relatively smooth feature map. In order to enhance the generalization ability of the model and improve the expression ability of the activation function, the embodiment of the present application introduces batch normalization in each convolutional layer during the downsampling and upsampling of the image, and introduces jump connections at the encoding and decoding ends to retain the most favorable image details from the two modalities. It can be understood that in other embodiments of the present application, CT images can also be used as a substitute for MR images to denoise PET images. Specifically, the decoder parameter settings are shown in Table 6 below:
表6解码器参数设置
Table 6 Decoder parameter settings
基于上述,本申请实施例的基于光流配准的PET成像方法利用PET图像和MR图像的轮廓信息,通过光流校正网络对PET图像和MR图像进行对齐,可以使得多模态数据更好地对齐,有效减少由于配准问题导致的图像失真和伪影,提高图像的清晰度和准确性;通过引入空间通道特征增强模块,该模块使用矩阵和向量来表示空间注意力和通道注意力,不仅基于当前模态的信息,还结合其他模态的影响进行计算,以确保在每个下采样阶段特征提取的有效性,可以更好地捕捉到图像中的重要特征,提高特征提取的准确性和鲁棒性。通过采用多头交叉注意力机制捕捉跨模态之间特征信息的多样性,能够更充分地利用多模态数据之间的相关性和互补性,可以有效地融合PET和MR的特征,得到更 加清晰的PET重建图像。Based on the above, the PET imaging method based on optical flow registration in the embodiment of the present application utilizes the contour information of the PET image and the MR image, and aligns the PET image and the MR image through the optical flow correction network, so that the multimodal data can be better aligned, effectively reducing the image distortion and artifacts caused by the registration problem, and improving the clarity and accuracy of the image; by introducing the spatial channel feature enhancement module, the module uses matrices and vectors to represent spatial attention and channel attention, which is calculated not only based on the information of the current modality, but also combined with the influence of other modalities to ensure the effectiveness of feature extraction at each downsampling stage, which can better capture the important features in the image and improve the accuracy and robustness of feature extraction. By adopting a multi-head cross-attention mechanism to capture the diversity of feature information across modalities, it is possible to more fully utilize the correlation and complementarity between multimodal data, and can effectively fuse the features of PET and MR to obtain a better Add clear PET reconstruction images.
请参阅图2,为本申请实施例的基于光流配准的PET成像装置结构示意图。本申请实施例的基于光流配准的PET成像装置40包括:Please refer to FIG2 , which is a schematic diagram of the structure of a PET imaging device based on optical flow registration according to an embodiment of the present application. The PET imaging device 40 based on optical flow registration according to an embodiment of the present application comprises:
图像采集模块41:用于分别采集放射部位的PET图像和MR图像;Image acquisition module 41: used to respectively acquire PET images and MR images of the radiation site;
光流配准模块42:用于通过光流配准网络获取所述PET图像和MR图像的光流信息,根据所述光流信息校正MR图像的位置后,分别提取所述PET图像和校正后MR图像的第一特征图和第二特征图;An optical flow registration module 42 is used to obtain optical flow information of the PET image and the MR image through an optical flow registration network, and after correcting the position of the MR image according to the optical flow information, extract a first feature map and a second feature map of the PET image and the corrected MR image respectively;
特征增强模块43:用于通过空间通道特征增强模块对所述第一特征图和第二特征图分别进行空间维度和通道维度的特征增强处理,并将空间维度和通道维度的特征增强处理结果进行融合,生成所述PET图像的第一多模态增强特征图和校正后MR图像的第二多模态增强特征图;The feature enhancement module 43 is used to perform feature enhancement processing of the first feature map and the second feature map in the spatial dimension and the channel dimension respectively through the spatial channel feature enhancement module, and fuse the feature enhancement processing results in the spatial dimension and the channel dimension to generate a first multimodal enhanced feature map of the PET image and a second multimodal enhanced feature map of the corrected MR image;
特征融合模块44:用于通过跨模态特征融合模块采用多头交叉注意力机制对所述第一多模态增强特征图和第二多模态增强特征图进行融合后,生成多模态融合特征图,根据所述多模态融合特征图生成PET重建图像。Feature fusion module 44: used to fuse the first multimodal enhanced feature map and the second multimodal enhanced feature map through a cross-modal feature fusion module using a multi-head cross-attention mechanism to generate a multimodal fusion feature map, and generate a PET reconstructed image according to the multimodal fusion feature map.
请参阅图3,为本申请实施例的设备结构示意图。该设备50包括:Please refer to FIG3 , which is a schematic diagram of the device structure of an embodiment of the present application. The device 50 includes:
存储有可执行程序指令的存储器51;A memory 51 storing executable program instructions;
与存储器51连接的处理器52;A processor 52 connected to the memory 51;
处理器52用于调用存储器51中存储的可执行程序指令并执行以下步骤:分别采集放射部位的PET图像和MR图像;通过光流配准网络获取所述PET图像和MR图像的光流信息,根据所述光流信息校正MR图像的位置后,分别提取所述PET图像和校正后MR图像的第一特征图和第二特征图;通过空间通道特征增强模块对所述第一特征图和第二特征图分别进行空间维度和通道维度的特征增强处理,并将空间维度和通道维度的特征增强处理结果进行融合,生成 所述PET图像的第一多模态增强特征图和校正后MR图像的第二多模态增强特征图;通过跨模态特征融合模块采用多头交叉注意力机制对所述第一多模态增强特征图和第二多模态增强特征图进行融合后,生成多模态融合特征图,根据所述多模态融合特征图生成PET重建图像。The processor 52 is used to call the executable program instructions stored in the memory 51 and perform the following steps: respectively collect PET images and MR images of the radiation site; obtain the optical flow information of the PET image and the MR image through the optical flow registration network, correct the position of the MR image according to the optical flow information, and respectively extract the first feature map and the second feature map of the PET image and the corrected MR image; respectively perform feature enhancement processing of the spatial dimension and the channel dimension on the first feature map and the second feature map through the spatial channel feature enhancement module, and fuse the feature enhancement processing results of the spatial dimension and the channel dimension to generate The first multimodal enhanced feature map of the PET image and the second multimodal enhanced feature map of the corrected MR image; after fusing the first multimodal enhanced feature map and the second multimodal enhanced feature map through a cross-modal feature fusion module using a multi-head cross-attention mechanism, a multimodal fusion feature map is generated, and a PET reconstructed image is generated according to the multimodal fusion feature map.
其中,处理器52还可以称为CPU(Central Processing Unit,中央处理单元)。处理器52可能是一种集成电路芯片,具有信号的处理能力。处理器52还可以是通用处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、现成可编程门阵列(FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The processor 52 may also be referred to as a CPU (Central Processing Unit). The processor 52 may be an integrated circuit chip having signal processing capabilities. The processor 52 may also be a general-purpose processor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), a field-programmable gate array (FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components. The general-purpose processor may be a microprocessor or the processor may also be any conventional processor, etc.
请参阅图4,图4为本申请实施例的存储介质的结构示意图。本申请实施例的存储介质存储有能够实现以下步骤的程序指令61:分别采集放射部位的PET图像和MR图像;通过光流配准网络获取所述PET图像和MR图像的光流信息,根据所述光流信息校正MR图像的位置后,分别提取所述PET图像和校正后MR图像的第一特征图和第二特征图;通过空间通道特征增强模块对所述第一特征图和第二特征图分别进行空间维度和通道维度的特征增强处理,并将空间维度和通道维度的特征增强处理结果进行融合,生成所述PET图像的第一多模态增强特征图和校正后MR图像的第二多模态增强特征图;通过跨模态特征融合模块采用多头交叉注意力机制对所述第一多模态增强特征图和第二多模态增强特征图进行融合后,生成多模态融合特征图,根据所述多模态融合特征图生成PET重建图像。其中,该程序指令61可以以软件产品的形式存储在上述存储介质中,包括若干指令用以使得一台设备(可以是个人计算机,服务器,或者网络设备等)或处理器(processor)执行本申请各个实施方式方法的全 部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序指令的介质,或者是计算机、服务器、手机、平板等终端设备。其中,服务器可以是独立的服务器,也可以是提供云服务、云数据库、云计算、云函数、云存储、网络服务、云通信、中间件服务、域名服务、安全服务、内容分发网络(Content Delivery Network,CDN)、以及大数据和人工智能平台等基础云计算服务的云服务器。Please refer to Figure 4, which is a schematic diagram of the structure of the storage medium of the embodiment of the present application. The storage medium of the embodiment of the present application stores a program instruction 61 that can implement the following steps: respectively collect PET images and MR images of the radiation site; obtain the optical flow information of the PET image and the MR image through the optical flow registration network, and extract the first feature map and the second feature map of the PET image and the corrected MR image respectively after correcting the position of the MR image according to the optical flow information; perform feature enhancement processing of the spatial dimension and the channel dimension on the first feature map and the second feature map respectively through the spatial channel feature enhancement module, and fuse the feature enhancement processing results of the spatial dimension and the channel dimension to generate the first multimodal enhancement feature map of the PET image and the second multimodal enhancement feature map of the corrected MR image; after fusing the first multimodal enhancement feature map and the second multimodal enhancement feature map through the cross-modal feature fusion module using a multi-head cross attention mechanism, a multimodal fusion feature map is generated, and a PET reconstructed image is generated according to the multimodal fusion feature map. The program instructions 61 may be stored in the above storage medium in the form of a software product, including several instructions for enabling a device (which may be a personal computer, a server, or a network device, etc.) or a processor to execute the entire method of each embodiment of the present application. The aforementioned storage medium includes: U disk, mobile hard disk, read-only memory (ROM), random access memory (RAM), disk or optical disk and other media that can store program instructions, or terminal devices such as computers, servers, mobile phones, tablets, etc. Among them, the server can be an independent server, or it can be a cloud server that provides basic cloud computing services such as cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communications, middleware services, domain name services, security services, content delivery networks (CDN), and big data and artificial intelligence platforms.
在本申请所提供的几个实施例中,应该理解到,所揭露的系统,装置和方法,可以通过其它的方式实现。例如,以上所描述的系统实施例仅仅是示意性的,例如,单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided in the present application, it should be understood that the disclosed systems, devices and methods can be implemented in other ways. For example, the system embodiments described above are only schematic. For example, the division of units is only a logical function division. There may be other division methods in actual implementation, such as multiple units or components can be combined or integrated into another system, or some features can be ignored or not executed. Another point is that the mutual coupling or direct coupling or communication connection shown or discussed can be an indirect coupling or communication connection through some interfaces, devices or units, which can be electrical, mechanical or other forms.
另外,在本申请各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。以上仅为本申请的实施方式,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。 In addition, each functional unit in each embodiment of the present application can be integrated into one processing unit, or each unit can exist physically separately, or two or more units can be integrated into one unit. The above integrated unit can be implemented in the form of hardware or in the form of software functional units. The above is only an implementation method of the present application, and does not limit the patent scope of the present application. Any equivalent structure or equivalent process transformation made by using the description and drawings of this application, or directly or indirectly used in other related technical fields, is also included in the patent protection scope of the present application.
Claims (10)
The cross-modal feature fusion module adopts a multi-head cross attention mechanism and a Transformer architecture to perform a linear projection operation on the first multimodal enhanced feature map F sc-pet and the second multimodal enhanced feature map F sc-mr to generate a query matrix Q, a key matrix K and a value matrix V for PET images and MR images, respectively, and perform an attention operation between the vectorized features of the two modalities:
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/CN2023/113053 WO2025035380A1 (en) | 2023-08-15 | 2023-08-15 | Pet imaging method and apparatus based on optical flow registration, and device and storage medium |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| PCT/CN2023/113053 WO2025035380A1 (en) | 2023-08-15 | 2023-08-15 | Pet imaging method and apparatus based on optical flow registration, and device and storage medium |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2025035380A1 true WO2025035380A1 (en) | 2025-02-20 |
Family
ID=94631915
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2023/113053 Pending WO2025035380A1 (en) | 2023-08-15 | 2023-08-15 | Pet imaging method and apparatus based on optical flow registration, and device and storage medium |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2025035380A1 (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119941813A (en) * | 2025-04-07 | 2025-05-06 | 西安交通大学医学院第一附属医院 | Multimodal large model-assisted laparoscopic soft tissue registration surgery navigation method and system |
| CN120254581A (en) * | 2025-04-02 | 2025-07-04 | 北京国力电气科技有限公司 | A method for monitoring the switch position status of a circuit breaker based on video recognition technology |
| CN120669192A (en) * | 2025-06-09 | 2025-09-19 | 中国人民解放军军事航天部队航天工程大学 | High-precision radio environment map reconstruction method based on multi-mode fusion |
| CN120726104A (en) * | 2025-08-19 | 2025-09-30 | 杭州电子科技大学 | A two-stage infrared and visible light image registration method, system and device |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101744618A (en) * | 2009-12-17 | 2010-06-23 | 北京亿仁赛博医疗设备有限公司 | One-equipment room transmission PET/CT/MR image collection, registering and imaging system and method |
| CN103854270A (en) * | 2012-11-28 | 2014-06-11 | 广州医学院第一附属医院 | CT and MR inter-machine three dimensional image fusion registration method and system |
| CN107644421A (en) * | 2016-07-20 | 2018-01-30 | 上海联影医疗科技有限公司 | Medical image cutting method and system |
| US20210150782A1 (en) * | 2019-11-15 | 2021-05-20 | The Board Of Trustees Of The Leland Stanford Junior University | Feature space based MR guided PET Reconstruction |
| CN114119546A (en) * | 2021-11-25 | 2022-03-01 | 推想医疗科技股份有限公司 | Method and device for detecting MRI images |
| CN115100306A (en) * | 2022-05-19 | 2022-09-23 | 浙江大学 | Four-dimensional cone-beam CT imaging method and device for pancreatic region |
-
2023
- 2023-08-15 WO PCT/CN2023/113053 patent/WO2025035380A1/en active Pending
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN101744618A (en) * | 2009-12-17 | 2010-06-23 | 北京亿仁赛博医疗设备有限公司 | One-equipment room transmission PET/CT/MR image collection, registering and imaging system and method |
| CN103854270A (en) * | 2012-11-28 | 2014-06-11 | 广州医学院第一附属医院 | CT and MR inter-machine three dimensional image fusion registration method and system |
| CN107644421A (en) * | 2016-07-20 | 2018-01-30 | 上海联影医疗科技有限公司 | Medical image cutting method and system |
| US20210150782A1 (en) * | 2019-11-15 | 2021-05-20 | The Board Of Trustees Of The Leland Stanford Junior University | Feature space based MR guided PET Reconstruction |
| CN114119546A (en) * | 2021-11-25 | 2022-03-01 | 推想医疗科技股份有限公司 | Method and device for detecting MRI images |
| CN115100306A (en) * | 2022-05-19 | 2022-09-23 | 浙江大学 | Four-dimensional cone-beam CT imaging method and device for pancreatic region |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN120254581A (en) * | 2025-04-02 | 2025-07-04 | 北京国力电气科技有限公司 | A method for monitoring the switch position status of a circuit breaker based on video recognition technology |
| CN119941813A (en) * | 2025-04-07 | 2025-05-06 | 西安交通大学医学院第一附属医院 | Multimodal large model-assisted laparoscopic soft tissue registration surgery navigation method and system |
| CN120669192A (en) * | 2025-06-09 | 2025-09-19 | 中国人民解放军军事航天部队航天工程大学 | High-precision radio environment map reconstruction method based on multi-mode fusion |
| CN120726104A (en) * | 2025-08-19 | 2025-09-30 | 杭州电子科技大学 | A two-stage infrared and visible light image registration method, system and device |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12118649B2 (en) | Deep learning based three-dimensional reconstruction method for low-dose PET imaging | |
| WO2025035380A1 (en) | Pet imaging method and apparatus based on optical flow registration, and device and storage medium | |
| WO2024066049A1 (en) | Pet image denoising method, terminal device, and readable storage medium | |
| CN111709897B (en) | Domain transformation-based positron emission tomography image reconstruction method | |
| Xiao et al. | STIR-Net: deep spatial-temporal image restoration net for radiation reduction in CT perfusion | |
| CN112488953A (en) | Medical image denoising method, system, terminal and storage medium | |
| Pan et al. | Ultra high speed SPECT bone imaging enabled by a deep learning enhancement method: a proof of concept | |
| CN114862881A (en) | A cross-modal attention tumor segmentation method and system based on PET-CT | |
| WO2021168920A1 (en) | Low-dose image enhancement method and system based on multiple dose levels, and computer device, and storage medium | |
| CN109215093B (en) | Low-dose PET image reconstruction method, device, equipment and storage medium | |
| CN115272507A (en) | A low-dose SPECT image reconstruction method based on conditional generative adversarial network model | |
| WO2022094911A1 (en) | Weight-sharing double-region generative adversarial network and image generation method therefor | |
| CN117036525A (en) | PET imaging method, device, equipment and storage medium based on optical flow registration | |
| CN111243052A (en) | Image reconstruction method and device, computer equipment and storage medium | |
| CN112634147B (en) | PET image noise reduction method, system, device and medium for self-supervision learning | |
| CN113053496B (en) | A Deep Learning Method for Low Dose Estimation in Medical Images | |
| CN118644570A (en) | A reconstruction method for ultra-sparse angular CT images | |
| Liang et al. | High quality low-dose SPECT reconstruction using CGAN-based transformer network with geometric tight framelet | |
| CN113436708B (en) | A method for generating delayed CT images based on deep learning algorithm | |
| Liang et al. | Tc-gan: a transformer-based conditional generative adversarial network for low-dose spect image reconstruction | |
| CN112652029A (en) | PET imaging method, device and equipment | |
| WO2022193276A1 (en) | Deep learning method for low dose estimation of medical image | |
| Zhuang et al. | Artifact removal for unpaired thorax CBCT images using a feature fusion residual network and contextual loss | |
| CN118411443B (en) | CT image generation method, system, device, and medium based on DR image | |
| CN120672890B (en) | A hybrid imaging method for four-dimensional computed tomography |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 23948813 Country of ref document: EP Kind code of ref document: A1 |