CN116797852A - Ultrasound image recognition model and ultrasound image recognition method - Google Patents
Ultrasound image recognition model and ultrasound image recognition method Download PDFInfo
- Publication number
- CN116797852A CN116797852A CN202310950122.6A CN202310950122A CN116797852A CN 116797852 A CN116797852 A CN 116797852A CN 202310950122 A CN202310950122 A CN 202310950122A CN 116797852 A CN116797852 A CN 116797852A
- Authority
- CN
- China
- Prior art keywords
- image
- channel
- module
- feature
- ultrasound image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Software Systems (AREA)
- General Physics & Mathematics (AREA)
- Computing Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Multimedia (AREA)
- Medical Informatics (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Biophysics (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
本申请适用于超声图像识别技术领域,提供了一种超声图像识别模型及超声图像识别方法,超声图像识别模型包括残差模块、类金字塔分解模块、概率池化模块、通道注意力模块、第一特征图像调节模块、第二特征图像调节模块以及图像识别模块;残差模块接收超声图像,输出端分别连接类金字塔分解模块、第一特征图像调节模块、第二特征图像调节模块,类金字塔分解模块连接概率池化模块,概率池化模块连接通道注意力模块,通道注意力模块连接第一特征图像调节模块,第一特征图像调节模块连接第二特征图像调节模块,第二特征图像调节模块连接图像识别模块,图像识别模块输出超声图像的识别结果。本申请能够提高超声图像识别的准确性。
This application is applicable to the technical field of ultrasonic image recognition, and provides an ultrasonic image recognition model and an ultrasonic image recognition method. The ultrasonic image recognition model includes a residual module, a pyramid-like decomposition module, a probability pooling module, a channel attention module, and a first Feature image adjustment module, second feature image adjustment module and image recognition module; the residual module receives the ultrasound image, and the output terminals are respectively connected to the pyramid-like decomposition module, the first feature image adjustment module, the second feature image adjustment module, and the pyramid-like decomposition module The probability pooling module is connected, the probability pooling module is connected to the channel attention module, the channel attention module is connected to the first feature image adjustment module, the first feature image adjustment module is connected to the second feature image adjustment module, and the second feature image adjustment module is connected to the image Recognition module, the image recognition module outputs the recognition result of the ultrasound image. This application can improve the accuracy of ultrasound image recognition.
Description
技术领域Technical field
本申请属于超声图像识别技术领域,尤其涉及一种超声图像识别模型及超声图像识别方法。The present application belongs to the technical field of ultrasonic image recognition, and in particular relates to an ultrasonic image recognition model and an ultrasonic image recognition method.
背景技术Background technique
超声检测通过接收一定频率声波在介质密度具有差异的分界面上的回波并成像,从而实现对物体的内部状况分析。由于超声无损检测具有动态检测、高灵敏性、高可控性以及经济性好等优势,长久以来在工业、医学以及土木等众多领域都具有不可替代的位置。然而,超声图像是由固定中心频率的回波采样、转化、赋值形成,不仅通道数量少而且噪声扰动多,导致有效信息量低。另外,由于超声图像的判识在各领域都需要专业人员来完成,不仅依靠个人经验,而且准确率、效率均难以满足社会发展的需要求。Ultrasonic testing achieves analysis of the internal condition of an object by receiving and imaging the echoes of sound waves of a certain frequency on interfaces with differences in medium density. Because ultrasonic non-destructive testing has the advantages of dynamic detection, high sensitivity, high controllability and good economy, it has long had an irreplaceable position in many fields such as industry, medicine, and civil engineering. However, ultrasound images are formed by echo sampling, conversion, and assignment of fixed center frequencies. Not only are there a small number of channels, but there are also many noise disturbances, resulting in a low amount of effective information. In addition, because the identification of ultrasound images requires professionals in various fields, it not only relies on personal experience, but also the accuracy and efficiency cannot meet the needs of social development.
近年来,以神经网络为代表的人工智能技术不断取得突破,正进入高速发展时期,在各行业应用也逐步深入。因此,在超声图像的识别分类任务中引入人工智能方法具有重要意义,将帮助专业人员决策,并缓解大量相难题。但是,由于超声图像成像过程与常规的光成像的差异化过大,直接使用卷积神经网络难以在超声图像中提取足到够特征,效果并不理想。因此,现有的超声图像识别方法的识别准确性有待提高。In recent years, artificial intelligence technology represented by neural networks has continued to make breakthroughs and is entering a period of rapid development, with its applications in various industries gradually deepening. Therefore, it is of great significance to introduce artificial intelligence methods in the recognition and classification tasks of ultrasound images, which will help professionals make decisions and alleviate a large number of phase problems. However, due to the large difference between the ultrasound image imaging process and conventional light imaging, it is difficult to directly use convolutional neural networks to extract sufficient features from ultrasound images, and the effect is not ideal. Therefore, the recognition accuracy of existing ultrasound image recognition methods needs to be improved.
发明内容Contents of the invention
本申请提供了一种超声图像识别模型及超声图像识别方法,可以解决目前超声图像识别方法的识别准确性低的问题。This application provides an ultrasound image recognition model and an ultrasound image recognition method, which can solve the problem of low recognition accuracy of current ultrasound image recognition methods.
第一方面,本申请提供了一种超声图像识别模型,包括残差模块、类金字塔分解模块、概率池化模块、通道注意力模块、第一特征图像调节模块、第二特征图像调节模块以及图像识别模块;In the first aspect, this application provides an ultrasound image recognition model, including a residual module, a pyramid-like decomposition module, a probability pooling module, a channel attention module, a first feature image adjustment module, a second feature image adjustment module, and an image identification module;
残差模块,用于对输入超声图像进行卷积处理,得到输入超声图像对应的初始多通道特征图像;The residual module is used to perform convolution processing on the input ultrasound image to obtain the initial multi-channel feature image corresponding to the input ultrasound image;
类金字塔分解模块,用于对初始多通道特征图像中每个通道对应的特征图像进行金字塔分解,得到每个通道对应的多个特征子图像;多个特征子图像的尺度互不相同;The pyramid-like decomposition module is used to perform pyramid decomposition on the feature image corresponding to each channel in the initial multi-channel feature image, and obtain multiple feature sub-images corresponding to each channel; the scales of the multiple feature sub-images are different from each other;
概率池化模块,用于分别对多个特征子图像中的每个特征子图像进行概率池化,得到每个特征子图像的初始权重系数;概率池化表示不同特征子图像对应的池化方式的概率互不相同,池化方式包括最大池化、平均池化、最小池化以及随机池化,概率服从标准正态分布;The probability pooling module is used to perform probability pooling on each feature sub-image in multiple feature sub-images to obtain the initial weight coefficient of each feature sub-image; probability pooling represents the pooling method corresponding to different feature sub-images. The probabilities are different from each other. The pooling methods include maximum pooling, average pooling, minimum pooling and random pooling. The probability obeys the standard normal distribution;
通道注意力模块,用于根据初始权重系数计算每个通道的通道权重,并对所有通道的通道权重进行激活,得到初始多通道特征图像的通道权重矩阵;The channel attention module is used to calculate the channel weight of each channel based on the initial weight coefficient, and activate the channel weights of all channels to obtain the channel weight matrix of the initial multi-channel feature image;
第一特征图像调节模块,用于将通道权重矩阵和初始多通道特征图像相乘,得到中间多通道特征图像;The first feature image adjustment module is used to multiply the channel weight matrix and the initial multi-channel feature image to obtain the intermediate multi-channel feature image;
第二特征图像调节模块,用于将中间多通道特征图像和初始多通道特征图像相加,得到最终多通道特征图像;The second feature image adjustment module is used to add the intermediate multi-channel feature image and the initial multi-channel feature image to obtain the final multi-channel feature image;
图像识别模块,用于对最终多通道特征图像进行识别,得到最终多通道特征图像的识别结果。The image recognition module is used to recognize the final multi-channel feature image and obtain the recognition result of the final multi-channel feature image.
可选的,残差模块的输入端接收超声图像,残差模块的输出端分别连接类金字塔分解模块的输入端、第一特征图像调节模块的第一输入端以及第二特征图像调节模块的第一输入端,类金字塔分解模块的输出端连接概率池化模块的输入端,概率池化模块的输出端连接通道注意力模块的输入端,通道注意力模块的输出端连接第一特征图像调节模块的第二输入端,第一特征图像调节模块的输出端连接第二特征图像调节模块的第二输入端,第二特征图像调节模块的输出端连接图像识别模块的输入端,图像识别模块的输出端输出超声图像的识别结果。Optionally, the input end of the residual module receives the ultrasound image, and the output end of the residual module is respectively connected to the input end of the pyramid-like decomposition module, the first input end of the first feature image adjustment module, and the third input end of the second feature image adjustment module. One input terminal, the output terminal of the pyramid-like decomposition module is connected to the input terminal of the probability pooling module, the output terminal of the probability pooling module is connected to the input terminal of the channel attention module, and the output terminal of the channel attention module is connected to the first feature image adjustment module The second input end of the first feature image adjustment module is connected to the second input end of the second feature image adjustment module. The output end of the second feature image adjustment module is connected to the input end of the image recognition module. The output of the image recognition module The terminal outputs the recognition results of the ultrasound image.
可选的,对初始多通道特征图像中每个通道对应的特征图像进行金字塔分解,得到每个通道对应的多个特征子图像,包括:Optionally, perform pyramid decomposition on the feature image corresponding to each channel in the initial multi-channel feature image to obtain multiple feature sub-images corresponding to each channel, including:
分别针对每个通道对应的特征图像,执行以下步骤:For the feature image corresponding to each channel, perform the following steps:
步骤i,将特征图像作为初始图像,对初始图像进行复制,得到中间图像;Step i, use the feature image as the initial image, copy the initial image, and obtain the intermediate image;
步骤ii,以预先设置的步长,对中间图像的最外圈像素进行截取,得到最终图像;Step ii: intercept the outermost pixels of the intermediate image with a preset step size to obtain the final image;
步骤iii,若最终图像的尺度大于等于预设尺度阈值,则将最终图像作为步骤i中的初始图像,返回执行步骤i;否则,将初始图像和执行步骤iii得到的所有最终图像,作为每个通道对应的多个特征子图像。Step iii, if the scale of the final image is greater than or equal to the preset scale threshold, use the final image as the initial image in step i, and return to step i; otherwise, use the initial image and all final images obtained by executing step iii as each Multiple feature sub-images corresponding to the channel.
可选的,尺度阈值的计算公式为Optional, the calculation formula of scale threshold is:
E=(min(M,N)-2)mod2*TE=(min(M,N)-2)mod2*T
其中,Mx,Nx均表示尺度阈值,E表示尺度阈值计算的判断条件,T表示步长,M,N均表示初始图像的尺度,mod表示取余,|·|表示取绝对值。Among them , M
可选的,根据初始权重系数计算每个通道的通道权重,并对所有通道的通道权重进行一维卷积处理,得到初始多通道特征图像的通道权重矩阵,包括:Optionally, calculate the channel weight of each channel based on the initial weight coefficient, and perform one-dimensional convolution processing on the channel weights of all channels to obtain the channel weight matrix of the initial multi-channel feature image, including:
分别针对每个通道,将通道对应的所有初始权重系数相加,得到通道权重;For each channel, add all the initial weight coefficients corresponding to the channel to obtain the channel weight;
利用一维卷积对所有通道的通道权重进行卷积处理,得到中间通道权重矩阵;其中,若通道的总数目小于等于128,则将一维卷积的卷积核设为5,否则将一维卷积的卷积核设为7;Use one-dimensional convolution to convolve the channel weights of all channels to obtain the intermediate channel weight matrix; among them, if the total number of channels is less than or equal to 128, set the convolution kernel of the one-dimensional convolution to 5, otherwise set one The convolution kernel of dimensional convolution is set to 7;
利用激活函数对中间通道权重矩阵进行激活,得到通道权重矩阵。Use the activation function to activate the intermediate channel weight matrix to obtain the channel weight matrix.
第二方面,本申请提供了一种超声图像识别方法,包括:In the second aspect, this application provides an ultrasound image recognition method, including:
获取超声图像训练集;超声图像训练集包括多个已知类型的超声图像;Obtain an ultrasound image training set; the ultrasound image training set includes multiple ultrasound images of known types;
将多个已知类型的超声图像逐个输入超声图像识别模型,得到每个超声图像的识别结果;超声图像识别模型为上述的超声图像识别模型;Input multiple known types of ultrasound images into the ultrasound image recognition model one by one to obtain the recognition results of each ultrasound image; the ultrasound image recognition model is the above-mentioned ultrasound image recognition model;
根据识别结果和每个超声图像的类型,计算超声图像识别模型的损失,并根据损失对超声图像识别模型进行反向传播,直至超声图像识别模型拟合,得到训练后的超声图像识别模型;According to the recognition results and the type of each ultrasound image, the loss of the ultrasound image recognition model is calculated, and the ultrasound image recognition model is backpropagated according to the loss until the ultrasound image recognition model is fitted, and the trained ultrasound image recognition model is obtained;
将待识别超声图像输入训练后的超声图像识别模型,得到待识别超声图像的识别结果。The ultrasound image to be recognized is input into the trained ultrasound image recognition model to obtain the recognition result of the ultrasound image to be recognized.
可选的,将多个已知类型的超声图像逐个输入超声图像识别模型,得到每个超声图像的识别结果,包括:Optionally, input multiple ultrasound images of known types into the ultrasound image recognition model one by one to obtain the recognition results of each ultrasound image, including:
通过计算公式By calculation formula
得到每个类型的输出概率Pk;其中,Pk表示类型k的输出概率,k=1,2,...,Q,Q表示超声图像训练集中超声图像的类型的总数量,zk表示第k个节点的输出值,节点的数量和类型的数量一一对应,,j表示第j个节点,n表示输出节点的个数,softmax(·)表示激活函数;Obtain the output probability P k of each type; where, P k represents the output probability of type k, k = 1, 2,..., Q, Q represents the total number of types of ultrasound images in the ultrasound image training set, and z k represents The output value of the k-th node has a one-to-one correspondence between the number of nodes and the number of types, j represents the j-th node, n represents the number of output nodes, and softmax(·) represents the activation function;
将输出概率最大的类型作为识别结果。The type with the highest output probability is used as the recognition result.
可选的,根据识别结果和每个超声图像的类型,计算超声图像识别模型的损失,并根据损失对超声图像识别模型进行反向传播,直至超声图像识别模型拟合,得到训练后的超声图像识别模型,包括:Optionally, calculate the loss of the ultrasound image recognition model based on the recognition result and the type of each ultrasound image, and perform backpropagation on the ultrasound image recognition model based on the loss until the ultrasound image recognition model is fitted to obtain the trained ultrasound image. Recognition models, including:
分别针对每个超声图像,通过计算公式For each ultrasound image, calculate the formula
FL=-ak(1-pk)γlog(pk)FL=-a k (1-p k ) γ log(p k )
得到超声图像识别模型的损失FL;其中,ak表示类型k的权重,pk表示类型k的输出概率,γ表示聚焦参数,Y∈[0,5];Obtain the loss FL of the ultrasound image recognition model; where a k represents the weight of type k, p k represents the output probability of type k, γ represents the focus parameter, Y∈[0,5];
根据损失EL对超声图像识别模型进行反向传播,直至反向传播得到的新超声图像识别模型的损失值小于预设损失阈值时,确定新超声图像识别模型拟合,得到训练后的超声图像识别模型;其中,反向传播的动量设置为0.9,反向传播的学习率采用热学习率策略获取,反向传播的L1正则化的权重衰减系数设置为5e-4。The ultrasound image recognition model is backpropagated according to the loss EL, until the loss value of the new ultrasound image recognition model obtained by backpropagation is less than the preset loss threshold, the fitting of the new ultrasound image recognition model is determined, and the trained ultrasound image recognition model is obtained. Model; among them, the momentum of back propagation is set to 0.9, the learning rate of back propagation is obtained using the thermal learning rate strategy, and the weight attenuation coefficient of L1 regularization of back propagation is set to 5e-4.
本申请的上述方案有如下的有益效果:The above solution of this application has the following beneficial effects:
本申请提供的超声图像识别模型利用类金字塔分解模块对每个通道对应的特征图像进行金字塔分解,能够使后续提取的特征更加精确,从而提高超声图像识别的准确性,利用概率池化模块对每个特征子图像进行概率池化,可以防止特征图像池化时过拟合,通过多种池化方式,能够选择性的提取重要的局部特征,基于这些重要的局部特征来对特征图像进行识别,能够显著提高超声图像识别的准确性。The ultrasonic image recognition model provided by this application uses a pyramid-like decomposition module to perform pyramid decomposition on the feature image corresponding to each channel, which can make the subsequent extracted features more accurate, thereby improving the accuracy of ultrasonic image recognition. It uses a probability pooling module to decompose each channel. Probabilistic pooling of feature sub-images can prevent over-fitting when pooling feature images. Through a variety of pooling methods, important local features can be selectively extracted, and feature images can be identified based on these important local features. It can significantly improve the accuracy of ultrasound image recognition.
本申请的其它有益效果将在随后的具体实施方式部分予以详细说明。Other beneficial effects of the present application will be described in detail in the subsequent specific embodiments section.
附图说明Description of the drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the embodiments or description of the prior art will be briefly introduced below. Obviously, the drawings in the following description are only for the purpose of the present application. For some embodiments, for those of ordinary skill in the art, other drawings can be obtained based on these drawings without exerting creative efforts.
图1为本申请一实施例提供的超声图像识别模型的结构示意图;Figure 1 is a schematic structural diagram of an ultrasound image recognition model provided by an embodiment of the present application;
图2为本申请一实施例提供的残差模块的结构示意图;Figure 2 is a schematic structural diagram of a residual module provided by an embodiment of the present application;
图3为本申请一实施例提供的特征图像金字的结构示意图;Figure 3 is a schematic structural diagram of a feature image character provided by an embodiment of the present application;
图4为本申请一实施例提供的超声图像识别的流程图。Figure 4 is a flow chart of ultrasound image recognition provided by an embodiment of the present application.
具体实施方式Detailed ways
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请实施例。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本申请。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本申请的描述。In the following description, for the purpose of explanation rather than limitation, specific details such as specific system structures and technologies are provided to provide a thorough understanding of the embodiments of the present application. However, it will be apparent to those skilled in the art that the present application may be practiced in other embodiments without these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present application with unnecessary detail.
应当理解,当在本申请说明书和所附权利要求书中使用时,术语“包括”指示所描述特征、整体、步骤、操作、元素和/或组件的存在,但并不排除一个或多个其它特征、整体、步骤、操作、元素、组件和/或其集合的存在或添加。It will be understood that, when used in this specification and the appended claims, the term "comprising" indicates the presence of the described features, integers, steps, operations, elements and/or components but does not exclude one or more other The presence or addition of features, integers, steps, operations, elements, components and/or collections thereof.
还应当理解,在本申请说明书和所附权利要求书中使用的术语“和/或”是指相关联列出的项中的一个或多个的任何组合以及所有可能组合,并且包括这些组合。It will also be understood that the term "and/or" as used in this specification and the appended claims refers to and includes any and all possible combinations of one or more of the associated listed items.
如在本申请说明书和所附权利要求书中所使用的那样,术语“如果”可以依据上下文被解释为“当...时”或“一旦”或“响应于确定”或“响应于检测到”。类似地,短语“如果确定”或“如果检测到[所描述条件或事件]”可以依据上下文被解释为意指“一旦确定”或“响应于确定”或“一旦检测到[所描述条件或事件]”或“响应于检测到[所描述条件或事件]”。As used in this specification and the appended claims, the term "if" may be interpreted as "when" or "once" or "in response to determining" or "in response to detecting" depending on the context. ". Similarly, the phrase "if determined" or "if [the described condition or event] is detected" may be interpreted, depending on the context, to mean "once determined" or "in response to a determination" or "once the [described condition or event] is detected ]" or "in response to detection of [the described condition or event]".
另外,在本申请说明书和所附权利要求书的描述中,术语“第一”、“第二”、“第三”等仅用于区分描述,而不能理解为指示或暗示相对重要性。In addition, in the description of this application and the appended claims, the terms "first", "second", "third", etc. are only used to distinguish the description, and cannot be understood as indicating or implying relative importance.
在本申请说明书中描述的参考“一个实施例”或“一些实施例”等意味着在本申请的一个或多个实施例中包括结合该实施例描述的特定特征、结构或特点。由此,在本说明书中的不同之处出现的语句“在一个实施例中”、“在一些实施例中”、“在其他一些实施例中”、“在另外一些实施例中”等不是必然都参考相同的实施例,而是意味着“一个或多个但不是所有的实施例”,除非是以其他方式另外特别强调。术语“包括”、“包含”、“具有”及它们的变形都意味着“包括但不限于”,除非是以其他方式另外特别强调。Reference in this specification to "one embodiment" or "some embodiments" or the like means that a particular feature, structure or characteristic described in connection with the embodiment is included in one or more embodiments of the application. Therefore, the phrases "in one embodiment", "in some embodiments", "in other embodiments", "in other embodiments", etc. appearing in different places in this specification are not necessarily References are made to the same embodiment, but rather to "one or more but not all embodiments" unless specifically stated otherwise. The terms “including,” “includes,” “having,” and variations thereof all mean “including but not limited to,” unless otherwise specifically emphasized.
针对目前超声图像识别模型的识别准确性低的问题,本申请提供了一种超声图像识别模型及超声图像识别方法,该超声图像识别模型利用类金字塔分解模块对每个通道对应的特征图像进行金字塔分解,能够使后续提取的特征更加精确,从而提高超声图像识别的准确性,利用概率池化模块对每个特征子图像进行概率池化,可以防止特征图像池化时过拟合,通过多种池化方式,能够选择性的提取重要的局部特征,基于这些重要的局部特征来对特征图像进行识别,能够显著提高超声图像识别的准确性。In order to solve the problem of low recognition accuracy of current ultrasonic image recognition models, this application provides an ultrasonic image recognition model and an ultrasonic image recognition method. The ultrasonic image recognition model uses a pyramid-like decomposition module to pyramid the feature images corresponding to each channel. Decomposition can make the subsequent extracted features more accurate, thereby improving the accuracy of ultrasound image recognition. Using the probability pooling module to perform probability pooling on each feature sub-image can prevent overfitting during feature image pooling. Through a variety of The pooling method can selectively extract important local features, and identify feature images based on these important local features, which can significantly improve the accuracy of ultrasound image recognition.
下面对本申请提供的超声图像识别模型进行示例性说明。The following is an exemplary description of the ultrasound image recognition model provided by this application.
如图1所示,本申请提供的超声图像识别模型包括残差模块(如图1中11所示)、类金字塔分解模块(如图1中12所示)、概率池化模块(如图1中13所示)、通道注意力模块(如图1中14所示)、第一特征图像调节模块(如图1中15所示)、第二特征图像调节模块(如图1中16所示)以及图像识别模块(如图1中17所示)。As shown in Figure 1, the ultrasound image recognition model provided by this application includes a residual module (shown as 11 in Figure 1), a pyramid-like decomposition module (shown as 12 in Figure 1), and a probability pooling module (shown as 12 in Figure 1) (shown as 13 in Figure 1), channel attention module (shown as 14 in Figure 1), first feature image adjustment module (shown as 15 in Figure 1), second feature image adjustment module (shown as 16 in Figure 1 ) and image recognition module (shown as 17 in Figure 1).
为了便于说明,在本申请的实施例中仅对超声图像识别模型中的各模块进行了功能说明,残差模块、类金字塔分解模块、概率池化模块、通道注意力模块、第一特征图像调节模块、第二特征图像调节模块以及图像识别模块作为一组功能单元,超声图像识别模型还包括多组上述功能单元。For the convenience of explanation, in the embodiment of this application, only the functions of each module in the ultrasound image recognition model are explained, including the residual module, the pyramid-like decomposition module, the probability pooling module, the channel attention module, and the first feature image adjustment. The module, the second characteristic image adjustment module and the image recognition module serve as a set of functional units, and the ultrasonic image recognition model also includes multiple sets of the above functional units.
各模块之间的连接关系如下:The connection relationship between each module is as follows:
残差模块的输入端接收超声图像,残差模块的输出端分别连接类金字塔分解模块的输入端、第一特征图像调节模块的第一输入端以及第二特征图像调节模块的第一输入端,类金字塔分解模块的输出端连接概率池化模块的输入端,概率池化模块的输出端连接通道注意力模块的输入端,通道注意力模块的输出端连接第一特征图像调节模块的第二输入端,第一特征图像调节模块的输出端连接第二特征图像调节模块的第二输入端,第二特征图像调节模块的输出端连接图像识别模块的输入端,图像识别模块的输出端输出超声图像的识别结果。The input end of the residual module receives the ultrasound image, and the output end of the residual module is respectively connected to the input end of the pyramid-like decomposition module, the first input end of the first feature image adjustment module, and the first input end of the second feature image adjustment module, The output end of the pyramid-like decomposition module is connected to the input end of the probability pooling module, the output end of the probability pooling module is connected to the input end of the channel attention module, and the output end of the channel attention module is connected to the second input of the first feature image adjustment module terminal, the output terminal of the first characteristic image adjustment module is connected to the second input terminal of the second characteristic image adjustment module, the output terminal of the second characteristic image adjustment module is connected to the input terminal of the image recognition module, and the output terminal of the image recognition module outputs the ultrasound image. identification results.
下面分别对各模块进行示例性说明。Each module is illustratively described below.
残差模块,用于对输入超声图像进行卷积处理,得到输入超声图像对应的初始多通道特征图像。The residual module is used to perform convolution processing on the input ultrasound image to obtain the initial multi-channel feature image corresponding to the input ultrasound image.
示例性的,如图2所示,在本申请的一实施例中,上述残差模块包括第一卷积层(如图2中21所示)、第一标准化层(如图2中22所示)、第一激活层(如图2中23所示)、第二卷积层(如图2中24所示)、第二标准化层(如图2中25所示)、第二激活层(如图2中26所示)、第三卷积层(如图2中27所示)、第三标准化层(如图2中28所示),上述各层依次连接。其中,第一卷积层为1*1卷积层,第二卷积层为3*3卷积层,第三卷积层为1*1卷积层,第二卷积层的步长设置为1,此举能够减少跨步长导致的像素关联信息丢失。Illustratively, as shown in Figure 2, in an embodiment of the present application, the above-mentioned residual module includes a first convolution layer (shown as 21 in Figure 2), a first normalization layer (shown as 22 in Figure 2) shown), the first activation layer (shown as 23 in Figure 2), the second convolution layer (shown as 24 in Figure 2), the second normalization layer (shown as 25 in Figure 2), the second activation layer (shown as 26 in Figure 2), the third convolution layer (shown as 27 in Figure 2), and the third normalization layer (shown as 28 in Figure 2), the above layers are connected in sequence. Among them, the first convolution layer is a 1*1 convolution layer, the second convolution layer is a 3*3 convolution layer, the third convolution layer is a 1*1 convolution layer, and the step size of the second convolution layer is set is 1, this can reduce the loss of pixel related information caused by the stride length.
需要说明的是,为了减少冗余参数,在本申请的实施例中,各卷积层的偏置项被去除。It should be noted that in order to reduce redundant parameters, in the embodiment of the present application, the bias terms of each convolution layer are removed.
类金字塔分解模块,用于对初始多通道特征图像中每个通道对应的特征图像进行金字塔分解,得到每个通道对应的多个特征子图像。The pyramid-like decomposition module is used to perform pyramid decomposition on the feature image corresponding to each channel in the initial multi-channel feature image, and obtain multiple feature sub-images corresponding to each channel.
其中,多个特征子图像的尺度互不相同。Among them, the scales of multiple feature sub-images are different from each other.
利用类金字塔分解模块对初始多通道特征图像中每个通道对应的特征图像进行金字塔分解,得到每个通道对应的多个特征子图像的具体过程如下:Use the pyramid-like decomposition module to perform pyramid decomposition on the feature image corresponding to each channel in the initial multi-channel feature image. The specific process of obtaining multiple feature sub-images corresponding to each channel is as follows:
步骤i,将特征图像作为初始图像,对初始图像进行复制,得到中间图像。In step i, the feature image is used as the initial image, and the initial image is copied to obtain the intermediate image.
步骤ii,以预先设置的步长,对中间图像的最外圈像素进行截取,得到最终图像。Step ii: intercept the outermost pixels of the intermediate image with a preset step size to obtain the final image.
步骤iii,若最终图像的尺度大于等于预设尺度阈值,则将最终图像作为步骤i中的初始图像,返回执行步骤i;否则,将初始图像和执行步骤iii得到的所有最终图像,作为每个通道对应的多个特征子图像。尺度阈值的计算公式为Step iii, if the scale of the final image is greater than or equal to the preset scale threshold, use the final image as the initial image in step i, and return to step i; otherwise, use the initial image and all final images obtained by executing step iii as each Multiple feature sub-images corresponding to the channel. The calculation formula of scale threshold is
E=(min(M,N)-2)mod2*TE=(min(M,N)-2)mod2*T
其中,Mx,Nx均表示尺度阈值,E表示尺度阈值计算的判断条件,T表示步长,M,N均表示初始图像的尺度,mod表示取余,|·|表示取绝对值。Among them , M
此外,可通过计算公式In addition, it can be calculated by the formula
得到特征子图像的总数量Q。应理解,所有特征子图像按照尺度从小到大的顺序排列,能够得到特征子图像类金字塔,具体如图3所示。Get the total number of feature sub-images Q. It should be understood that all feature sub-images are arranged in order from small to large scales, and a feature sub-image pyramid can be obtained, as shown in Figure 3.
在对特征图像进行类金字塔分解时,可通过计算公式When performing pyramid-like decomposition on feature images, the calculation formula can be used
PC=C*PQ P C =C*P Q
得到经过类金字塔分解后的像素数量PC,其中,C表示特征图像的通道数量,PQ表示单通道特征图像包含的像素数量,k表示第k个特征子图像。The number of pixels P C after quasi-pyramid decomposition is obtained, where C represents the number of channels of the feature image, P Q represents the number of pixels contained in the single-channel feature image, and k represents the k-th feature sub-image.
概率池化模块,用于分别对多个特征子图像中的每个特征子图像进行概率池化,得到每个特征子图像的初始权重系数。The probability pooling module is used to perform probability pooling on each feature sub-image in multiple feature sub-images to obtain the initial weight coefficient of each feature sub-image.
上述概率池化表示不同特征子图像对应的池化方式的概率互不相同,池化方式包括最大池化、平均池化、最小池化以及随机池化,概率服从标准正态分布。The above probability pooling means that the probabilities of the pooling methods corresponding to different feature sub-images are different from each other. The pooling methods include maximum pooling, average pooling, minimum pooling and random pooling, and the probabilities obey the standard normal distribution.
由于相同的池化在中心对称性的特征图像上产生的结果也具有相似性,若采用统一的池化方式容易使局部特征的过度放大,这不仅会覆盖其它特征,而且难以提取到多维度信息。有鉴于此,本申请对金字塔分解后的不同尺度的特征子图像使用不同的池化方式。Since the results produced by the same pooling on feature images with central symmetry are also similar, if a unified pooling method is used, it is easy to over-amplify local features, which will not only cover other features, but also make it difficult to extract multi-dimensional information. . In view of this, this application uses different pooling methods for feature sub-images of different scales after pyramid decomposition.
具体的,设置不同池化方式的概率服从标准正态分布。示例性的,分布在区间(μ-σ,μ+σ)的池化方式(全局最大池化)的概率为68%,分布在区间(-∞,μ-σ)的池化方式(全局平均池化)的概率为16%,分布在区间(μ+σ,+∞)的池化方式(全局最小池化)的概率为12%,当位于μ-σ或μ+σ上时,采用全局随机池化,概率为4%。实际应用时,优先选取全局最大池化和全局平均池化,因为全局最大池化能够保留特征图像代表性的信息,全局平均池化能够模糊特征图像的信息。Specifically, the probabilities of different pooling methods are set to obey the standard normal distribution. For example, the probability of the pooling method (global maximum pooling) distributed in the interval (μ-σ,μ+σ) is 68%, and the probability of the pooling method (global average) distributed in the interval (-∞,μ-σ) The probability of pooling (pooling) is 16%, and the probability of pooling (global minimum pooling) distributed in the interval (μ+σ,+∞) is 12%. When it is located on μ-σ or μ+σ, the global Random pooling with 4% probability. In practical applications, global maximum pooling and global average pooling are preferred, because global maximum pooling can retain the representative information of feature images, and global average pooling can blur the information of feature images.
通道注意力模块,用于根据初始权重系数计算每个通道的通道权重,并对所有通道的通道权重进行一维卷积处理,得到初始多通道特征图像的通道权重矩阵。The channel attention module is used to calculate the channel weight of each channel based on the initial weight coefficient, and perform one-dimensional convolution processing on the channel weights of all channels to obtain the channel weight matrix of the initial multi-channel feature image.
在本申请的实施例中,通道注意力模块根据初始权重系数计算每个通道的通道权重,并对所有通道的通道权重进行一维卷积处理,得到初始多通道特征图像的通道权重矩阵的具体过程如下:In the embodiment of this application, the channel attention module calculates the channel weight of each channel based on the initial weight coefficient, and performs one-dimensional convolution processing on the channel weights of all channels to obtain the specific channel weight matrix of the initial multi-channel feature image. The process is as follows:
分别针对每个通道,将通道对应的所有初始权重系数相加,得到通道权重。For each channel, add all the initial weight coefficients corresponding to the channel to obtain the channel weight.
示例性的,通过计算公式Wi=wi,1+wi,2+...+wi,j+...+wi,Q,得到通道权重Wi;其中,Wi表示第i个通道的通道权重,wi,j表示第i个通道的第j个初始权重系数。For example, the channel weight Wi is obtained by calculating the formula Wi = wi,1 + wi,2 +...+ wi,j +...+wi ,Q ; where Wi represents the The channel weight of the i channel, w i,j represents the j-th initial weight coefficient of the i-th channel.
利用一维卷积对所有通道的通道权重进行卷积处理,得到中间通道权重矩阵。其中,若通道的总数目小于等于128,则将一维卷积的卷积核设为5,否则将一维卷积的卷积核设为7。One-dimensional convolution is used to convolve the channel weights of all channels to obtain the intermediate channel weight matrix. Among them, if the total number of channels is less than or equal to 128, the convolution kernel of the one-dimensional convolution is set to 5, otherwise the convolution kernel of the one-dimensional convolution is set to 7.
值得一提的是,在此根据通道的总数目设置一维卷积的卷积核,有利于应对超声图像的复杂变动情况。同时,一维卷积对所有通道的通道权重进行卷积处理,能够提取相邻通道之间的依赖关系,可以使通道权重矩阵更加准确。It is worth mentioning that the convolution kernel of one-dimensional convolution is set according to the total number of channels, which is beneficial to deal with the complex changes of ultrasound images. At the same time, one-dimensional convolution performs convolution processing on the channel weights of all channels, which can extract the dependence between adjacent channels and make the channel weight matrix more accurate.
利用激活函数对中间通道权重矩阵进行激活,得到通道权重矩阵。Use the activation function to activate the intermediate channel weight matrix to obtain the channel weight matrix.
示例性的,激活函数的表达式为S(x)=1/(1+e-x)。For example, the expression of the activation function is S(x)=1/(1+e -x ).
需要说明的是,通道权重矩阵中的每一行对应一个通道权重。It should be noted that each row in the channel weight matrix corresponds to a channel weight.
第一特征图像调节模块,用于将通道权重矩阵和初始多通道特征图像相乘,得到中间多通道特征图像The first feature image adjustment module is used to multiply the channel weight matrix and the initial multi-channel feature image to obtain the intermediate multi-channel feature image.
该部分本质上是对每个通道的特征图像进行加权处理。通道权重矩阵中的每个元素表示了对应通道的权重值,通过与初始多通道特征图像相乘,可以调整每个通道的重要性,突出或抑制不同通道的特征信息。这样可以提取出更具有区分度和重要性的特征。This part essentially weights the feature image of each channel. Each element in the channel weight matrix represents the weight value of the corresponding channel. By multiplying with the initial multi-channel feature image, the importance of each channel can be adjusted to highlight or suppress the feature information of different channels. This can extract more distinguishing and important features.
第二特征图像调节模块,用于将中间多通道特征图像和初始多通道特征图像相加,得到最终多通道特征图像。The second feature image adjustment module is used to add the intermediate multi-channel feature image and the initial multi-channel feature image to obtain the final multi-channel feature image.
该步骤可以增强特征图像的表达能力和鲁棒性。通过相加操作,可以将中间多通道特征图像中的特征信息与初始多通道特征图像中的原始信息进行融合,从而得到更丰富、更准确的特征表示。This step can enhance the expressiveness and robustness of the feature image. Through the addition operation, the feature information in the intermediate multi-channel feature image can be fused with the original information in the initial multi-channel feature image, thereby obtaining a richer and more accurate feature representation.
图像识别模块,用于对最终多通道特征图像进行识别,得到最终多通道特征图像的识别结果。The image recognition module is used to recognize the final multi-channel feature image and obtain the recognition result of the final multi-channel feature image.
在本申请的实施例中,图像识别模块为全连接层,在进行特征图像识别时,全连接层中每个神经元将输出一个概率,将概率最大的神经元对应的超声图像类型作为识别结果;其中,神经元和超声图像类型一一对应,即每个神经元对应一种超声图像类型。In the embodiment of this application, the image recognition module is a fully connected layer. When performing feature image recognition, each neuron in the fully connected layer will output a probability, and the ultrasound image type corresponding to the neuron with the highest probability will be used as the recognition result. ; Among them, neurons and ultrasound image types have a one-to-one correspondence, that is, each neuron corresponds to an ultrasound image type.
可见,本申请提供的超声图像识别模型利用类金字塔分解模块对每个通道对应的特征图像进行金字塔分解,能够使后续提取的特征更加精确,从而提高超声图像识别的准确性,利用概率池化模块对每个特征子图像进行概率池化,可以防止特征图像池化时过拟合,通过多种池化方式,能够选择性的提取重要的局部特征,基于这些重要的局部特征来对特征图像进行识别,能够显著提高超声图像识别的准确性。It can be seen that the ultrasound image recognition model provided in this application uses a pyramid-like decomposition module to perform pyramid decomposition on the feature image corresponding to each channel, which can make the subsequently extracted features more accurate, thereby improving the accuracy of ultrasound image recognition, and uses the probability pooling module Probabilistic pooling of each feature sub-image can prevent over-fitting when pooling feature images. Through a variety of pooling methods, important local features can be selectively extracted, and feature images can be processed based on these important local features. Recognition can significantly improve the accuracy of ultrasound image recognition.
下面对本申请提供的超声图像识别方法进行示例性说明。The following is an exemplary description of the ultrasound image recognition method provided by this application.
如图4所示,本申请提供的超声图像识别方法包括以下步骤:As shown in Figure 4, the ultrasound image recognition method provided by this application includes the following steps:
步骤41,获取超声图像训练集。Step 41: Obtain the ultrasound image training set.
上述超声图像训练集包括多个已知类型的超声图像。The above-mentioned ultrasound image training set includes multiple ultrasound images of known types.
下面对获取超声图像训练集的过程进行示例性说明。The following is an exemplary explanation of the process of obtaining the ultrasound image training set.
步骤41.1,从超声图像开放数据集中获取1500张超声图像,构成超声图像训练集。Step 41.1: Obtain 1500 ultrasound images from the ultrasound image open data set to form an ultrasound image training set.
步骤41.2,相关领域的专家对超声图像训练集进行识别,确定每个超声图像的类型。Step 41.2: Experts in related fields identify the ultrasound image training set and determine the type of each ultrasound image.
示例性的,超声图像训练集中的超声图像识别为肝脏对应的超声图像、脾脏对应的超声图像、肾脏对应的超声图像、肠道对应的超声图像、胆囊对应的超声图像、膀胱对应的超声图像、小腿肌肉对应的超声图像以及乳腺对应的超声图像。相应的,超声图像的类型包括肝脏、脾脏、肾脏、肠道、胆囊、膀胱、小腿肌肉以及乳腺。For example, the ultrasound images in the ultrasound image training set are identified as ultrasound images corresponding to the liver, ultrasound images corresponding to the spleen, ultrasound images corresponding to the kidneys, ultrasound images corresponding to the intestines, ultrasound images corresponding to the gallbladder, and ultrasound images corresponding to the bladder, Ultrasound images corresponding to the calf muscles and ultrasonic images corresponding to the breast. Accordingly, the types of ultrasound images include liver, spleen, kidney, intestine, gallbladder, bladder, calf muscles, and breast.
步骤41.3,对分好类的超声图像进行数据扩增。Step 41.3: Perform data amplification on the classified ultrasound images.
本申请实施例中可采用数据扩增方式包括:Data amplification methods that can be used in the embodiments of this application include:
方式1,0°~30°内的任意角度旋转;Mode 1, any angle rotation within 0°~30°;
方式2,随机采用上下或左右翻转的一种;Method 2: Randomly flip up and down or left and right;
方式3,保持图像长宽比放大至原始的1.5倍,并裁剪至原始尺寸;Method 3: Maintain the aspect ratio of the image and enlarge it to 1.5 times the original size, and crop it to the original size;
方式4,裁剪出小部分原始图像,再放大至原始尺寸;Method 4: Crop out a small part of the original image and then enlarge it to the original size;
方式5,在图像中随机挑选一块20*20~50*50大小的像素区域,以该区域的像素均值填充该区域的所有像素;Method 5: Randomly select a pixel area of 20*20~50*50 in the image, and fill all the pixels in the area with the average pixel value of the area;
方式6,随机采用索贝尔算子(Sobel)、罗伯茨算子(Roberts)、Prewitt(一种图像边缘检测的微分算子)以及拉普拉斯算子(Laplacian)中的一种算子对超声图像进行锐化处理;Method 6: Randomly use one of Sobel operator (Sobel), Roberts operator (Roberts operator), Prewitt (a differential operator for image edge detection) and Laplacian operator (Laplacian) for ultrasound The image is sharpened;
方式7,在超声图像中随机添加高斯噪声、椒盐噪声或伽马噪声中的一种。Method 7: Randomly add one of Gaussian noise, salt-and-pepper noise, or gamma noise to the ultrasound image.
步骤41.4,对数据扩增后的超声图像进行缩放。Step 41.4: Scale the ultrasound image after data amplification.
具体的,可将每个类别的超声图像的尺寸缩放为64*64,再对所有相同尺寸的超声图像进行标准化处理,使每张超声图像的任意像素值的均值为0,标准差为1,该过程的计算公式为Pstand=(PS-μ)/σ,其中,Ps表示原始像素值,Pstand表示标准化后的新像素值,μ表示像素的均值,σ表示像素的标准差。Specifically, the size of the ultrasound images of each category can be scaled to 64*64, and then all ultrasound images of the same size can be standardized so that the mean of any pixel value of each ultrasound image is 0 and the standard deviation is 1. The calculation formula of this process is P stand = (P S -μ)/σ, where P s represents the original pixel value, P stand represents the new pixel value after normalization, μ represents the mean value of the pixel, and σ represents the standard deviation of the pixel.
步骤42,将多个已知类型的超声图像逐个输入超声图像识别模型,得到每个超声图像的识别结果。Step 42: Input multiple ultrasound images of known types into the ultrasound image recognition model one by one to obtain the recognition result of each ultrasound image.
其中,超声图像识别模型为上述的超声图像识别模型。Wherein, the ultrasound image recognition model is the above-mentioned ultrasound image recognition model.
下面对步骤42(将多个已知类型的超声图像逐个输入超声图像识别模型,得到每个超声图像的识别结果)的过程进行示例性说明。The following is an exemplary description of the process of step 42 (input multiple known types of ultrasound images into the ultrasound image recognition model one by one to obtain the recognition result of each ultrasound image).
具体的,过计算公式Specifically, through calculation formula
得到每个类型的输出概率Pk。Get the output probability P k of each type.
其中,Pk表示类型k的输出概率,k=1,2,...,Q,Q表示超声图像训练集中超声图像的类型的总数量,zk表示第k个节点的输出值,节点的数量和类型的数量一一对应,j表示第j个节点,n表示输出节点的个数,softmax(·)表示激活函数。Among them, P k represents the output probability of type k, k = 1, 2,..., Q, Q represents the total number of types of ultrasound images in the ultrasound image training set, z k represents the output value of the k-th node, the node's There is a one-to-one correspondence between the number and the number of types. j represents the j-th node, n represents the number of output nodes, and softmax(·) represents the activation function.
示例性的,在本申请的一实施例中,超声图像识别模型的输出概率情况为Exemplarily, in an embodiment of the present application, the output probability of the ultrasound image recognition model is
将输出概率最大的类型作为识别结果。即,将输入的超声图像识别为小腿肌肉对应的超声图像。The type with the highest output probability is used as the recognition result. That is, the input ultrasound image is recognized as an ultrasound image corresponding to the calf muscle.
步骤43,根据识别结果和每个超声图像的类型,计算超声图像识别模型的损失,并根据损失对超声图像识别模型进行反向传播,直至超声图像识别模型拟合,得到训练后的超声图像识别模型。Step 43: Calculate the loss of the ultrasound image recognition model based on the recognition result and the type of each ultrasound image, and perform backpropagation on the ultrasound image recognition model based on the loss until the ultrasound image recognition model is fitted to obtain the trained ultrasound image recognition Model.
具体的,分别针对每个超声图像,通过计算公式Specifically, for each ultrasound image, the calculation formula
FL=-ak(1-pk)γlog(pk)FL=-a k (1-p k ) γ log(p k )
得到超声图像识别模型的损失FL;其中,ak表示类型k的权重,pk表示类型k的输出概率,γ表示聚焦参数,γ∈[0,5];Obtain the loss FL of the ultrasound image recognition model; where a k represents the weight of type k, p k represents the output probability of type k, γ represents the focus parameter, γ∈[0,5];
根据损失FL对超声图像识别模型进行反向传播,直至反向传播得到的新超声图像识别模型的损失值小于预设损失阈值时,确定新超声图像识别模型拟合,得到训练后的超声图像识别模型;其中,反向传播的动量设置为0.9,反向传播的学习率采用热学习率策略获取,反向传播的L1正则化的权重衰减系数设置为5e-4。The ultrasound image recognition model is backpropagated according to the loss FL, until the loss value of the new ultrasound image recognition model obtained by backpropagation is less than the preset loss threshold, the new ultrasound image recognition model is determined to fit, and the trained ultrasound image recognition model is obtained. Model; among them, the momentum of back propagation is set to 0.9, the learning rate of back propagation is obtained using the thermal learning rate strategy, and the weight attenuation coefficient of L1 regularization of back propagation is set to 5e-4.
步骤44,将待识别超声图像输入训练后的超声图像识别模型,得到待识别超声图像的识别结果。Step 44: Input the ultrasound image to be recognized into the trained ultrasound image recognition model to obtain the recognition result of the ultrasound image to be recognized.
上述待识别超声图像是指不确定类型的超声图像。The above-mentioned ultrasound image to be recognized refers to an ultrasound image of an uncertain type.
在本申请的实施例中,在超声图像识别模型训练后,为了进一步提高超声图像识别模型的准确度,可通过计算识别结果对应的平均准确率、精度、敏感度以及F-1分数(是统计学中用来衡量二分类模型精确度的一种指标)来衡量超声图像识别模型的效果,取效果最好(平均准确率最高、精度最高、敏感度最高以及F-1分数最高)的超声图像识别模型作为最终的超声图像识别模型。In the embodiment of the present application, after the ultrasound image recognition model is trained, in order to further improve the accuracy of the ultrasound image recognition model, the average accuracy, precision, sensitivity and F-1 score (which is a statistic) corresponding to the recognition results can be calculated. (an index used in science to measure the accuracy of a binary classification model) to measure the effect of the ultrasound image recognition model, and select the ultrasound image with the best effect (highest average accuracy, highest precision, highest sensitivity, and highest F-1 score) The recognition model serves as the final ultrasound image recognition model.
平均准确率的计算公式如下:The average accuracy is calculated as follows:
精度的计算公式如下:The calculation formula for accuracy is as follows:
敏感度的计算公式如下:The sensitivity is calculated as follows:
F-1分数的计算公式如下:The F-1 score is calculated as follows:
其中,表示正确地预测为正的样例数,表示错误地预测为正的样例数,表示错误地预测为负的样例数,表示正确地预测为负的样例数。Among them, represents the number of samples correctly predicted to be positive, represents the number of samples incorrectly predicted to be positive, represents the number of samples incorrectly predicted to be negative, and represents the number of samples correctly predicted to be negative.
本申请提供的超声图像识别模型及超声图像识别方法与其他现有技术的识别效果对比情况如下表所示:The comparison of the recognition effects of the ultrasound image recognition model and ultrasound image recognition method provided by this application with other existing technologies is as shown in the following table:
由上表结果可知,本申请提供的超声图像识别模型及超声图像识别方法相比于其它现有技术具有一定的先进性,超声图像的识别效果更加准确。It can be seen from the results in the above table that the ultrasonic image recognition model and ultrasonic image recognition method provided by this application are more advanced than other existing technologies, and the ultrasonic image recognition effect is more accurate.
以上所述是本申请的优选实施方式,应当指出,对于本技术领域的普通技术人员来说,在不脱离本申请所述原理的前提下,还可以作出若干改进和润饰,这些改进和润饰也应视为本申请的保护范围。The above is the preferred embodiment of the present application. It should be pointed out that for those of ordinary skill in the art, several improvements and modifications can be made without departing from the principles described in the present application. These improvements and modifications can also be made. should be regarded as the scope of protection of this application.
Claims (8)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310950122.6A CN116797852B (en) | 2023-07-31 | 2023-07-31 | Ultrasonic image recognition model and ultrasonic image recognition method |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310950122.6A CN116797852B (en) | 2023-07-31 | 2023-07-31 | Ultrasonic image recognition model and ultrasonic image recognition method |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN116797852A true CN116797852A (en) | 2023-09-22 |
| CN116797852B CN116797852B (en) | 2025-08-08 |
Family
ID=88043978
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202310950122.6A Active CN116797852B (en) | 2023-07-31 | 2023-07-31 | Ultrasonic image recognition model and ultrasonic image recognition method |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN116797852B (en) |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112819096A (en) * | 2021-02-26 | 2021-05-18 | 西北大学 | Method for constructing fossil image classification model based on composite convolutional neural network |
| US20220383623A1 (en) * | 2021-05-18 | 2022-12-01 | Samsung Electronics Co., Ltd. | Method and apparatus for training neural network models to increase performance of the neural network models |
-
2023
- 2023-07-31 CN CN202310950122.6A patent/CN116797852B/en active Active
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112819096A (en) * | 2021-02-26 | 2021-05-18 | 西北大学 | Method for constructing fossil image classification model based on composite convolutional neural network |
| US20220383623A1 (en) * | 2021-05-18 | 2022-12-01 | Samsung Electronics Co., Ltd. | Method and apparatus for training neural network models to increase performance of the neural network models |
Non-Patent Citations (1)
| Title |
|---|
| 杨静亚;李景霞;王振宇;程海;: "基于卷积神经网络的花朵品种的识别", 黑龙江大学工程学报, no. 04, 25 December 2019 (2019-12-25) * |
Also Published As
| Publication number | Publication date |
|---|---|
| CN116797852B (en) | 2025-08-08 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN112818969B (en) | Knowledge distillation-based face pose estimation method and system | |
| US20240020514A1 (en) | Improper neural network input detection and handling | |
| CN109101897A (en) | Object detection method, system and the relevant device of underwater robot | |
| CN111046964B (en) | Convolutional neural network-based human and vehicle infrared thermal image identification method | |
| CN111507884A (en) | Self-adaptive image steganalysis method and system based on deep convolutional neural network | |
| CN114332075A (en) | Rapid structural defect identification and classification method based on lightweight deep learning model | |
| CN103268607B (en) | A kind of common object detection method under weak supervision condition | |
| CN111145145B (en) | Image surface defect detection method based on MobileNet | |
| CN111079739A (en) | Multi-scale attention feature detection method | |
| CN109389171A (en) | Medical image classification method based on more granularity convolution noise reduction autocoder technologies | |
| CN108985200A (en) | A kind of In vivo detection algorithm of the non-formula based on terminal device | |
| CN116310519A (en) | A Surface Defect Classification Method Based on Semi-Supervised Deep Learning | |
| CN115861651A (en) | Target detection method based on NAM and YOLOv3 | |
| CN115410258B (en) | Attention image-based facial expression recognition method | |
| CN113436125A (en) | Side-scan sonar simulation image generation method, device and equipment based on style migration | |
| CN110020658A (en) | A kind of well-marked target detection method based on multitask deep learning | |
| CN113870254A (en) | Target object detection method and device, electronic equipment and storage medium | |
| CN113723482A (en) | Hyperspectral target detection method based on multi-example twin network | |
| CN112883931A (en) | Real-time true and false motion judgment method based on long and short term memory network | |
| CN116052254A (en) | Visual continuous emotion recognition method based on extended Kalman filtering neural network | |
| CN109284765A (en) | A Natural Image Classification Method Based on Convolutional Neural Networks Based on Negative-valued Features | |
| CN115829996A (en) | Unsupervised synthetic aperture radar image change detection method based on depth feature map | |
| CN115630675A (en) | Target identification method, system, electronic equipment and storage medium | |
| CN115294382A (en) | A method for multi-damage intelligent identification of structures with limited training samples | |
| CN105809200A (en) | Biologically-inspired image meaning information autonomous extraction method and device |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |