CN110598600A - Remote sensing image cloud detection method based on UNET neural network - Google Patents
Remote sensing image cloud detection method based on UNET neural network Download PDFInfo
- Publication number
- CN110598600A CN110598600A CN201910814865.4A CN201910814865A CN110598600A CN 110598600 A CN110598600 A CN 110598600A CN 201910814865 A CN201910814865 A CN 201910814865A CN 110598600 A CN110598600 A CN 110598600A
- Authority
- CN
- China
- Prior art keywords
- cloud detection
- remote sensing
- sensing image
- network
- cloud
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/13—Satellite images
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Molecular Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Evolutionary Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Astronomy & Astrophysics (AREA)
- Remote Sensing (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
本发明公开了一种基于UNET神经网络的遥感图像云检测方法,包括以下步骤:建立具有5个下采样层、5个上采样层的云检测网络,其中所述云检测网络的前四个下采样层,每层后接卷积层和池化层,后四个上采样层每层后接有反卷积层;对原始遥感图像集进行云标注、人工复检、数据增强处理,将处理后的所述遥感图像集分为训练集、评估集和测试集;利用所述训练集与测试集不断优化云检测网络的;利用所述云检测网络对遥感图像进行云检测,输出所述云检测的结果图片。本发明解决了由于云特征提取不够充分而导致检测结果不够理想的问题,提高了检测精度,加强了算法的普适性。
The invention discloses a remote sensing image cloud detection method based on UNET neural network, comprising the following steps: establishing a cloud detection network with 5 down-sampling layers and 5 up-sampling layers, wherein the first four down-sampling layers of the cloud detection network are Sampling layer, each layer is followed by a convolutional layer and a pooling layer, and each of the last four upsampling layers is followed by a deconvolution layer; cloud labeling, manual re-inspection, and data enhancement processing are performed on the original remote sensing image set, and the processing The final remote sensing image set is divided into a training set, an evaluation set and a test set; the cloud detection network is continuously optimized by using the training set and the test set; the cloud detection network is used to detect the remote sensing image, and the cloud is output A picture of the test result. The invention solves the problem of unsatisfactory detection results due to insufficient extraction of cloud features, improves the detection accuracy, and strengthens the universality of the algorithm.
Description
技术领域technical field
本发明涉及遥感图像检测领域,更具体的说,是涉及一种基于UNET神经网络的遥感图像云检测方法。The invention relates to the field of remote sensing image detection, and more specifically, relates to a remote sensing image cloud detection method based on a UNET neural network.
背景技术Background technique
随着卫星遥感技术的快速发展,遥感影像在环境、农业、气象等领域得到广泛的应用。但是光学成像的卫星遥感图像往往容易受到天气影响,导致图像存在云的遮挡,影响图像的进一步应用和分析;只有在气象条件允许且完全无云的时候才能拍摄图像,既影响时效性,又增加了卫星的运行成本。With the rapid development of satellite remote sensing technology, remote sensing images have been widely used in the fields of environment, agriculture, and meteorology. However, satellite remote sensing images of optical imaging are often easily affected by the weather, resulting in cloud occlusion in the image, which affects the further application and analysis of the image; images can only be taken when the weather conditions permit and there is no cloud at all, which not only affects timeliness, but also increases cost of operating the satellite.
现有技术中对遥感图像的处理,需要从图片中提取各种参数,或者对图片进行变换映射,然后再进行阈值判断,此过程非常复杂并且对图像数据的要求比较大,容错性不高,不利于推广。现有技术中,对不同背景的图像进行检测,有着不同的成功率,对于光谱的精度和预处理水平都有一定要求,否则误差会较大,适应性不强。In the prior art, the processing of remote sensing images needs to extract various parameters from the pictures, or transform and map the pictures, and then perform threshold judgment. This process is very complicated and requires relatively large image data, and the fault tolerance is not high. Not conducive to promotion. In the prior art, the detection of images with different backgrounds has different success rates, and there are certain requirements for the accuracy of the spectrum and the level of preprocessing, otherwise the error will be large and the adaptability is not strong.
发明内容Contents of the invention
有鉴于此,本发明的主要目的在于提供一种基于UNET神经网络的遥感图像云检测方法,解决了由于云特征提取不够充分而导致检测结果不够理想的问题,提高了检测精度,增强了算法的普适性。In view of this, the main purpose of the present invention is to provide a remote sensing image cloud detection method based on the UNET neural network, which solves the problem of unsatisfactory detection results due to insufficient cloud feature extraction, improves detection accuracy, and strengthens the algorithm. universality.
本发明的基于UNET神经网络的遥感图像云检测方法,具体技术方案包括:The remote sensing image cloud detection method based on UNET neural network of the present invention, specific technical scheme comprises:
建立具有5个下采样层、5个上采样层、4条跳跃连接链的云检测网络,其中所述云检测网络的前四个下采样层每层后接有池化层,后四个上采样层每层后接有反卷积层;Set up a cloud detection network with 5 downsampling layers, 5 upsampling layers, and 4 skip connection chains, wherein the first four downsampling layers of the cloud detection network are each followed by a pooling layer, and the last four upsampling layers are each followed by a pooling layer. Each layer of the sampling layer is followed by a deconvolution layer;
对原始遥感图像集进行云标注、人工复检、数据增强处理,将处理后的所述遥感图像集分为训练集、评估集和测试集;Perform cloud labeling, manual re-inspection, and data enhancement processing on the original remote sensing image set, and divide the processed remote sensing image set into a training set, an evaluation set, and a test set;
利用所述训练集中的图像对所述云检测网络进行训练,不断更新所述云检测网络的参数;Using the images in the training set to train the cloud detection network, and constantly update the parameters of the cloud detection network;
利用所述测试集中的图像对训练好的所述云检测网络中进行准确度测试,进行所述云检测网络的参数调整及再训练;Using the images in the test set to test the accuracy of the trained cloud detection network, and perform parameter adjustment and retraining of the cloud detection network;
利用所述云检测网络对遥感图像进行云检测,输出所述云检测的结果图片。The cloud detection network is used to perform cloud detection on the remote sensing image, and the result picture of the cloud detection is output.
进一步地,包括:所述云检测网络的第五个下采样层与第一个上采样层相连接。Further, the method includes: the fifth down-sampling layer of the cloud detection network is connected to the first up-sampling layer.
进一步地,包括:所述云检测网络的最后一个上采样层连接全连接层。Further, it includes: the last upsampling layer of the cloud detection network is connected to the fully connected layer.
进一步地,对原始遥感图像集进行云标注包括:利用ENVI软件对所述遥感图像集进行云标注。Further, performing cloud labeling on the original remote sensing image set includes: using ENVI software to perform cloud labeling on the remote sensing image set.
进一步地,对原始遥感图像集进行数据增强包括:对所述遥感图像集进行旋转平移、弹性形变、对比度增强。Further, performing data enhancement on the original remote sensing image set includes: performing rotation translation, elastic deformation, and contrast enhancement on the remote sensing image set.
进一步地,利用所述测试集中的图像对训练好的所述云检测网络中进行准确度测试,进行所述云检测网络的参数调整及再训练包括:根据训练中损失函数的表现,对训练的次数及训练的批次大小进行调整。Further, using the images in the test set to test the accuracy of the trained cloud detection network, performing parameter adjustment and retraining of the cloud detection network includes: according to the performance of the loss function in training, the training The number of times and the batch size of training are adjusted.
进一步地,将处理后的所述遥感图像集分为训练集、评估集和测试集进一步地包括:从所述遥感图像集中随机选取70%的图像作为训练集,20%的图像做为评估集,剩余的10%作为测试集。Further, dividing the processed remote sensing image set into a training set, an evaluation set and a test set further includes: randomly selecting 70% of the images from the remote sensing image set as a training set, and 20% of the images as an evaluation set , and the remaining 10% is used as the test set.
进一步地,包括:所述云检测网络的激活函数为ReLU函数。Further, it includes: the activation function of the cloud detection network is a ReLU function.
进一步地,利用所述训练集中的图像对所述云检测网络进行训练,不断更新所述云检测网络的参数,包括:所述云检测网络的参数可以为卷积的权重和卷积的偏置值。Further, the cloud detection network is trained using the images in the training set, and the parameters of the cloud detection network are continuously updated, including: the parameters of the cloud detection network can be convolution weights and convolution biases value.
进一步地,利用所述测试集中的图像对训练好的所述云检测网络中进行准确度测试,进行所述云检测网络的参数调整及再训练,包括:学习速率设置为0.001。Further, use the images in the test set to test the accuracy of the trained cloud detection network, and perform parameter adjustment and retraining of the cloud detection network, including: setting the learning rate to 0.001.
本发明利用深度学习模型UNET网络,对光学遥感图像进行下采样提取图像特征,然后通过一些跳跃连接与上采样结合一起生成目标检测图像。解决了由于云特征提取不够充分而导致检测结果不够理想的问题,提高了检测精度,加强了算法的普适性,更有利进行科研和经济活动的发展,具有较高的智能型与便捷性。The present invention utilizes a deep learning model UNET network to down-sample an optical remote sensing image to extract image features, and then combine some skip connections with up-sampling to generate a target detection image. It solves the problem of unsatisfactory detection results due to insufficient cloud feature extraction, improves the detection accuracy, strengthens the universality of the algorithm, and is more conducive to the development of scientific research and economic activities, with high intelligence and convenience.
附图说明Description of drawings
为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据提供的附图获得其他的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present invention or the prior art, the following will briefly introduce the drawings that need to be used in the description of the embodiments or the prior art. Obviously, the accompanying drawings in the following description are only It is an embodiment of the present invention, and those skilled in the art can also obtain other drawings according to the provided drawings without creative work.
图1为本发明的流程示意图;Fig. 1 is a schematic flow sheet of the present invention;
图2为本发明的云检测网络结构图;Fig. 2 is a cloud detection network structural diagram of the present invention;
图3为本发明的云检测网络结构图中虚线部分示意图;Fig. 3 is a schematic diagram of the dotted line in the cloud detection network structure diagram of the present invention;
图4为本发明实施例的训练集样本图,其中左半部分为遥感图像原图,右半部为对应的标注图;Fig. 4 is the training set sample diagram of the embodiment of the present invention, wherein the left half is the original image of the remote sensing image, and the right half is the corresponding labeled diagram;
图5为本发明实施例进行检测的遥感图像原图;Fig. 5 is the original picture of the remote sensing image detected by the embodiment of the present invention;
图6为本发明实施例对遥感图像原图进行标注后的标注图;Fig. 6 is an annotated diagram after annotating the original image of the remote sensing image according to the embodiment of the present invention;
图7为本发明实施例遥感图像原图的网络检测结果图。Fig. 7 is a network detection result diagram of the original image of the remote sensing image according to the embodiment of the present invention.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The following will clearly and completely describe the technical solutions in the embodiments of the application with reference to the drawings in the embodiments of the application. Apparently, the described embodiments are only some of the embodiments of the application, not all of them. Based on the embodiments in this application, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the scope of protection of this application.
本发明所提出的基于UNET神经网络的遥感图像云检测方法,最主要的工作是搭建并训练一个UNET网络,该网络以遥感图像作为输入,输出遥感图像云的掩模。利用深度学习平台实现所网络的搭建,网络的训练包括数据集的制作及训练调参过程。本发明实施例中数据集的制作不需要像传统监督学习那样做出图片与标签的配对,只需要收集大量卫星原始遥感图像,使用ENVI软件对原始遥感图像进行初步云标注,然后人工检查修改,得到与原始遥感图像一一对应的云检测掩模图,将云检测掩模图作为训练集。The main task of the UNET neural network-based remote sensing image cloud detection method proposed by the present invention is to build and train a UNET network, which uses remote sensing images as input and outputs the remote sensing image cloud mask. Use the deep learning platform to realize the construction of the network, and the training of the network includes the production of data sets and the process of training and tuning parameters. The production of the data set in the embodiment of the present invention does not require pairing of pictures and labels like traditional supervised learning. It only needs to collect a large number of satellite original remote sensing images, use ENVI software to perform preliminary cloud annotation on the original remote sensing images, and then manually check and modify them. The cloud detection mask map corresponding to the original remote sensing image is obtained, and the cloud detection mask map is used as the training set.
本发明实施例提供的一种基于UNET神经网络的遥感图像云检测方法,具体包括如下步骤:A kind of remote sensing image cloud detection method based on UNET neural network that the embodiment of the present invention provides, specifically comprises the following steps:
一、建立UNET网络模型:1. Establish a UNET network model:
UNET采用的是一个包含下采样和上采样的网络结构,UNET网络可以简单看为先下采样,经过不同程度的卷积,学习了深层次的特征,在经过上采样还原为原图大小,上采样用反卷积实现。下采样用来逐渐展现环境信息,而上采样的过程是结合下采样各层信息和上采样的输入信息来还原细节信息,并且逐步还原图像精度。UNET adopts a network structure including down-sampling and up-sampling. The UNET network can be simply regarded as down-sampling first, and after different degrees of convolution, deep-level features are learned. After up-sampling, it is restored to the size of the original image. Sampling is implemented using deconvolution. Downsampling is used to gradually reveal environmental information, and the process of upsampling is to combine downsampled layers of information and upsampled input information to restore detailed information and gradually restore image accuracy.
UNET网络的上采样阶段与下采样阶段采用了相同数量层次的卷积操作,且使用跳跃连接链结构将下采样层与上采样层相连,使得下采样层提取到的特征可以直接传递到上采样层,这使得UNET网络的像素定位更加准确,分割精度更高。The up-sampling stage and the down-sampling stage of the UNET network use the same number of convolution operations, and use the skip connection chain structure to connect the down-sampling layer to the up-sampling layer, so that the features extracted by the down-sampling layer can be directly transferred to the up-sampling layer. layer, which makes the pixel positioning of the UNET network more accurate and the segmentation accuracy higher.
如附图2所示,本发明实施例建立的云检测网络基于UNET神经网络,包含有5个下采样层、5个上采样层,其中前四个下采样层每层后接有池化层,后四个上采样层每层后接有反卷积层,同时具有四条跳跃连接链,第五个下采样层与第一个上采样层相连接,最后一个上采样层接全连接层输出,激活函数为ReLU函数。As shown in Figure 2, the cloud detection network established by the embodiment of the present invention is based on the UNET neural network, including 5 down-sampling layers and 5 up-sampling layers, wherein the first four down-sampling layers are each followed by a pooling layer , each of the last four upsampling layers is followed by a deconvolution layer, and has four skip connection chains, the fifth downsampling layer is connected to the first upsampling layer, and the last upsampling layer is connected to the fully connected layer output , the activation function is the ReLU function.
下采样层的作用是通过大量卷积核从遥感图像所对应的矩阵中提取高维特征图,可以提取出多张特征图,每张特征图是从图片中提取出来的局部感知,综合这些特征图可以提取出图片中感兴趣的部分。本发明实施例采用大小为3*3,步长为1的卷积核,把原始遥感图像中3*3窗口内的图像变成一个像素,对每个卷积层进行两次卷积,每层的卷积核数量每次下采样后翻倍,后一个下采样层的卷积核数量是前一个下采样层卷积核数量的2倍,可以有效提取出图片的特征信息,实现了对图像特征的多尺度特征识别。The role of the downsampling layer is to extract high-dimensional feature maps from the matrix corresponding to the remote sensing image through a large number of convolution kernels, and multiple feature maps can be extracted. Each feature map is a local perception extracted from the picture, and these features are integrated Graphs can extract interesting parts of a picture. In the embodiment of the present invention, a convolution kernel with a size of 3*3 and a step size of 1 is used to convert the image in the 3*3 window of the original remote sensing image into one pixel, and perform two convolutions on each convolution layer, each The number of convolution kernels in each layer is doubled after each downsampling, and the number of convolution kernels in the next downsampling layer is twice the number of convolution kernels in the previous downsampling layer, which can effectively extract the feature information of the image and realize the Multiscale Feature Recognition of Image Features.
在多层神经网络中,激励函数的作用是对卷积得到的特征向量图进行非线性映射,因为经过多层卷积后,特征向量图的数值在训练的时候变化不大,会导致梯度消失,进而导致卷积网络模型不可训练。在每次卷积后使用ReLU激励函数,可以对卷积层的线性计算的结果进行非线性映射,这样训练前后特征图的变化不会太少,进而可以利用卷积网络模型进行训练。In a multi-layer neural network, the function of the activation function is to perform nonlinear mapping on the eigenvector map obtained by convolution, because after multi-layer convolution, the value of the eigenvector map does not change much during training, which will cause the gradient to disappear , which in turn makes the convolutional network model untrainable. After each convolution, the ReLU activation function can be used to nonlinearly map the results of the linear calculation of the convolutional layer, so that the feature map changes before and after training will not be too small, and then the convolutional network model can be used for training.
对于前四层卷积层,每个卷积层后面接上池化层,其作用是特征降维,压缩数据和参数的数量,减小训练过拟合,同时提高云检测网络的容错性。本发明实施例使用尺寸为2*2,步长为2的池化核,池化后特征图的长度和宽度都会缩短一半。For the first four convolutional layers, each convolutional layer is followed by a pooling layer. Its function is to reduce the feature dimension, compress the number of data and parameters, reduce training overfitting, and improve the fault tolerance of the cloud detection network. The embodiment of the present invention uses a pooling kernel with a size of 2*2 and a step size of 2, and the length and width of the feature map after pooling will be reduced by half.
上采样层的作用是从图片的特征向量图还原成图片。其中每层上采样层的特征图和对应层数的下采样层特征图进行跳跃连接,即附图2中所示的剪接,再进行卷积,上采样层也是尺寸为3*3,步长为2的卷积核,这样做的目的是让图片在还原过程中,能够参考更低维度的特征,还原后的图片更加贴近原图。每个上采样层卷积核的个数与同层的下采样层的卷积核个数对应。The role of the upsampling layer is to restore the picture from the feature vector map of the picture. The feature map of the upsampling layer of each layer and the feature map of the downsampling layer corresponding to the number of layers are skipped and connected, that is, the splicing shown in Figure 2, and then convolution is performed. The size of the upsampling layer is also 3*3, and the step size The convolution kernel is 2. The purpose of this is to allow the image to refer to lower-dimensional features during the restoration process, and the restored image is closer to the original image. The number of convolution kernels of each upsampling layer corresponds to the number of convolution kernels of the downsampling layer of the same layer.
反卷积层的作用可以看作池化层的反作用,采用插值的方法使得特征图扩充,使用尺寸为2*2的,步长为2的反卷积核。The role of the deconvolution layer can be seen as the reaction of the pooling layer. The interpolation method is used to expand the feature map, and the deconvolution kernel with a size of 2*2 and a step size of 2 is used.
全连接层:上采样层的最后一层连接两个全连接层,全连接层使用尺寸为1*1的卷积核,分别得到分割图的前景和背景,然后组合输出结果掩模。Fully connected layer: The last layer of the upsampling layer connects two fully connected layers. The fully connected layer uses a convolution kernel with a size of 1*1 to obtain the foreground and background of the segmentation map respectively, and then combines the output mask.
如附图2所示,本申请实施例的云检测网络具有5个下采样层、5个上采样层,前4个下采样层每层后接卷积层和池化层,第5个下采样层与第1个上采样层相连接,后4个上采样层每层后接有反卷积层,前4层中相应层次的上采样层与下采样层之间具有跳跃连接链。输入一张任意尺寸的遥感图像到云检测网络中,利用下采样层进行特征提取,本发明实施例采用尺寸为3*3,步长为1的卷积核从遥感图像的矩阵中提取高维特征图,每层进行2次卷积操作,每次卷积后利用激励函数将卷积得到的特征向量图进行非线性映射。第一个下采样层使用32个卷积核,进行2次卷积操作后,利用池化层降低下一个下采样层待处理的数据和参数的数量,本发明实施例的池化层使用大小为2*2,步长为2的过滤器,池化后特征图的长度和宽度均为原特征图的1/2,也就是池化后的数据量减少到池化前的1/4,从而可以预防网络过拟合。下采样层作为特征提取部分,每经过一个池化层就一个尺度,本申请实施例包括原图尺度,一共有5个尺度,4个池化层实现了网络对图像特征的多尺度特征识别。As shown in Figure 2, the cloud detection network of the embodiment of the present application has 5 down-sampling layers and 5 up-sampling layers, each of the first 4 down-sampling layers is followed by a convolutional layer and a pooling layer, and the fifth down-sampling layer The sampling layer is connected to the first upsampling layer, and each of the last 4 upsampling layers is followed by a deconvolution layer. There is a skip connection chain between the upsampling layer and the downsampling layer of the corresponding layers in the first 4 layers. Input a remote sensing image of any size into the cloud detection network, and use the downsampling layer to perform feature extraction. The embodiment of the present invention uses a convolution kernel with a size of 3*3 and a step size of 1 to extract high-dimensional images from the matrix of remote sensing images. For the feature map, two convolution operations are performed on each layer, and after each convolution, the activation function is used to perform nonlinear mapping on the feature vector map obtained by convolution. The first downsampling layer uses 32 convolution kernels. After performing two convolution operations, the pooling layer is used to reduce the amount of data and parameters to be processed in the next downsampling layer. The pooling layer in the embodiment of the present invention uses a size For a filter of 2*2 and a step size of 2, the length and width of the feature map after pooling are 1/2 of the original feature map, that is, the amount of data after pooling is reduced to 1/4 of that before pooling, This prevents the network from overfitting. The downsampling layer is used as the feature extraction part, and each pooling layer has a scale. The embodiment of this application includes the scale of the original image. There are 5 scales in total. The 4 pooling layers realize the multi-scale feature recognition of image features by the network.
后一个下采样层的卷积核数量是前一个下采样层的卷积核数量的2倍,如附图2所示,第二个下采样层的卷积核数量为64个,第三个下采样层的卷积核数量为128个,第四个下采样层的卷积核数量为256个,最后一个下采样层的卷积核为512个,最后一个下采样层与第一个上采样相连接。上采样部分,每进行一次上采样,将对应层次的下采样层特征图剪接,然后将上一个上采样层的特征图与对应层次的下采样层特征图进行拼接,再采用尺寸为3*3,步长为2的卷积核进行两次卷积操作。每个上采样层的两次卷积操作完成之后,再进行反卷积操作,采用插值的方法扩充特征图,本发明实施例采用大小为2*2,步长为2的反卷积核,反卷积核就是原卷积核的转置矩阵,使用反卷积填充图像内容,使得图像内容变得丰富。最后一个上采样层连接两个全连接层,全连接层使用尺寸为1*1的卷积核,分别得到分割图的前景和背景,然后组合输出结果,得到原始遥感图像的掩模图像。The number of convolution kernels in the latter downsampling layer is twice the number of convolution kernels in the previous downsampling layer. As shown in Figure 2, the number of convolution kernels in the second downsampling layer is 64, and the number of convolution kernels in the third downsampling layer is 64. The number of convolution kernels in the downsampling layer is 128, the number of convolution kernels in the fourth downsampling layer is 256, and the number of convolution kernels in the last downsampling layer is 512. connected to the sampling. In the upsampling part, every time upsampling is performed, the feature map of the downsampling layer of the corresponding level is spliced, and then the feature map of the previous upsampling layer is spliced with the feature map of the downsampling layer of the corresponding level, and then the size is 3*3 , the convolution kernel with a step size of 2 performs two convolution operations. After the two convolution operations of each upsampling layer are completed, the deconvolution operation is performed again, and the feature map is expanded by interpolation. The embodiment of the present invention uses a deconvolution kernel with a size of 2*2 and a step size of 2. The deconvolution kernel is the transposition matrix of the original convolution kernel, which uses deconvolution to fill the image content, making the image content richer. The last upsampling layer connects two fully connected layers. The fully connected layer uses a convolution kernel with a size of 1*1 to obtain the foreground and background of the segmented image respectively, and then combines the output results to obtain the mask image of the original remote sensing image.
如附图3所示,上采样部分会融合特征提取部分的输出,这样做实际上是将多尺度特征融合在了一起,以第二个上采样层为例,它的特征既来自第4个下采样层输出的特征,即附图3中虚线箭头右边的虚线矩形部分,也有来自第一个上采样层的输出,这样的拼接是贯穿整个云检测网络的,本申请实施例的网络中有四次拼接过程。As shown in Figure 3, the upsampling part will fuse the output of the feature extraction part, which actually fuses multi-scale features together. Taking the second upsampling layer as an example, its features come from the fourth The characteristics of the output of the downsampling layer, that is, the dotted rectangle on the right side of the dotted arrow in Figure 3, also has the output from the first upsampling layer. Such splicing runs through the entire cloud detection network. In the network of the embodiment of the present application, there are Four splicing process.
本发明实施例中的UNET网路的每个卷积层得到的特征图都会连接到对应的上采样层,从而实现对每层特征图都有效使用到后续计算中,避免了直接在高级特征图中进行监督和loss计算,而是结合了低级特征图中的特征,从而可以使得最终所得到的特征图中既包含了高层次的特征,也保护很多低层次的特征,实现了不同层次下特征的融合,提高模型的结果精确度。The feature map obtained by each convolutional layer of the UNET network in the embodiment of the present invention will be connected to the corresponding upsampling layer, so that the feature map of each layer can be effectively used in subsequent calculations, avoiding the direct calculation of high-level feature maps Supervision and loss calculation are carried out in the middle, but the features in the low-level feature map are combined, so that the final feature map can not only contain high-level features, but also protect many low-level features, and realize different levels of features. The fusion can improve the accuracy of the model results.
二、数据集的制作过程:2. The production process of the data set:
本申请实施例将Landsat卫星图片作为数据集,以背景大致分为雪地、草地、荒漠、海洋和城镇五类图片集;In the embodiment of the present application, the Landsat satellite image is used as a data set, and the background is roughly divided into five types of image sets: snow, grassland, desert, ocean, and town;
本申请实施例利用Landsat数据集中对应图片的多光谱图片集和传感器数据,使用ENVI5.4软件中云层识别工具,调整云区域向外扩展的大小(Kernel Size)以及云-无云陆地和水域的可能性阈值(Cloud Probability Threshold)参数,可以初步将图片中的云识别出来,但仍存在误判漏判的可能性。其中,云-无云陆地和水域的可能性阈值越大,可能检测到的云区域越小。The embodiment of the present application utilizes the multi-spectral picture set and sensor data of the corresponding pictures in the Landsat data set, and uses the cloud layer identification tool in the ENVI5.4 software to adjust the outward expansion of the cloud area (Kernel Size) and the cloud-cloudless land and water areas. The cloud probability threshold (Cloud Probability Threshold) parameter can initially identify the cloud in the picture, but there is still the possibility of misjudgment or omission. Among them, the greater the possibility threshold of cloud-cloudless land and water area, the smaller the cloud area that may be detected.
对经过初步自动处理后的图片进行人工复检,对于存在云的误判漏判的掩模图片,利用图片编辑工具对其进行补足,作为训练的标签,每张原始遥感图像都有一张经人工标注的云检测掩模图像与其对应,原始遥感图像及其对应的云检测掩模图像组成训练数据集,如附图3所示的训练集样本,左边是原始遥感图像,右边是与其对应的标注图。Manually re-inspect the pictures after preliminary automatic processing. For the mask pictures with cloud misjudgments and omissions, use picture editing tools to make up for them. As training labels, each original remote sensing image has a manual The marked cloud detection mask image corresponds to it, and the original remote sensing image and its corresponding cloud detection mask image form a training data set, as shown in the training set sample shown in Figure 3. The left side is the original remote sensing image, and the right side is the corresponding annotation picture.
本申请实施例将人工标注后得到的460张图片,利用python的openCV库对每张图片进行旋转、形变、增强对比度等处理,对原始图片及掩模图片可进行随机角度的旋转,例如旋转90°或旋转180°,利用getAffineTransform函数对图片进行形变处理,通过增强对比度的最大值最小值来增强图片的对比度。对部分图片进行二值化处理,因为有些图片的二值化效果没有意义,如云和水连在一起,所以没有对所有图片进行二值化处理。In the embodiment of the present application, the 460 pictures obtained after manual labeling are processed by using python's openCV library to rotate, deform, and enhance the contrast of each picture, and the original picture and the mask picture can be rotated at random angles, such as rotating 90 ° or rotate 180°, use the getAffineTransform function to deform the picture, and enhance the contrast of the picture by enhancing the maximum and minimum values of the contrast. Binarize some pictures, because the binarization effect of some pictures is meaningless, such as clouds and water are connected together, so all pictures are not binarized.
通过数据扩充和增强得到一定数量的图片集,即进行一定程度的旋转平移、弹性形变和灰度值变化等,丰富数据集的多样性,增强网络模型的鲁棒性。A certain number of image sets are obtained through data expansion and enhancement, that is, a certain degree of rotation and translation, elastic deformation, and gray value changes, etc., to enrich the diversity of data sets and enhance the robustness of network models.
若有新的数据集要加入,云检测结果图可以用ENVi软件进行粗检测后人工进行校验得到。If there is a new data set to be added, the cloud detection result map can be manually verified after rough detection with ENVi software.
三、利用数据集进行网络训练的过程如下:3. The process of using the data set for network training is as follows:
将遥感图像数据集用Dropout的方法进行网络的训练,重复训练上述搭建的网络,在训练次数达到预设阈值或测试的准确度达到目标数值后表明所搭建的UNET网络模型已经符合要求。Use the dropout method to train the network on the remote sensing image data set, and repeat the training of the above-built network. After the number of training times reaches the preset threshold or the accuracy of the test reaches the target value, it indicates that the built UNET network model has met the requirements.
本申请实施例的UNET网络模型的损失函数为:The loss function of the UNET network model of the embodiment of the present application is:
其中,E为交叉熵计算函数,x为输入图片的像素点,pl(x)为近似最大函数,wc(x)为类别频率权重,d1(x)表示像素点x到离它最近的云区边界的距离,d2(x)表示像素点x到离它第二近的云区边界的距离,参数w0和标准差σ的值是可调整的。Among them, E is the cross-entropy calculation function, x is the pixel point of the input image, p l(x) is the approximate maximum function, w c (x) is the category frequency weight, and d 1 (x) indicates that the pixel point x is closest to it d 2 (x) represents the distance from the pixel point x to the second closest cloud boundary, the parameter w 0 and the standard deviation σ are adjustable.
利用损失函数计算误差损失,用来评估两个向量的差距,损失函数中近似最大函数pl(x)的对象是人工标注图和云检测网络所输出的图像。在本申请实施例中利用损失函数评估云检测网络生成的掩模与真实标注图像间的差距,在训练过程中反馈给云检测网络进行参数更新。为了使损失函数的结果更贴近实际应用的需求,对交叉熵函数进行加权重设计,用来补偿训练数据集中每类像素的不同频率。The loss function is used to calculate the error loss, which is used to evaluate the gap between two vectors. The object of the approximate maximum function p l(x) in the loss function is the manually labeled image and the image output by the cloud detection network. In the embodiment of the present application, the loss function is used to evaluate the gap between the mask generated by the cloud detection network and the real labeled image, and it is fed back to the cloud detection network for parameter update during the training process. In order to make the result of the loss function closer to the needs of practical applications, the cross-entropy function is weighted to compensate for the different frequencies of each type of pixel in the training data set.
本申请实施例利用训练数据集对云检测网络进行训练时,每次输入一张原始遥感图片到当前网络中,正向传递计算得到当前网络的输出图片,利用损失函数,计算输出图片与其对应的已人工标注的掩模图片之间的误差,利用链式法则将这个误差反向传播到网络中,在反向传播的过程中,利用Adam优化器对网络中的参数如卷积的权重、卷积的偏置等进行一轮更新,完成一次学习。本发明实施例将1050张原始图片作为训练样本,每张图片的大小为572*572,训练时将1050张图片训练一次做为一个周期,本申请实施例共训练了50个周期。When the embodiment of the present application uses the training data set to train the cloud detection network, each time an original remote sensing picture is input into the current network, the forward pass calculation is used to obtain the output picture of the current network, and the loss function is used to calculate the output picture and its corresponding The error between the manually marked mask images is backpropagated to the network by using the chain rule. In the process of backpropagation, the parameters in the network such as the weight of convolution and A round of updating is performed on the offset of the product, and a learning is completed. In the embodiment of the present invention, 1050 original pictures are used as training samples, and the size of each picture is 572*572. During training, 1050 pictures are trained once as a cycle, and a total of 50 cycles are trained in the embodiment of the present application.
四、对网络模型进行调参、再训练:4. Adjust parameters and retrain the network model:
将测试数据集和评估数据集中的图像输入到上述云检测网络中,进行准确度的测试,然后对云检测网络进行训练,调整卷积的权重、卷积的偏置等参数的值,不断重复训练过程,直到达到预期效果。Input the images in the test data set and evaluation data set into the above cloud detection network to test the accuracy, then train the cloud detection network, adjust the convolution weight, convolution bias and other parameters, and repeat training process until the desired effect is achieved.
在深度学习中,神经网路的权重初始化方法对模型的收敛速度和性能有着至关重要的影响,神经网络对权重参数的不停迭代更新,以期达到较好的性能。神经网路的训练过程中的参数学习是基于梯度下降法进行优化的。梯度下降法需要在开始训练时给每一个参数赋一个初始值。在实际应用中,参数服从高斯分布或均匀分布式比较有效的初始化方式。在深度神经网络中,随着层数的增多,在梯度下降的过程中,容易出现梯度消失或梯度爆炸。权重可以采用全0、全1或固定值的形式进行初始化,本发明实施例中使用固定值进行权重进行初始化,所采取的固定值为0.1。In deep learning, the weight initialization method of the neural network has a crucial impact on the convergence speed and performance of the model. The neural network iteratively updates the weight parameters in order to achieve better performance. The parameter learning in the training process of the neural network is optimized based on the gradient descent method. The gradient descent method needs to assign an initial value to each parameter at the beginning of training. In practical applications, parameters obeying Gaussian distribution or uniform distribution are more effective initialization methods. In a deep neural network, as the number of layers increases, gradient disappearance or gradient explosion is prone to occur during gradient descent. The weight can be initialized in the form of all 0s, all 1s, or a fixed value. In the embodiment of the present invention, the weight is initialized with a fixed value, and the adopted fixed value is 0.1.
根据训练损失函数的表现,可以对训练的次数,及训练的批次大小进行调整,调整时要考虑训练速度。进行学习速率的调整,过大的速率会使网络过早进入局部优化,效果不理想,本发明实施例的学习速率最终设置为0.001。According to the performance of the training loss function, the number of training times and the batch size of training can be adjusted, and the training speed should be considered when adjusting. Adjust the learning rate. If the rate is too high, the network will enter local optimization prematurely, and the effect is not ideal. The learning rate in the embodiment of the present invention is finally set to 0.001.
本发明采用了深度学习的网络进行检测,省去了复杂算法的设计与实现,可以更方便直接达到目的。同时,针对现有的人工智能方法处理此类型问题的时候,通常网络结构太过简单,或者受限于结构,增加层数也不能达到较高的精度,容易出现漏判误判的情况。与传统多层卷积全连接网络相比,UNET的上采样依然有大量的通道与下采样连接,这使得网络将上下文信息向更高层分辨率传播,作为结果,扩展路径与收缩路径对称,形成一个U型的形状,有效保留底层的细节,进行更高精度的检测。The present invention uses a deep learning network for detection, which saves the design and implementation of complex algorithms, and can achieve the goal more conveniently and directly. At the same time, when dealing with this type of problem with the existing artificial intelligence methods, the network structure is usually too simple, or is limited by the structure, and the increase in the number of layers cannot achieve higher accuracy, which is prone to misjudgment and misjudgment. Compared with the traditional multi-layer convolutional fully connected network, the upsampling of UNET still has a large number of channels and downsampling connections, which makes the network propagate the context information to a higher layer resolution. As a result, the expansion path and the contraction path are symmetrical, forming A U-shaped shape effectively preserves the underlying details for higher-precision detection.
本发明所提出的基于UNET神经网络的遥感图像云检测方法,是直接从图片输入到图片掩模输出的方法,UNET神经网络的上采样依然有大量的通道与下采样连接,这使得网络将上下文信息向更高层分辨率传播,扩展路径与收缩路径对称,形成一个U型的形状,有效保留底层的细节,进行更高精度的检测。The remote sensing image cloud detection method based on the UNET neural network proposed by the present invention is a method directly from the image input to the image mask output, and the up-sampling of the UNET neural network still has a large number of channels connected to the down-sampling, which makes the network the context The information propagates to higher-level resolution, and the expansion path is symmetrical to the contraction path, forming a U-shaped shape, effectively retaining the details of the bottom layer, and performing higher-precision detection.
本发明在结合使用深度学习卷积神经网络的基础上,对于遥感图片进行云检测,然后对同一地区进行分时拍摄,将有云区域与其无云时的区域进行替换,整合成无云的图像,这种方法的关键点包括:直接对遥感图像进行处理,不用作前期图片的预处理或参数的预提取;直接输出掩模图片不用后期生成。Based on the combined use of deep learning convolutional neural network, the present invention performs cloud detection on remote sensing pictures, and then takes time-sharing photographs of the same area, replaces the cloudy area with its cloudless area, and integrates it into a cloudless image , the key points of this method include: direct processing of remote sensing images, not used as pre-processing of previous images or pre-extraction of parameters; direct output of mask images without post-generation.
以上所述,以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。As mentioned above, the above embodiments are only used to illustrate the technical solutions of the present invention, rather than to limit them; although the present invention has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that: it can still understand the foregoing The technical solutions recorded in each embodiment are modified, or some of the technical features are replaced equivalently; and these modifications or replacements do not make the essence of the corresponding technical solutions deviate from the spirit and scope of the technical solutions of the various embodiments of the present invention.
Claims (10)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910814865.4A CN110598600A (en) | 2019-08-27 | 2019-08-27 | Remote sensing image cloud detection method based on UNET neural network |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910814865.4A CN110598600A (en) | 2019-08-27 | 2019-08-27 | Remote sensing image cloud detection method based on UNET neural network |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN110598600A true CN110598600A (en) | 2019-12-20 |
Family
ID=68856789
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201910814865.4A Pending CN110598600A (en) | 2019-08-27 | 2019-08-27 | Remote sensing image cloud detection method based on UNET neural network |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN110598600A (en) |
Cited By (20)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111428726A (en) * | 2020-06-10 | 2020-07-17 | 中山大学 | Panoptic segmentation method, system, device and storage medium based on graph neural network |
| CN111553289A (en) * | 2020-04-29 | 2020-08-18 | 中国科学院空天信息创新研究院 | A kind of remote sensing image cloud detection method and system |
| CN111611932A (en) * | 2020-05-22 | 2020-09-01 | 哈尔滨工业大学(深圳) | Remote sensing image cloud detection method, terminal and storage medium based on fully convolutional network |
| CN111915592A (en) * | 2020-08-04 | 2020-11-10 | 西安电子科技大学 | Remote sensing image cloud detection method based on deep learning |
| CN111951400A (en) * | 2020-08-06 | 2020-11-17 | 天津大学 | Narrow beam X-ray excited luminescence tomography method based on U-net network |
| CN112101354A (en) * | 2020-09-23 | 2020-12-18 | 广州虎牙科技有限公司 | Text recognition model training method, text positioning method and related device |
| CN112101309A (en) * | 2020-11-12 | 2020-12-18 | 北京道达天际科技有限公司 | Ground object target identification method and device based on deep learning segmentation network |
| CN112348744A (en) * | 2020-11-24 | 2021-02-09 | 电子科技大学 | Data enhancement method based on thumbnail |
| CN112729562A (en) * | 2021-01-26 | 2021-04-30 | 河南工业大学 | Sea ice distribution detection method based on improved U-shaped convolutional neural network |
| CN113034498A (en) * | 2021-04-28 | 2021-06-25 | 江苏欧密格光电科技股份有限公司 | LED lamp bead defect detection and assessment method and device, computer equipment and medium |
| CN113392915A (en) * | 2021-06-23 | 2021-09-14 | 宁波聚华光学科技有限公司 | Industrial part defect detection method based on deep learning |
| CN113421258A (en) * | 2021-07-22 | 2021-09-21 | 湖南星图空间信息技术有限公司 | Automatic cloud detection method based on high-resolution remote sensing image |
| CN113569926A (en) * | 2021-07-13 | 2021-10-29 | 中国资源卫星应用中心 | A method and device for training a cloud segmentation model with high mobility |
| CN113610032A (en) * | 2021-08-16 | 2021-11-05 | 北京市城市规划设计研究院 | Method and device for building identification based on remote sensing images |
| CN113792653A (en) * | 2021-09-13 | 2021-12-14 | 山东交通学院 | A method, system, device and storage medium for remote sensing image cloud detection |
| CN114022408A (en) * | 2021-09-22 | 2022-02-08 | 中国空间技术研究院 | Remote sensing image cloud detection method based on multi-scale convolution neural network |
| CN114494821A (en) * | 2021-12-16 | 2022-05-13 | 广西壮族自治区自然资源遥感院 | Remote sensing image cloud detection method based on feature multi-scale perception and self-adaptive aggregation |
| CN114694035A (en) * | 2022-03-07 | 2022-07-01 | 中国科学院空天信息创新研究院 | Multi-angle polarization satellite cloud detection method based on integrated spatial and spectral features |
| CN116071622A (en) * | 2023-04-06 | 2023-05-05 | 广州思德医疗科技有限公司 | Stomach image recognition model construction method and system based on deep learning |
| CN117036984A (en) * | 2023-10-09 | 2023-11-10 | 武汉大学 | Cascade U-shaped network cloud detection method and system integrating attention mechanisms |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109255294A (en) * | 2018-08-02 | 2019-01-22 | 中国地质大学(北京) | A kind of remote sensing image clouds recognition methods based on deep learning |
| CN110119728A (en) * | 2019-05-23 | 2019-08-13 | 哈尔滨工业大学 | Remote sensing images cloud detection method of optic based on Multiscale Fusion semantic segmentation network |
-
2019
- 2019-08-27 CN CN201910814865.4A patent/CN110598600A/en active Pending
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN109255294A (en) * | 2018-08-02 | 2019-01-22 | 中国地质大学(北京) | A kind of remote sensing image clouds recognition methods based on deep learning |
| CN110119728A (en) * | 2019-05-23 | 2019-08-13 | 哈尔滨工业大学 | Remote sensing images cloud detection method of optic based on Multiscale Fusion semantic segmentation network |
Non-Patent Citations (1)
| Title |
|---|
| JACOB HØXBROEJEPPESEN,ET.AL: "A cloud detection algorithm for satellite imagery based on deep learning", 《REMOTE SENSING OF ENVIRONMENT》 * |
Cited By (29)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111553289A (en) * | 2020-04-29 | 2020-08-18 | 中国科学院空天信息创新研究院 | A kind of remote sensing image cloud detection method and system |
| CN111611932A (en) * | 2020-05-22 | 2020-09-01 | 哈尔滨工业大学(深圳) | Remote sensing image cloud detection method, terminal and storage medium based on fully convolutional network |
| CN111428726A (en) * | 2020-06-10 | 2020-07-17 | 中山大学 | Panoptic segmentation method, system, device and storage medium based on graph neural network |
| CN111428726B (en) * | 2020-06-10 | 2020-09-11 | 中山大学 | Panoptic segmentation method, system, device and storage medium based on graph neural network |
| CN111915592A (en) * | 2020-08-04 | 2020-11-10 | 西安电子科技大学 | Remote sensing image cloud detection method based on deep learning |
| CN111915592B (en) * | 2020-08-04 | 2023-08-22 | 西安电子科技大学 | Remote sensing image cloud detection method based on deep learning |
| CN111951400A (en) * | 2020-08-06 | 2020-11-17 | 天津大学 | Narrow beam X-ray excited luminescence tomography method based on U-net network |
| CN112101354A (en) * | 2020-09-23 | 2020-12-18 | 广州虎牙科技有限公司 | Text recognition model training method, text positioning method and related device |
| CN112101309A (en) * | 2020-11-12 | 2020-12-18 | 北京道达天际科技有限公司 | Ground object target identification method and device based on deep learning segmentation network |
| CN112348744A (en) * | 2020-11-24 | 2021-02-09 | 电子科技大学 | Data enhancement method based on thumbnail |
| CN112729562A (en) * | 2021-01-26 | 2021-04-30 | 河南工业大学 | Sea ice distribution detection method based on improved U-shaped convolutional neural network |
| CN112729562B (en) * | 2021-01-26 | 2022-08-26 | 河南工业大学 | Sea ice distribution detection method based on improved U-shaped convolutional neural network |
| CN113034498A (en) * | 2021-04-28 | 2021-06-25 | 江苏欧密格光电科技股份有限公司 | LED lamp bead defect detection and assessment method and device, computer equipment and medium |
| CN113034498B (en) * | 2021-04-28 | 2023-11-28 | 江苏欧密格光电科技股份有限公司 | LED lamp bead defect detection and assessment method, device, computer equipment and medium |
| CN113392915A (en) * | 2021-06-23 | 2021-09-14 | 宁波聚华光学科技有限公司 | Industrial part defect detection method based on deep learning |
| CN113569926A (en) * | 2021-07-13 | 2021-10-29 | 中国资源卫星应用中心 | A method and device for training a cloud segmentation model with high mobility |
| CN113421258A (en) * | 2021-07-22 | 2021-09-21 | 湖南星图空间信息技术有限公司 | Automatic cloud detection method based on high-resolution remote sensing image |
| CN113610032A (en) * | 2021-08-16 | 2021-11-05 | 北京市城市规划设计研究院 | Method and device for building identification based on remote sensing images |
| CN113792653B (en) * | 2021-09-13 | 2023-10-20 | 山东交通学院 | Method, system, equipment and storage medium for cloud detection of remote sensing image |
| CN113792653A (en) * | 2021-09-13 | 2021-12-14 | 山东交通学院 | A method, system, device and storage medium for remote sensing image cloud detection |
| CN114022408A (en) * | 2021-09-22 | 2022-02-08 | 中国空间技术研究院 | Remote sensing image cloud detection method based on multi-scale convolution neural network |
| CN114022408B (en) * | 2021-09-22 | 2025-02-07 | 中国空间技术研究院 | Cloud Detection Method in Remote Sensing Images Based on Multi-scale Convolutional Neural Network |
| CN114494821A (en) * | 2021-12-16 | 2022-05-13 | 广西壮族自治区自然资源遥感院 | Remote sensing image cloud detection method based on feature multi-scale perception and self-adaptive aggregation |
| CN114494821B (en) * | 2021-12-16 | 2022-11-18 | 广西壮族自治区自然资源遥感院 | Remote sensing image cloud detection method based on feature multi-scale perception and self-adaptive aggregation |
| CN114694035A (en) * | 2022-03-07 | 2022-07-01 | 中国科学院空天信息创新研究院 | Multi-angle polarization satellite cloud detection method based on integrated spatial and spectral features |
| CN116071622A (en) * | 2023-04-06 | 2023-05-05 | 广州思德医疗科技有限公司 | Stomach image recognition model construction method and system based on deep learning |
| CN116071622B (en) * | 2023-04-06 | 2024-01-12 | 广州思德医疗科技有限公司 | Stomach image recognition model construction method and system based on deep learning |
| CN117036984A (en) * | 2023-10-09 | 2023-11-10 | 武汉大学 | Cascade U-shaped network cloud detection method and system integrating attention mechanisms |
| CN117036984B (en) * | 2023-10-09 | 2024-01-09 | 武汉大学 | A cascaded U-shaped network cloud detection method and system integrating attention mechanism |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN110598600A (en) | Remote sensing image cloud detection method based on UNET neural network | |
| CN109299274B (en) | Natural scene text detection method based on full convolution neural network | |
| CN112287978B (en) | Hyperspectral remote sensing image classification method based on self-attention context network | |
| CN110263705B (en) | Two phases of high-resolution remote sensing image change detection system for the field of remote sensing technology | |
| CN110136154B (en) | Remote sensing image semantic segmentation method based on full convolution network and morphological processing | |
| CN109584248B (en) | Infrared target instance segmentation method based on feature fusion and dense connection network | |
| CN111898621B (en) | A Contour Shape Recognition Method | |
| CN111401384B (en) | Transformer equipment defect image matching method | |
| CN112149619B (en) | A natural scene text recognition method based on Transformer model | |
| CN109886066B (en) | Rapid target detection method based on multi-scale and multi-layer feature fusion | |
| CN114998566B (en) | An interpretable multi-scale infrared small target detection network design method | |
| CN108154192B (en) | High-resolution SAR terrain classification method based on multi-scale convolution and feature fusion | |
| CN114820579A (en) | Semantic segmentation based image composite defect detection method and system | |
| CN113313810B (en) | 6D attitude parameter calculation method for transparent object | |
| CN110188685A (en) | A target counting method and system based on double-attention multi-scale cascade network | |
| CN111079683A (en) | Remote sensing image cloud and snow detection method based on convolutional neural network | |
| CN114048822A (en) | An Image Attention Mechanism Feature Fusion Segmentation Method | |
| CN109035260A (en) | A kind of sky areas dividing method, device and convolutional neural networks | |
| CN111191583A (en) | Spatial target recognition system and method based on convolutional neural network | |
| CN112906706A (en) | Improved image semantic segmentation method based on coder-decoder | |
| CN112837320B (en) | A semantic segmentation method for remote sensing images based on parallel atrous convolution | |
| CN118865178B (en) | A flood extraction and location method based on deep learning and spatial information fusion | |
| CN113837956A (en) | Unpaired Supervised Cloud Detection and Large Area Thick Cloud Removal | |
| CN112990041B (en) | A method for extracting buildings from remote sensing images based on improved U-net | |
| WO2023087597A1 (en) | Image processing method and system, device, and medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| RJ01 | Rejection of invention patent application after publication | ||
| RJ01 | Rejection of invention patent application after publication |
Application publication date: 20191220 |