CN118865175A - Automatic labeling system and method of UAV aerial survey data based on artificial intelligence - Google Patents
Automatic labeling system and method of UAV aerial survey data based on artificial intelligence Download PDFInfo
- Publication number
- CN118865175A CN118865175A CN202411009792.9A CN202411009792A CN118865175A CN 118865175 A CN118865175 A CN 118865175A CN 202411009792 A CN202411009792 A CN 202411009792A CN 118865175 A CN118865175 A CN 118865175A
- Authority
- CN
- China
- Prior art keywords
- data
- labeling
- areas
- aerial survey
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C11/00—Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C11/00—Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
- G01C11/04—Interpretation of pictures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
- G06V20/17—Terrestrial scenes taken from planes or by drones
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Software Systems (AREA)
- Medical Informatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Remote Sensing (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- General Health & Medical Sciences (AREA)
- Radar, Positioning & Navigation (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域Technical Field
本发明涉及航测标注技术领域,具体为基于人工智能的无人机航测数据自动标注系统及方法。The present invention relates to the field of aerial survey annotation technology, and in particular to an automatic annotation system and method for unmanned aerial survey data based on artificial intelligence.
背景技术Background Art
随着无人机技术的迅速发展,无人机航测在地理信息、城市规划、农业监测等领域得到了广泛应用。然而,大量的航测数据需要进行准确且高效的标注,以提取有价值的信息。现有技术中,尽管使用了人工智能对无人机航测数据进行自动标注,仍然会出现未能识别的区域。这些未识别区域通常需要直接上报,由工作人员进行人工识别和标注,从而导致人工标注的工作量和强度依然较高,未能有效降低。With the rapid development of drone technology, drone aerial surveys have been widely used in the fields of geographic information, urban planning, agricultural monitoring, etc. However, a large amount of aerial survey data needs to be accurately and efficiently labeled to extract valuable information. In the existing technology, although artificial intelligence is used to automatically label drone aerial survey data, there are still unidentified areas. These unidentified areas usually need to be reported directly and manually identified and labeled by staff, resulting in the workload and intensity of manual labeling being still high and not effectively reduced.
发明内容Summary of the invention
为了解决上述问题,本发明提供了基于人工智能的无人机航测数据自动标注系统及方法。In order to solve the above problems, the present invention provides an automatic labeling system and method for UAV aerial survey data based on artificial intelligence.
本发明采用以下技术方案,基于人工智能的无人机航测数据自动标注方法,包括:The present invention adopts the following technical scheme, a method for automatic labeling of drone aerial survey data based on artificial intelligence, including:
记录无人机航测时每张航测图像对应的拍摄环境数据与拍摄姿态数据;Record the shooting environment data and shooting posture data corresponding to each aerial survey image during drone aerial survey;
提取航测图像中i个未能自动标注的区域,标记为未识别区域;Extract i areas in the aerial survey image that have not been automatically annotated and mark them as unrecognized areas;
将未能自动标注的区域对应的拍摄环境数据与拍摄姿态数据,输入训练完成的第一机器学习模型中,输出对应的标准像素块数量,所述像素块为依次相邻的像素块;Input the shooting environment data and shooting posture data corresponding to the area that cannot be automatically labeled into the first machine learning model that has been trained, and output the corresponding number of standard pixel blocks, where the pixel blocks are pixel blocks that are adjacent in sequence;
以标准像素块数量将i个未识别区域依次划分,得到每个未识别区域对应的子区域;Divide the i unrecognized regions in turn by the number of standard pixel blocks to obtain sub-regions corresponding to each unrecognized region;
将航测图像对应的拍摄环境数据、拍摄姿态数据及n个子区域依次输入训练完成的第二机器学习模型中,获得每个子区域对应的标注结果,标注结果包括标注名称与标注失败。The shooting environment data, shooting posture data and n sub-areas corresponding to the aerial survey image are input into the trained second machine learning model in sequence to obtain the labeling results corresponding to each sub-area, and the labeling results include the labeling name and labeling failure.
作为上述技术方案的进一步描述:划分子区域的方法为:以标准像素块数量为长度得到正方形区域,以正方形区域对未识别区域进行划分,得到为n个子区域,n为大于或等于1的整数;不足以构成正方形区域也作为一个子区域。As a further description of the above technical solution: the method of dividing the sub-regions is: a square region is obtained with the number of standard pixel blocks as the length, and the unidentified region is divided into n sub-regions by the square region, where n is an integer greater than or equal to 1; a region that is not large enough to form a square region is also regarded as a sub-region.
作为上述技术方案的进一步描述:所述第一机器学习模型的训练方法包括:As a further description of the above technical solution: the training method of the first machine learning model includes:
预先收集k组识别单位训练数据,识别单位训练数据包括识别单位特征数据以及识别单位特征数据对应的标准像素块数量,识别单位特征数据包括拍摄环境数据与拍摄姿态数据,将识别单位训练数据转换为第一特征向量;Pre-collect k groups of recognition unit training data, the recognition unit training data including recognition unit feature data and the number of standard pixel blocks corresponding to the recognition unit feature data, the recognition unit feature data including shooting environment data and shooting posture data, and convert the recognition unit training data into a first feature vector;
将每组第一特征向量作为第一机器学习模型的输入,所述第一机器学习模型以每组识别单位特征数据向量对应的预测标准像素块数量向量作为输出,以每组识别单位特征数据向量对应的实际标准像素块数量向量作为预测目标;以最小化所有识别单位特征数据向量的第一预测误差之和作为训练目标;对第一机器学习模型进行训练,直至第一预测误差之和达到收敛时停止训练;所述第一机器学习模型为随机森林模型、梯度提升树模型或循环神经网络模型。Each group of first feature vectors is used as the input of the first machine learning model. The first machine learning model uses the predicted standard pixel block number vector corresponding to each group of identification unit feature data vectors as the output, and the actual standard pixel block number vector corresponding to each group of identification unit feature data vectors as the prediction target; minimizing the sum of the first prediction errors of all identification unit feature data vectors is used as the training target; the first machine learning model is trained until the sum of the first prediction errors reaches convergence, and the training is stopped; the first machine learning model is a random forest model, a gradient boosting tree model or a recurrent neural network model.
作为上述技术方案的进一步描述:所述第一预测误差的计算公式为ZK=(αK-μK)2,其中ZK为第一预测误差,K为识别单位特征数据向量对应第一特征向量的组号,αK为第K组识别单位特征数据向量对应的预测标准像素块数量向量,μK为第K组识别单位特征数据向量对应的实际标准像素块数量向量。As a further description of the above technical solution: the calculation formula of the first prediction error is Z K =(α K -μ K ) 2 , where Z K is the first prediction error, K is the group number of the identification unit feature data vector corresponding to the first feature vector, α K is the predicted standard pixel block number vector corresponding to the Kth group of identification unit feature data vectors, and μ K is the actual standard pixel block number vector corresponding to the Kth group of identification unit feature data vectors.
作为上述技术方案的进一步描述:所述拍摄环境数据包括季节、环境亮度与天气,所述拍摄姿态数据包括无人机的距地高度、倾斜角度与速度:所述倾斜角度包括俯仰、偏航与滚动。As a further description of the above technical solution: the shooting environment data includes season, ambient brightness and weather, and the shooting attitude data includes the height above the ground, tilt angle and speed of the drone: the tilt angle includes pitch, yaw and roll.
作为上述技术方案的进一步描述:所述标注名称高速公路、河流、建筑物、湖泊、农田、机场、工业区与公园。As a further description of the above technical solution: the labeled names are highways, rivers, buildings, lakes, farmlands, airports, industrial areas and parks.
作为上述技术方案的进一步描述:还包括:As a further description of the above technical solution: it also includes:
将标注名称的添加至航测图像中对应子区域位置处,形成二次标注图片;Add the annotation name to the corresponding sub-region position in the aerial survey image to form a secondary annotation image;
将二次标注图片输入训练完成的合理性识别模型中,输出不合理标注名称。Input the secondary annotated images into the trained rationality recognition model and output the unreasonable annotation names.
作为上述技术方案的进一步描述:所述二次标注图片包括标注名称以及每个标注名称对应的区域。As a further description of the above technical solution: the secondary labeled image includes a labeled name and an area corresponding to each labeled name.
作为上述技术方案的进一步描述:所述合理性识别模型的训练方法包括:As a further description of the above technical solution: the training method of the rationality recognition model includes:
预先收集识别数据,识别数据包括二次标注图片与不合理标注名称,将识别数据转换为第二特征向量,将第二特征向量划分为训练集与测试集,使用第二特征数据中的二次标注图片向量作为合理性识别模型的输入,将训练集中的不合理标注名称向量作为合理性识别模型的输出,对合理性识别模型进行训练,得到初始合理性识别模型,利用测试集对初始合理性识别模型进行测试,输出满足预设准确度的合理性识别模型,合理性识别模型为卷积神经网络模型。Collect recognition data in advance, the recognition data includes secondary annotated pictures and unreasonable annotated names, convert the recognition data into a second feature vector, divide the second feature vector into a training set and a test set, use the secondary annotated picture vector in the second feature data as the input of the rationality recognition model, use the unreasonable annotated name vector in the training set as the output of the rationality recognition model, train the rationality recognition model to obtain an initial rationality recognition model, test the initial rationality recognition model with the test set, and output a rationality recognition model that meets the preset accuracy. The rationality recognition model is a convolutional neural network model.
基于人工智能的无人机航测数据自动标注系统,用于实现所述的基于人工智能的无人机航测数据自动标注方法,系统包括:The automatic labeling system for drone aerial survey data based on artificial intelligence is used to implement the automatic labeling method for drone aerial survey data based on artificial intelligence. The system includes:
记录模块,用于记录无人机航测时每张航测图像对应的拍摄环境数据与拍摄姿态数据;The recording module is used to record the shooting environment data and shooting posture data corresponding to each aerial survey image during the UAV aerial survey;
第一处理模块,提取航测图像中i个未能自动标注的区域,标记为未识别区域;The first processing module extracts i areas in the aerial survey image that have not been automatically marked and marks them as unrecognized areas;
第二处理模块,将未能自动标注的区域对应的拍摄环境数据与拍摄姿态数据,输入训练完成的第一机器学习模型中,输出对应的标准像素块数量,所述像素块为依次相邻的像素块;The second processing module inputs the shooting environment data and shooting posture data corresponding to the area that cannot be automatically marked into the trained first machine learning model, and outputs the corresponding number of standard pixel blocks, where the pixel blocks are pixel blocks that are adjacent in sequence;
第三处理模块,用于以标准像素块数量将i个未识别区域依次划分,得到每个未识别区域对应的子区域;The third processing module is used to divide the i unrecognized areas in sequence according to the number of standard pixel blocks to obtain sub-areas corresponding to each unrecognized area;
第四处理模块,用于将航测图像对应的拍摄环境数据、拍摄姿态数据及n个子区域依次输入训练完成的第二机器学习模型中,获得每个子区域对应的标注结果,标注结果包括标注名称与标注失败;A fourth processing module is used to input the shooting environment data, shooting posture data and n sub-areas corresponding to the aerial survey image into the trained second machine learning model in sequence, and obtain a labeling result corresponding to each sub-area, where the labeling result includes a labeling name and a labeling failure;
第五处理模块,用于将标注名称的添加至航测图像中对应子区域位置处,形成二次标注图片;A fifth processing module is used to add the annotation name to the corresponding sub-region position in the aerial survey image to form a secondary annotation image;
合理性处理模块,用于将二次标注图片输入训练完成的合理性识别模型中,输出不合理标注名称。The rationality processing module is used to input the secondary labeled images into the trained rationality recognition model and output the unreasonable labeling names.
有益效果:Beneficial effects:
针对未能自动识别的区域,通过获取航测图像的拍摄环境数据和拍摄姿态数据,选择合适的标准像素块数量,以提升后续标注结果的准确性;选择合适的标准像素块数量可以有效地提高识别精度。For areas that cannot be automatically identified, the shooting environment data and shooting posture data of the aerial survey image are obtained, and the appropriate number of standard pixel blocks is selected to improve the accuracy of subsequent annotation results; selecting the appropriate number of standard pixel blocks can effectively improve the recognition accuracy.
首先,将未识别区域按照标准像素块数量划分为正方形子区域,得到若干子区域,这种细化划分能够更加关注细节,提高对微小地物和复杂场景的识别能力,使用标准识别模型对每个子区域进行处理,由于子区域较小,模型可以更专注于区域内的特征,减少干扰因素,提高识别精度。First, the unrecognized area is divided into square sub-regions according to the number of standard pixel blocks to obtain several sub-regions. This detailed division can pay more attention to details and improve the recognition ability of tiny objects and complex scenes. Each sub-region is processed using a standard recognition model. Since the sub-region is small, the model can focus more on the features within the region, reduce interference factors, and improve recognition accuracy.
以航测图像对应的拍摄环境数据、拍摄姿态数据及若干子区域作为输入数据,模型能够更加全面地分析未识别区域,并准确标注名称,通过自动化处理,降低了人工对未识别区域的标注强度,同时也提高了航测图像整体标注的完整性和准确性。Taking the shooting environment data, shooting posture data and several sub-areas corresponding to the aerial survey image as input data, the model can analyze the unidentified areas more comprehensively and accurately label the names. Through automated processing, it reduces the intensity of manual labeling of unidentified areas, and also improves the completeness and accuracy of the overall labeling of the aerial survey image.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
图1为本发明实施例1中的基于人工智能的无人机航测数据自动标注系统的模块连接图;FIG1 is a module connection diagram of an automatic annotation system for drone aerial survey data based on artificial intelligence in Embodiment 1 of the present invention;
图2为本发明实施例2中基于人工智能的无人机航测数据自动标注系统模块连接图;FIG2 is a module connection diagram of an automatic annotation system for drone aerial survey data based on artificial intelligence in Embodiment 2 of the present invention;
图3为本发明实施例3中基于人工智能的无人机航测数据自动标注方法流程示意图。FIG3 is a schematic diagram of a flow chart of an automatic labeling method for drone aerial survey data based on artificial intelligence in Embodiment 3 of the present invention.
具体实施方式DETAILED DESCRIPTION
为了使本发明实现的技术手段、创作特征、达成目的与功效易于明白了解,下面结合具体图示,进一步阐述本发明。需要说明的是,在不冲突的情况下,本申请中的实施例及实施例中的特征可以相互结合。In order to make the technical means, creative features, objectives and effects of the present invention easy to understand, the present invention is further described below with reference to specific diagrams. It should be noted that the embodiments and features in the embodiments of the present application can be combined with each other without conflict.
请参阅图1,本发明实施例提供基于人工智能的无人机航测数据自动标注系统,包括:记录模块、第一处理模块、第二处理模块、第三处理模块与第四处理模块,各个模块通过有线和/或无线连接。Please refer to Figure 1. An embodiment of the present invention provides an automatic annotation system for drone aerial survey data based on artificial intelligence, including: a recording module, a first processing module, a second processing module, a third processing module and a fourth processing module, and each module is connected by wire and/or wirelessly.
记录模块,用于记录无人机航测时每张航测图像对应的拍摄环境数据与拍摄姿态数据;拍摄环境数据包括季节、环境亮度与天气,拍摄姿态数据包括无人机的距地高度、倾斜角度与速度:倾斜角度包括俯仰、偏航、滚动。The recording module is used to record the shooting environment data and shooting attitude data corresponding to each aerial survey image during drone aerial survey; the shooting environment data includes season, ambient brightness and weather, and the shooting attitude data includes the drone's height from the ground, tilt angle and speed: the tilt angle includes pitch, yaw and roll.
第一处理模块,提取航测图像中i个未能自动标注的区域,标记为未识别区域。The first processing module extracts i areas in the aerial survey image that cannot be automatically marked and marks them as unrecognized areas.
第二处理模块,将未能自动标注的区域对应的拍摄环境数据与拍摄姿态数据,输入训练完成的第一机器学习模型中,输出对应的标准像素块数量,所述像素块为依次相邻的像素块。The second processing module inputs the shooting environment data and shooting posture data corresponding to the area that cannot be automatically marked into the trained first machine learning model, and outputs the corresponding number of standard pixel blocks, where the pixel blocks are pixel blocks adjacent to each other in sequence.
在针对未能自动标注的区域,如何划分小区域再次识别,分小区即下文的子区域,子区域大小将直接关系到后续再次标注结果的准确性,发明人将季节、环境亮度、天气、无人机的距地高度、倾斜角度与速度,作为识别单位模型的输入,原因如下:For areas that have not been automatically labeled, how to divide them into small areas for re-identification, that is, sub-areas below. The size of the sub-areas will directly affect the accuracy of the subsequent re-labeling results. The inventor uses season, ambient brightness, weather, altitude, tilt angle and speed of the drone as inputs to the recognition unit model for the following reasons:
天气如晴天、阴天、雾霾等均会影响图像的清晰度和对比度,例如,阴天会导致图像整体偏暗,而雾霾会导致图像模糊,图像质量较低时,分割成较小区域以提取更多细节,减少天气对识别结果的负面影响;环境亮度影响图像的阴影和亮度,例如,强烈的阳光会在地物上投射明显的阴影,而低光照可能使得某些地物难以辨识,在光照条件较差的情况下,分割成较小区域可以减少阴影和光照变化对识别的影响,提高识别精度;季节变化会导致地物外观的变化,如植物的生长和枯萎,积雪覆盖等,例如,夏季的农田与冬季覆盖积雪的农田在视觉上有很大差异,在季节变化明显的情况下,分割成较小区域可以捕捉到更多细节,有助于区分不同季节的地物特征。Weather conditions such as sunny, cloudy, and haze will affect the clarity and contrast of the image. For example, a cloudy day will cause the image to be dark overall, while haze will cause the image to be blurred. When the image quality is low, it is divided into smaller areas to extract more details and reduce the negative impact of weather on the recognition results. The ambient brightness affects the shadows and brightness of the image. For example, strong sunlight will cast obvious shadows on the objects, while low light may make some objects difficult to identify. In poor lighting conditions, dividing into smaller areas can reduce the impact of shadows and lighting changes on recognition and improve recognition accuracy. Seasonal changes can cause changes in the appearance of objects, such as plant growth and withering, snow coverage, etc. For example, there is a big visual difference between farmland in summer and farmland covered with snow in winter. In the case of obvious seasonal changes, dividing into smaller areas can capture more details and help distinguish the characteristics of objects in different seasons.
无人机的倾斜角度会导致图像的几何畸变,如拉伸、压缩、旋转等,较大的倾斜角度需要分割成较小区域,以减少畸变对识别的影响,使用较小区域可以帮助更好地校正图像并保持识别精度;无人机的速度会影响图像的清晰度和重叠度,较快的速度会导致图像模糊,而较慢的速度则可以获得更清晰的图像,在较快速度下,分割成较小区域可以减少模糊对识别的影响,确保模型能够捕捉到更多细节,反之则相反。分割成较小区域对应的标准像素块数量较少,反之则相反。The tilt angle of the drone will cause geometric distortion of the image, such as stretching, compression, rotation, etc. A larger tilt angle needs to be divided into smaller areas to reduce the impact of distortion on recognition. Using smaller areas can help better correct the image and maintain recognition accuracy. The speed of the drone will affect the clarity and overlap of the image. A faster speed will cause the image to be blurred, while a slower speed can obtain a clearer image. At a faster speed, dividing into smaller areas can reduce the impact of blur on recognition and ensure that the model can capture more details, and vice versa. The number of standard pixel blocks corresponding to dividing into smaller areas is smaller, and vice versa.
综上所述,拍摄环境数据和拍摄姿态数据与选择单位数量像素块数量密切相关。In summary, the shooting environment data and the shooting posture data are closely related to the number of selected unit number pixel blocks.
所述第一机器学习模型的训练方法包括:The training method of the first machine learning model includes:
预先收集k组识别单位训练数据,识别单位训练数据包括识别单位特征数据以及识别单位特征数据对应的标准像素块数量,识别单位特征数据包括拍摄环境数据与拍摄姿态数据,将识别单位训练数据转换为第一特征向量。Collect k groups of recognition unit training data in advance, the recognition unit training data includes recognition unit feature data and the number of standard pixel blocks corresponding to the recognition unit feature data, the recognition unit feature data includes shooting environment data and shooting posture data, and convert the recognition unit training data into a first feature vector.
将每组第一特征向量作为第一机器学习模型的输入,所述第一机器学习模型以每组识别单位特征数据向量对应的预测标准像素块数量向量作为输出,以每组识别单位特征数据向量对应的实际标准像素块数量向量作为预测目标;以最小化所有识别单位特征数据向量的第一预测误差之和作为训练目标;其中,第一预测误差的计算公式为ZK=(αK-μK)2,其中ZK为第一预测误差,K为识别单位特征数据向量对应第一特征向量的组号,αK为第K组识别单位特征数据向量对应的预测标准像素块数量向量,μK为第K组识别单位特征数据向量对应的实际标准像素块数量向量;对第一机器学习模型进行训练,直至第一预测误差之和达到收敛时停止训练;所述第一机器学习模型为随机森林模型、梯度提升树模型(如XGBoost,LightGBM)或循环神经网络模型。Each group of first feature vectors is used as the input of the first machine learning model. The first machine learning model uses the predicted standard pixel block number vector corresponding to each group of identification unit feature data vectors as the output, and uses the actual standard pixel block number vector corresponding to each group of identification unit feature data vectors as the prediction target; minimizing the sum of the first prediction errors of all identification unit feature data vectors is used as the training target; wherein the calculation formula of the first prediction error is Z K = (α K - μ K ) 2 , wherein Z K is the first prediction error, K is the group number of the first feature vector corresponding to the identification unit feature data vector, α K is the predicted standard pixel block number vector corresponding to the Kth group of identification unit feature data vectors, and μ K is the actual standard pixel block number vector corresponding to the Kth group of identification unit feature data vectors; the first machine learning model is trained until the sum of the first prediction errors converges and the training is stopped; the first machine learning model is a random forest model, a gradient boosting tree model (such as XGBoost, LightGBM) or a recurrent neural network model.
第三处理模块,用于以标准像素块数量将i个未识别区域依次划分,得到每个未识别区域对应的子区域。The third processing module is used to divide the i unidentified areas in sequence according to the number of standard pixel blocks to obtain sub-areas corresponding to each unidentified area.
划分子区域的方法为:以标准像素块数量为长度得到正方形区域,以正方形区域对未识别区域进行划分,得到为n个子区域,n为大于或等于1的整数;不足以构成正方形区域也作为一个子区域。The method of dividing the sub-regions is: a square region is obtained with the number of standard pixel blocks as the length, and the unrecognized region is divided into n sub-regions by the square region, where n is an integer greater than or equal to 1; a region that is not enough to form a square region is also regarded as a sub-region.
第四处理模块,用于将航测图像对应的拍摄环境数据、拍摄姿态数据及n个子区域依次输入训练完成的第二机器学习模型中,获得每个子区域对应的标注结果,标注结果包括标注名称与标注失败;针对标注失败对应的子区域,进行上报,由人工进行查看进行标注;标注名称高速公路、河流、建筑物、湖泊、农田、机场、工业区、公园等。The fourth processing module is used to input the shooting environment data, shooting posture data and n sub-areas corresponding to the aerial survey image into the trained second machine learning model in sequence to obtain the labeling results corresponding to each sub-area, and the labeling results include labeling names and labeling failures; the sub-areas corresponding to the labeling failures are reported and manually reviewed and labeled; the labeling names include highways, rivers, buildings, lakes, farmlands, airports, industrial areas, parks, etc.
本实施例针对未能自动识别区域,通过获取航测图像对应的拍摄环境数据与拍摄姿态数据,选择合适的标准像素块数量,合适的标准像素块数量有效提升后续再次标注结果的准确性;再以标准像素块数量为长度得到正方形区域,以正方形区域对未识别区域进行划分,得到为n个子区域,小范围的子区域识别能够更加关注细节,提高对微小地物和复杂场景的识别能力,进行处理时,第一机器学习模型可以更专注于区域内的特征,减少干扰因素,提高识别精度;再以航测图像对应的拍摄环境数据、拍摄姿态数据及n个子区域为输入数据,得到未识别区域的标注名称,降低人工对未识别区域标注强度,同时也有效降低提升航测图像整体标注的完整性和准确性。In this embodiment, for areas that cannot be automatically identified, the shooting environment data and shooting posture data corresponding to the aerial survey image are obtained to select an appropriate number of standard pixel blocks. The appropriate number of standard pixel blocks effectively improves the accuracy of subsequent re-annotation results; a square area is then obtained with the number of standard pixel blocks as the length, and the unidentified area is divided into n sub-areas by the square area. The identification of sub-areas in a small range can pay more attention to details and improve the recognition ability of tiny objects and complex scenes. During processing, the first machine learning model can focus more on the features in the area, reduce interference factors, and improve recognition accuracy; the shooting environment data, shooting posture data and n sub-areas corresponding to the aerial survey image are then used as input data to obtain the labeling name of the unidentified area, thereby reducing the intensity of manual labeling of the unidentified area, and at the same time effectively improving the integrity and accuracy of the overall labeling of the aerial survey image.
实施例2Example 2
请参阅图2所示,本实施例提供的基于人工智能的无人机航测数据自动标注系统,还包括:Please refer to FIG. 2 , the automatic annotation system for drone aerial survey data based on artificial intelligence provided in this embodiment also includes:
第五处理模块,用于将标注名称的添加至航测图像中对应子区域位置处,形成二次标注图片,二次标注图片包括标注名称以及每个标注名称对应的区域。The fifth processing module is used to add the annotation name to the corresponding sub-region position in the aerial survey image to form a secondary annotation image, which includes the annotation name and the region corresponding to each annotation name.
合理性处理模块,用于将二次标注图片输入训练完成的合理性识别模型中,输出不合理标注名称,将不合理标注名称进行上报,由人工进行上报,进行核对,重新标注,有效提升自动标注的准确性及合理性筛查效率。The rationality processing module is used to input the secondary labeled images into the trained rationality recognition model, output the unreasonable labeling names, report the unreasonable labeling names manually, check and re-label them, effectively improving the accuracy of automatic labeling and the efficiency of rationality screening.
不合理标注名称例如,在一大片农田标注区域内,出现了高层建筑标注,农田和高层建筑在地理和功能上有明显差异,高层建筑不可能出现在农田中间,导致出错的原因,可能是将农田中的某些结构(如灌溉设施)误认为是建筑物;在一条高速公路标注区域内,出现了机场跑道的标注,高速公路和机场跑道的建设和用途完全不同,机场跑道不可能紧邻高速公路,导致出错的原因,是模型对线状地物特征的混淆,例如长度、宽度和方向相似;在一个公园绿地标注区域内,出现了工业区标注,公园内部通常为绿地、步道和娱乐设施,不可能有大面积的工业设施,导致出错的原因,是模型对公园和工业区的一些绿化部分混淆。For example, in a large farmland marking area, a high-rise building marking appears. There are obvious geographical and functional differences between farmland and high-rise buildings. It is impossible for a high-rise building to appear in the middle of farmland. The reason for the error may be that some structures in the farmland (such as irrigation facilities) are mistaken for buildings; in a highway marking area, an airport runway marking appears. The construction and use of highways and airport runways are completely different. It is impossible for an airport runway to be adjacent to a highway. The reason for the error is that the model confuses the characteristics of linear features, such as similar lengths, widths, and directions; in a park green space marking area, an industrial area marking appears. The interior of a park is usually green space, trails, and entertainment facilities. It is impossible for there to be large-scale industrial facilities. The reason for the error is that the model confuses some green areas of the park and the industrial area.
合理性识别模型的训练方法包括:The training methods of the rationality recognition model include:
预先收集识别数据,识别数据包括二次标注图片与不合理标注名称,将识别数据转换为第二特征向量,将第二特征向量划分为训练集与测试集,使用第二特征数据中的二次标注图片向量作为合理性识别模型的输入,将训练集中的不合理标注名称向量作为合理性识别模型的输出,对合理性识别模型进行训练,得到初始合理性识别模型,利用测试集对初始合理性识别模型进行测试,输出满足预设准确度的合理性识别模型,合理性识别模型为卷积神经网络模型,如VGG16、ResNet等。Collect recognition data in advance, the recognition data includes secondary annotated pictures and unreasonable annotated names, convert the recognition data into a second feature vector, divide the second feature vector into a training set and a test set, use the secondary annotated picture vector in the second feature data as the input of the rationality recognition model, use the unreasonable annotated name vector in the training set as the output of the rationality recognition model, train the rationality recognition model to obtain an initial rationality recognition model, test the initial rationality recognition model with the test set, and output a rationality recognition model that meets the preset accuracy. The rationality recognition model is a convolutional neural network model, such as VGG16, ResNet, etc.
实施例3Example 3
请参阅图3所示,本实施例提供基于人工智能的无人机航测数据自动标注方法,包括:As shown in FIG3 , this embodiment provides an automatic labeling method for drone aerial survey data based on artificial intelligence, including:
记录无人机航测时每张航测图像对应的拍摄环境数据与拍摄姿态数据;Record the shooting environment data and shooting posture data corresponding to each aerial survey image during drone aerial survey;
提取航测图像中i个未能自动标注的区域,标记为未识别区域;Extract i areas in the aerial survey image that have not been automatically annotated and mark them as unrecognized areas;
将未能自动标注的区域对应的拍摄环境数据与拍摄姿态数据,输入训练完成的第一机器学习模型中,输出对应的标准像素块数量,所述像素块为依次相邻的像素块;Input the shooting environment data and shooting posture data corresponding to the area that cannot be automatically labeled into the first machine learning model that has been trained, and output the corresponding number of standard pixel blocks, where the pixel blocks are pixel blocks that are adjacent in sequence;
以标准像素块数量将i个未识别区域依次划分,得到每个未识别区域对应的子区域;Divide the i unrecognized regions in turn by the number of standard pixel blocks to obtain sub-regions corresponding to each unrecognized region;
将航测图像对应的拍摄环境数据、拍摄姿态数据及n个子区域依次输入训练完成的第二机器学习模型中,获得每个子区域对应的标注结果,标注结果包括标注名称与标注失败。The shooting environment data, shooting posture data and n sub-areas corresponding to the aerial survey image are input into the trained second machine learning model in sequence to obtain the labeling results corresponding to each sub-area, and the labeling results include the labeling name and labeling failure.
以上显示和描述了本发明的基本原理、主要特征和优点。本行业的技术人员应该了解,本发明不受上述实施例的限制,上述实施例和说明书中的描述的只是说明本发明的原理,在不脱离本发明精神和范围的前提下,本发明还会有各种变化和改进,这些变化和改进都落入要求保护的本发明范围内。本发明要求保护范围由所附的权利要求书及其等效物界定。The basic principles, main features and advantages of the present invention are shown and described above. Those skilled in the art should understand that the present invention is not limited to the above embodiments. The above embodiments and the description in the specification are only to illustrate the principles of the present invention. Without departing from the spirit and scope of the present invention, the present invention may have various changes and improvements, which fall within the scope of the present invention to be protected. The scope of protection of the present invention is defined by the attached claims and their equivalents.
Claims (10)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202411009792.9A CN118865175B (en) | 2024-07-26 | 2024-07-26 | Automatic labeling system and method for UAV aerial survey data based on artificial intelligence |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202411009792.9A CN118865175B (en) | 2024-07-26 | 2024-07-26 | Automatic labeling system and method for UAV aerial survey data based on artificial intelligence |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN118865175A true CN118865175A (en) | 2024-10-29 |
| CN118865175B CN118865175B (en) | 2025-01-24 |
Family
ID=93168943
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202411009792.9A Active CN118865175B (en) | 2024-07-26 | 2024-07-26 | Automatic labeling system and method for UAV aerial survey data based on artificial intelligence |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN118865175B (en) |
Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20030161504A1 (en) * | 2002-02-27 | 2003-08-28 | Nec Corporation | Image recognition system and recognition method thereof, and program |
| US20190311203A1 (en) * | 2018-04-09 | 2019-10-10 | Accenture Global Solutions Limited | Aerial monitoring system and method for identifying and locating object features |
| WO2020088076A1 (en) * | 2018-10-31 | 2020-05-07 | 阿里巴巴集团控股有限公司 | Image labeling method, device, and system |
| WO2020163455A1 (en) * | 2019-02-05 | 2020-08-13 | Urugus S.A. | Automatic optimization of machine learning algorithms in the presence of target datasets |
| CN111967313A (en) * | 2020-07-08 | 2020-11-20 | 北京航空航天大学 | Unmanned aerial vehicle image annotation method assisted by deep learning target detection algorithm |
| CN112329550A (en) * | 2020-10-16 | 2021-02-05 | 中国科学院空间应用工程与技术中心 | Weak supervision learning-based disaster-stricken building rapid positioning evaluation method and device |
| CN114004977A (en) * | 2021-10-25 | 2022-02-01 | 成都飞机工业(集团)有限责任公司 | Aerial photography data target positioning method and system based on deep learning |
| CN114550129A (en) * | 2022-01-26 | 2022-05-27 | 江苏联合职业技术学院苏州工业园区分院 | Machine learning model processing method and system based on data set |
| CN114581800A (en) * | 2022-03-04 | 2022-06-03 | 中科合肥智慧农业协同创新研究院 | Rice lodging detection method and device based on unmanned aerial vehicle image and deep learning and storage medium |
| CN114663652A (en) * | 2022-03-25 | 2022-06-24 | 广州极飞科技股份有限公司 | Image processing method, image processing apparatus, management system, electronic device, and storage medium |
| CN115719419A (en) * | 2022-11-30 | 2023-02-28 | 广东工业大学 | A Method for Segmenting Highway Image Regions Based on Semi-Supervised Learning |
| CN117217491A (en) * | 2023-10-11 | 2023-12-12 | 浙江省自然资源征收中心 | Regional digital mapping system and method based on digital land acquisition and relocation |
| CN117372907A (en) * | 2023-10-24 | 2024-01-09 | 江苏东蓝信息技术有限公司 | Clustering method based on artificial intelligent unmanned aerial vehicle simulation training |
| CN118089669A (en) * | 2024-02-29 | 2024-05-28 | 芜湖途中信息科技有限公司 | Topography mapping system and method based on aviation mapping technology |
-
2024
- 2024-07-26 CN CN202411009792.9A patent/CN118865175B/en active Active
Patent Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20030161504A1 (en) * | 2002-02-27 | 2003-08-28 | Nec Corporation | Image recognition system and recognition method thereof, and program |
| US20190311203A1 (en) * | 2018-04-09 | 2019-10-10 | Accenture Global Solutions Limited | Aerial monitoring system and method for identifying and locating object features |
| WO2020088076A1 (en) * | 2018-10-31 | 2020-05-07 | 阿里巴巴集团控股有限公司 | Image labeling method, device, and system |
| WO2020163455A1 (en) * | 2019-02-05 | 2020-08-13 | Urugus S.A. | Automatic optimization of machine learning algorithms in the presence of target datasets |
| CN111967313A (en) * | 2020-07-08 | 2020-11-20 | 北京航空航天大学 | Unmanned aerial vehicle image annotation method assisted by deep learning target detection algorithm |
| CN112329550A (en) * | 2020-10-16 | 2021-02-05 | 中国科学院空间应用工程与技术中心 | Weak supervision learning-based disaster-stricken building rapid positioning evaluation method and device |
| CN114004977A (en) * | 2021-10-25 | 2022-02-01 | 成都飞机工业(集团)有限责任公司 | Aerial photography data target positioning method and system based on deep learning |
| CN114550129A (en) * | 2022-01-26 | 2022-05-27 | 江苏联合职业技术学院苏州工业园区分院 | Machine learning model processing method and system based on data set |
| CN114581800A (en) * | 2022-03-04 | 2022-06-03 | 中科合肥智慧农业协同创新研究院 | Rice lodging detection method and device based on unmanned aerial vehicle image and deep learning and storage medium |
| CN114663652A (en) * | 2022-03-25 | 2022-06-24 | 广州极飞科技股份有限公司 | Image processing method, image processing apparatus, management system, electronic device, and storage medium |
| CN115719419A (en) * | 2022-11-30 | 2023-02-28 | 广东工业大学 | A Method for Segmenting Highway Image Regions Based on Semi-Supervised Learning |
| CN117217491A (en) * | 2023-10-11 | 2023-12-12 | 浙江省自然资源征收中心 | Regional digital mapping system and method based on digital land acquisition and relocation |
| CN117372907A (en) * | 2023-10-24 | 2024-01-09 | 江苏东蓝信息技术有限公司 | Clustering method based on artificial intelligent unmanned aerial vehicle simulation training |
| CN118089669A (en) * | 2024-02-29 | 2024-05-28 | 芜湖途中信息科技有限公司 | Topography mapping system and method based on aviation mapping technology |
Non-Patent Citations (1)
| Title |
|---|
| 李岁月;陈柳;: "基于特征提取的无人机图像建筑区域提取", 无线互联科技, no. 07, 10 April 2018 (2018-04-10) * |
Also Published As
| Publication number | Publication date |
|---|---|
| CN118865175B (en) | 2025-01-24 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN112633535B (en) | Intelligent inspection method and system for photovoltaic power station based on unmanned aerial vehicle image | |
| CN106127204B (en) | A multi-directional water meter reading area detection algorithm based on fully convolutional neural network | |
| CN107316001A (en) | Small and intensive method for traffic sign detection in a kind of automatic Pilot scene | |
| CN114332020A (en) | A method and system for positioning and defect detection of photovoltaic panels based on visible light images | |
| Zhu et al. | Extreme weather recognition using convolutional neural networks | |
| CN114581307B (en) | Multi-image stitching method, system, device and medium for target tracking and recognition | |
| CN116385911A (en) | Lightweight target detection method for unmanned aerial vehicle inspection insulator | |
| CN116631158A (en) | Landslide early warning device based on image recognition and computer vision | |
| CN116682045A (en) | A Fault Detection Method of Beam Pumping Unit Based on Intelligent Video Analysis | |
| CN118887805A (en) | An intelligent transportation infrastructure inspection management platform based on artificial intelligence | |
| CN116152685B (en) | A pedestrian detection method and system based on the field of view of a drone | |
| CN115661932A (en) | Fishing behavior detection method | |
| Liu et al. | Discovering potential illegal construction within building roofs from UAV images using semantic segmentation and object-based change detection | |
| CN116246181A (en) | Extra-high voltage dense channel drift monitoring method based on satellite remote sensing technology | |
| CN120107799A (en) | A method for evaluating the quality of rice transplanting position based on image analysis and geometric modeling | |
| Li et al. | HRHD-HK: A benchmark dataset of high-rise and high-density urban scenes for 3D semantic segmentation of photogrammetric point clouds | |
| CN114529721B (en) | Urban remote sensing image vegetation coverage recognition method based on deep learning | |
| CN119671178A (en) | A method for optimizing the layout of photovoltaic equipment in mountain parks | |
| CN118865175A (en) | Automatic labeling system and method of UAV aerial survey data based on artificial intelligence | |
| CN117726955A (en) | A method and system for identifying termite indicators using optical and infrared drone data | |
| CN118230182A (en) | An automatic identification method of chimneys and condensation towers in urban areas based on YOLOv7 | |
| CN114550016A (en) | Unmanned aerial vehicle positioning method and system based on context information perception | |
| Wang et al. | Real-time detection and analysis system for tea leafhopper | |
| CN119229328B (en) | A method and system for low-altitude UAV waterlogging inspection based on deep learning | |
| CN118840659A (en) | Large-scale mountain forest aerial photography target identification positioning method based on image quality enhancement |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant |