CN111506772B - Image searching method and system based on image feature extraction - Google Patents
Image searching method and system based on image feature extraction Download PDFInfo
- Publication number
- CN111506772B CN111506772B CN201910095245.XA CN201910095245A CN111506772B CN 111506772 B CN111506772 B CN 111506772B CN 201910095245 A CN201910095245 A CN 201910095245A CN 111506772 B CN111506772 B CN 111506772B
- Authority
- CN
- China
- Prior art keywords
- video
- feature
- sequence
- frame
- picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/23—Clustering techniques
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域technical field
本发明涉及以图搜影技术领域,具体的说是一种基于图像特征提取的以图搜影方法及系统。The present invention relates to the technical field of image search, in particular to a method and system for image search based on image feature extraction.
背景技术Background technique
随着影视产业的发展,为了记录人类文化和艺术遗产,或者一些以营利为目的影视观看平台或者广电集团,均保存有大量的视频。但是由于时间的积累,在进行归类保存后,仍然存在庞大的数据。With the development of the film and television industry, in order to record human cultural and artistic heritage, or some for-profit film and television viewing platforms or radio and television groups, a large number of videos are preserved. However, due to the accumulation of time, there is still a huge amount of data after classification and preservation.
随着人们大量的使用和智能化的发展,人们开始寻找更便捷、快速、准确的视频搜索方法。在现有技术中,在进行视频保存时,通常是给每一个视频通过关键字进行命名或者设置记录时间等文字信息进行保存。人们通过输入关键字或者时间进行搜索,但是由于数据量大,关键字往往存在重复,即使缩小范围检索,仍存在很多数据,在进行查找时,人们只能通过观看相关视频内容进行对应查找,数据量大、查找速度慢、浪费时间。With the development of a large number of people's use and intelligence, people begin to look for more convenient, fast and accurate video search methods. In the prior art, when video is saved, it is usually to name each video by a keyword or set text information such as recording time to save. People search by entering keywords or time, but due to the large amount of data, keywords are often repeated. Even if the search is narrowed down, there is still a lot of data. When searching, people can only search by watching related video content. Data The amount is large, the search speed is slow, and time is wasted.
随着发展,人们开始提出采用图片进行影视视频靶向搜索,即通过随机获取的任意一个视频内的图片,从庞大的数据库中进行查找是属于哪一个视频的,在现有技术中,还停留在对每个视频的帧与获取的图片进行逐一对比,耗时时间长,并且服务器要求高,内存占用大,适用性低。With the development, people began to propose the use of pictures for targeted search of film and television videos, that is, to search for which video belongs to a huge database by randomly obtaining pictures in any video. It takes a long time to compare the frames of each video with the acquired pictures one by one, and has high server requirements, large memory usage, and low applicability.
故在现有技术中,针对上述以图搜影的技术还存在改进方向,并有必要提出一种以图搜影技术,来加快搜影速度。Therefore, in the prior art, there is still an improvement direction for the above-mentioned image search technology, and it is necessary to propose a image search technology to speed up the image search.
发明内容Contents of the invention
针对上述问题,本发明提供了一种基于图像特征提取的以图搜影方法及系统,对视频进行图像特征提取,进行特征简洁保存。In view of the above problems, the present invention provides a method and system for image search video based on image feature extraction, which extracts image features from videos and saves features concisely.
为达到上述目的,本发明采用的具体技术方案如下:In order to achieve the above object, the concrete technical scheme that the present invention adopts is as follows:
一种基于图像特征提取的以图搜影方法,其关键技术在于,具体步骤包括:A method of image search based on image feature extraction, the key technology of which is that the specific steps include:
步骤1:对M个视频中的每一帧图片进行特征提取、压缩后,得到M个视频特征点集合;Step 1: After performing feature extraction and compression on each frame of pictures in the M videos, M video feature point sets are obtained;
步骤2:将M个视频特征点集合送入压缩视频数据库,经聚类操作得到聚类簇群和M个视频特征点的簇序列集合;Step 2: Send M video feature point collections into the compressed video database, and obtain cluster clusters and cluster sequence sets of M video feature points through clustering operations;
步骤3:图片获取端对待查询图片进行获取,经图片特征提取模块后得到待查询图片特征序列;并送入压缩视频数据库进行聚类,得到待查询图片簇序列;Step 3: The picture acquisition terminal acquires the picture to be queried, and obtains the feature sequence of the picture to be queried after the picture feature extraction module; and sends it to the compressed video database for clustering, and obtains the cluster sequence of the picture to be queried;
步骤4:将待查询图片簇序列与M个视频特征点的簇序列集合进行对比,得到序列相似度;并提取与待查询图片簇序列的序列相似度最高对应的视频特征点集合以及帧特征序列;Step 4: Compare the image cluster sequence to be queried with the cluster sequence set of M video feature points to obtain the sequence similarity; and extract the video feature point set and frame feature sequence corresponding to the highest sequence similarity of the image cluster sequence to be queried ;
步骤5:根据视频特征点集合以及帧特征序列的索引,提取对应的视频中的帧图片,并将该帧图片与待查询图片进行对比,得到图片相似度;Step 5: According to the video feature point set and the index of the frame feature sequence, extract the frame picture in the corresponding video, and compare the frame picture with the picture to be queried to obtain the picture similarity;
步骤6:若图片相似度大于图片相似度阈值,则输出对应视频或者对应视频的索引。Step 6: If the picture similarity is greater than the picture similarity threshold, output the corresponding video or the index of the corresponding video.
通过上述设计,M个视频经过特征提取、压缩后,对应得到M个视频特征点集合。并在压缩视频数据库中,通过聚类形成聚类簇群和M个视频对应的M个视频特征点的簇序列集合。当获取到任意一张图片后,经特征提取、聚类后,得到待查询图片簇序列,通过将待查询图片簇序列与M个视频特征点的簇序列集合相对比,从M个视频特征点的簇序列集合中寻找到最大相似度的帧图片以及对应视频的索引。通过调取符合要求的帧图片和待查询图片进行实际图片进行对比后,来进一步确实查找到的视频的准确性,提高查找精度。通过设计新的模块,实现视频特征提取和压缩,将缩小视频内存占用量,利用视频中,连续的帧相似度大的特征,删除相似度大的帧图片,对视频中的帧进行大大缩减。并且在结合图片特征序列的独特性,每一张图片提取的特征序列的不同,进行特征提取后,保留的视频特征点集合内存进一步被缩减。其中,M为正整数。Through the above design, after feature extraction and compression of M videos, M video feature point sets are correspondingly obtained. And in the compressed video database, a clustering cluster and a cluster sequence set of M video feature points corresponding to the M videos are formed by clustering. When any picture is obtained, after feature extraction and clustering, the cluster sequence of the picture to be queried is obtained. By comparing the cluster sequence of the picture to be queried with the cluster sequence set of M video feature points, the M video feature points Find the frame picture with the maximum similarity and the index of the corresponding video in the cluster sequence set of . By comparing the frame pictures that meet the requirements with the actual pictures to be queried, the accuracy of the found videos is further confirmed and the search accuracy is improved. By designing a new module to realize video feature extraction and compression, the memory usage of video will be reduced, and the features of continuous frame similarity in video will be used to delete frame pictures with large similarity, and the frames in video will be greatly reduced. And combining the uniqueness of the picture feature sequence, the feature sequence extracted from each picture is different, after feature extraction, the memory of the reserved video feature point set is further reduced. Wherein, M is a positive integer.
进一步的,步骤1中,实现对M个视频中的每一帧图片进行特征提取、压缩后,得到M个视频特征点集合的具体步骤为:Further, in step 1, after implementing feature extraction and compression for each frame of pictures in the M videos, the specific steps for obtaining M video feature point sets are:
步骤11:视频获取端依次从M个视频特征点获取一段视频Ax,并保存在待提取存储模块内;Step 11: The video acquisition end sequentially acquires a section of video Ax from M video feature points, and stores it in the storage module to be extracted;
步骤12:待提取存储模块得到视频Ax的总帧数N,并将视频Ax中的每一帧依次发送至特征提取与对比模块;Step 12: The storage module to be extracted obtains the total number of frames N of the video Ax, and sends each frame in the video Ax to the feature extraction and comparison module in turn;
步骤13:特征提取与对比模块用于指定基准帧,并根据基准帧对视频Ax进行预处理、图像特征提取、对比、筛选、删除、压缩后,得到视频特征点集合,具体内容为:Step 13: The feature extraction and comparison module is used to specify the reference frame, and perform preprocessing, image feature extraction, comparison, screening, deletion, and compression on the video Ax according to the reference frame to obtain a set of video feature points. The specific content is:
S131:特征提取与对比模块获取第一帧,令该第一帧为基准帧,提取第一帧图像特征后对提取的特征进行排序,得到基准帧特征序列,并将该基准帧特征序列保存在特征序列存储模块内;S131: The feature extraction and comparison module obtains the first frame, makes the first frame a reference frame, sorts the extracted features after extracting the image features of the first frame, obtains a reference frame feature sequence, and saves the reference frame feature sequence in In the feature sequence storage module;
S132:特征提取与对比模块获取下一帧,并令该帧为待对比帧,提取待对比帧图像特征后对提取的特征进行排序,得到待对比帧特征序列;S132: The feature extraction and comparison module obtains the next frame, and makes the frame the frame to be compared, extracts the image features of the frame to be compared, and sorts the extracted features to obtain a feature sequence of the frame to be compared;
S133:将待对比帧特征序列与所述基准帧特征序列进行对比;S133: Compare the feature sequence of the frame to be compared with the feature sequence of the reference frame;
若二者相似度大于等于设定的相似度阈值,则抛弃该待对比帧特征序列,进入步骤S134;If the similarity between the two is greater than or equal to the set similarity threshold, discard the feature sequence of the frame to be compared and enter step S134;
若相似度小于设定的相似度阈值,则令该待对比帧特征序列为新的基准帧特征序列,即:令对应的待对比帧为基准帧;并将新的基准帧特征序列依次保存在特征序列存储模块内,进入步骤S134;If the similarity is less than the set similarity threshold, then make the feature sequence of the frame to be compared a new reference frame feature sequence, that is: let the corresponding frame to be compared be a reference frame; and save the new reference frame feature sequence in sequence In the feature sequence storage module, enter step S134;
S134:判断视频Ax中的第N帧是否已经对比完毕;若是,特征序列存储模块将保存的所有基准帧特征序列组成视频特征点集合,并输出该视频Ax的视频特征点集合;否则返回步骤S132;S134: Determine whether the Nth frame in the video Ax has been compared; if so, the feature sequence storage module forms a video feature point set with all reference frame feature sequences preserved, and outputs the video feature point set of the video Ax; otherwise, return to step S132 ;
步骤14:判断M个视频是否全部压缩完毕,若是结束步骤1,否则返回步骤11。Step 14: Determine whether all the M videos have been compressed, if so, end step 1, otherwise return to
由于视频是由依次排列的帧组成,并且要组成动画的过程,相邻的帧的图像相似度大。通过上述设计,在实现以图搜影功能时,为了精简压缩视频数据库,提取视频每一帧的图像特征,将相邻的帧进行逐一对比,删除重复出现的帧,最后保留下来的为重复率低的视频特征点集合。在进行以图搜影功能时,提取到该图片的图像特征后,可以快速找到对应的视频特征点集合。相对于现有技术,以图搜影过程周期短,便于大众推广使用。图片出现的在视频的时间位置都能够较为精确的定位。Since the video is composed of frames arranged in sequence, and the process of forming an animation, the image similarity of adjacent frames is large. Through the above design, in order to simplify and compress the video database, extract the image features of each frame of the video, compare the adjacent frames one by one, delete the repeated frames, and finally retain the repetition rate. Low set of video feature points. When performing the image search function, after extracting the image features of the image, you can quickly find the corresponding video feature point set. Compared with the existing technology, the process cycle of image search is short, which is convenient for popularization and use. The time position of the picture appearing in the video can be more accurately positioned.
所述特征提取与对比模块在获取到任意帧的图片时,均需要先对图片进行预处理,其中预处理内容包括:图片大小重置处理和灰度处理,其中,图片大小重置处理即将图片重置成统一大小,其中图片大小可根据系统自定义设定。其中,进行灰度处理的灰度值也自定义设定,灰度值在0~255。When the feature extraction and comparison module acquires a picture of any frame, it needs to preprocess the picture first, wherein the preprocessing content includes: picture size reset processing and grayscale processing, wherein the picture size reset process is about to process the picture Reset to a uniform size, where the picture size can be customized according to the system. Among them, the grayscale value for grayscale processing is also customized, and the grayscale value ranges from 0 to 255.
其中相似度阈值自定义设定,将待对比帧特征序列与基准帧特征序列进行对比后,若相似度大于相似度阈值,则认为两帧只需要保存基准帧即可,依次类推,一段视频进行视频特征提取,并只保存部分特征序列后,用于以图搜影的视频占用容量大大降低,且不影响实现以图搜影,并且加快了搜索速度。Among them, the similarity threshold is custom-set. After comparing the feature sequence of the frame to be compared with the feature sequence of the reference frame, if the similarity is greater than the similarity threshold, it is considered that only the reference frame needs to be saved for the two frames, and so on. After video feature extraction and saving only part of the feature sequence, the video occupation capacity used for image search is greatly reduced without affecting the realization of image search, and the search speed is accelerated.
其中Ax中的x为正整数。Where x in Ax is a positive integer.
再进一步的,步骤2中,M个视频特征点集合经聚类操作得到聚类簇群的具体步骤为:Furthermore, in
步骤21a:从M个视频特征点集合任意选择k个特征作为初始聚类簇的聚类中心;Step 21a: randomly select k features from M video feature point sets as the cluster center of the initial cluster;
步骤21b:分别计算M个视频特征点集合中所有特征到k个特征作为聚类簇中心的聚类距离,并设定最小聚类距离对特征进行划分;Step 21b: Calculate the clustering distances from all the features in the M video feature point sets to k features as cluster centers, and set the minimum clustering distance to divide the features;
步骤21c:选择出聚类中心发生变化的聚类簇,并计算对应聚类簇的聚类距离均值,根据聚类距离均值确定该聚类簇的聚类中心,Step 21c: Select the cluster whose cluster center changes, and calculate the mean value of the cluster distance of the corresponding cluster, and determine the cluster center of the cluster according to the mean value of the cluster distance,
步骤21d:若聚类中心不再发生变化,则终止聚类,输出k聚类簇形成的聚类簇群以及对应的聚类中心,否则返回步骤21b;Step 21d: If the cluster center does not change any more, then terminate the clustering, output the cluster cluster formed by k clusters and the corresponding cluster center, otherwise return to step 21b;
步骤2中,M个视频特征点集合经聚类操作得到M个视频特征点的簇序列集合的具体步骤为:对所述聚类簇群中的k个聚类簇进行编号;In
依次计算M个视频特征点集合中,每个特征到k个聚类簇的聚类距离;Calculate the clustering distances from each feature to k clusters in the M video feature point sets in turn;
对每个特征的k个聚类距离进行排序,并将该特征归类到最小聚类距离对应聚类簇,并获取该特征对应的聚类簇编号,直至每一个视频特征点集合中所有的特征均归类完毕后,得到M个簇序列集合。Sort the k clustering distances of each feature, and classify the feature into the cluster corresponding to the minimum clustering distance, and obtain the cluster number corresponding to the feature, until all the video feature points in each video feature point set After the features are classified, M cluster sequence sets are obtained.
通过聚类分析算法,对视频的特征进行了归类,得到视频特征独有的聚类簇群,并对M个视频的特征进行了归类。在对比待查询图片时,提供了对比依据。提取待查询图片中的特征后,放入聚类簇群对每个特征进行归类后,与M个视频对应的M个簇序列集合进行一一对比,对比M个簇序列集合中是否存在与待查询图片特征归类相同的序列。通过聚类算法,将图片进行特征聚类,便于特征相似对比,为以图搜影提供了对比依据。Through the clustering analysis algorithm, the features of the video are classified, and the unique cluster clusters of the video features are obtained, and the features of M videos are classified. When comparing the images to be queried, a comparison basis is provided. After extracting the features in the picture to be queried, put them into the clustering clusters to classify each feature, and compare them one by one with the M cluster sequence sets corresponding to the M videos, and compare whether there are any similarities in the M cluster sequence sets. The image features to be queried are classified into the same sequence. Through the clustering algorithm, the features of the pictures are clustered, which is convenient for feature similarity comparison, and provides a comparison basis for image search.
再进一步描述,步骤3中,图片获取端对待查询图片进行获取,经特征提取后得到待查询图片特征序列;并送入压缩视频数据库进行聚类,得到待查询图片簇序列的具体步骤为:To further describe, in
步骤31:图片获取端获取待查询图片后,将待查询图片发送至图片特征提取模块;Step 31: After the image acquisition terminal obtains the image to be queried, it sends the image to be queried to the image feature extraction module;
步骤32:图片特征提取模块结合第三方视觉库,提取到待查询图片特征序列后送入所述压缩视频数据库;Step 32: the picture feature extraction module combines the third-party visual library to extract the picture feature sequence to be queried and send it to the compressed video database;
步骤33:计算待查询图片特征序列中的所有特征聚类距离,并将该特征归类到最小聚类距离对应聚类簇,得到待查询图片特征序列对应的待查询图片簇序列。Step 33: Calculate the clustering distances of all the features in the feature sequence of the picture to be queried, and classify the feature into the cluster corresponding to the minimum clustering distance, and obtain the cluster sequence of the picture to be queried corresponding to the feature sequence of the picture to be queried.
采用上述方案,将待查询图片进行特征提取和归类,用于与M个簇序列集合进行对比,以找出对应的帧图片和源视频。Using the above scheme, feature extraction and classification are performed on the image to be queried, and it is used for comparison with M cluster sequence sets to find out the corresponding frame image and source video.
再进一步描述,所述特征提取与对比模块、图片特征提取模块均与第三方视觉库连接;所述特征提取与对比模块至少设置有帧发送单元和特征序列接收单元,该帧发送单元用于将所述待对比帧发送至所述第三方视觉库进行待对比帧图像特征提取,并将得到的待对比帧特征序列反馈给特征序列接收单元;所述图片特征提取模块设置有图片发送单元和待查询图片特征序列接收单元,该图片发送单元用于将待查询图片发送至所述第三方视觉库进行特征提取,并将得到的待查询图片特征序列反馈给待查询图片特征序列接收单元。To further describe, the feature extraction and comparison module and the picture feature extraction module are all connected to a third-party visual library; the feature extraction and comparison module is at least provided with a frame sending unit and a feature sequence receiving unit, and the frame sending unit is used for The frame to be compared is sent to the third-party visual library for image feature extraction of the frame to be compared, and the obtained frame feature sequence to be compared is fed back to the feature sequence receiving unit; the picture feature extraction module is provided with a picture sending unit and a frame to be compared. The query picture feature sequence receiving unit, the picture sending unit is used to send the picture to be queried to the third-party visual library for feature extraction, and feed back the obtained picture feature sequence to be queried to the picture feature sequence receiving unit to be queried.
第三方视觉库为计算机视觉开源库,可以是OpenCV、JavaCV、Torch3Vision、ImLab、CIMG、Generic Image Library(GIL)-boost integration等等。The third-party vision library is an open source library for computer vision, which can be OpenCV, JavaCV, Torch3Vision, ImLab, CIMG, Generic Image Library (GIL)-boost integration, etc.
再进一步的技术方案为:所述第三方视觉库为OpenCV视觉库。A further technical solution is: the third-party vision library is an OpenCV vision library.
OpenCV是一个开放源代码的计算机视觉应用平台,由英特尔公司下属研发中心俄罗斯团队发起该项目,开源BSD证书,OpenCV的目标是实现实时计算机视觉,,是一个跨平台的计算机视觉库。从开发之日起就得到了迅猛发展,获得了众多公司和业界大牛的鼎力支持与贡献,因为是BSD开源许可,因此可以免费应用在科研和商业应用领域。OpenCV作为强大的计算机视觉开源库,很大程度上参考了MatLab的实现细节和风格,比如说,在OpenCV2.x版本以后,越来越多的函数实现了MatLab具有的功能,甚至干脆连函数名都一模一样(如imread,imshow,imwriter等)。这一做法,不仅拉近了产品开发与学术研究的距离,并极大程度的提高了开发人员的研发效率,不得不说,Intel公司真的是一个伟大的公司。在计算机内存中,数字图像以矩阵的形式存储和运算,比如,在MatLab中,图像读取之后对应一个矩阵,在OpenCV中,同样也是如此。在早期的OpenCV1.x版本中,图像的处理是通过IplImage(该名称源于Intel的另一个开源库Intel Image Processing Library,缩写成IplImage)结构来实现的。早期的OpenCV是用C语言编写,因此提供的借口也是C语言接口,其源代码完全是C的编程风格。IplImage结构是OpenCV矩阵运算的基本数据结构。到OpenCV2.x版本,OpenCV开源库引入了面向对象编程思想,大量源代码用C++重写,Mat类(Matrix的缩写)是OpenCV用于处理图像而引入的一个封装类。从功能上讲,Mat类在IplImage结构的基础上进一步增强,并且,由于引入C++高级编程特性,Mat类的扩展性大大提高,Mat类的内容在后期的版本中不断丰富,通过查看Mat类的定义,会发现其设计实现十分全面而具体,基本覆盖计算机视觉对于图像处理的基本要求。OpenCV is an open source computer vision application platform. The project was initiated by the Russian team of Intel Corporation's R&D Center. It has an open source BSD certificate. The goal of OpenCV is to achieve real-time computer vision. It is a cross-platform computer vision library. It has developed rapidly since the day it was developed, and has received strong support and contributions from many companies and industry leaders. Because it is a BSD open source license, it can be used in scientific research and commercial applications for free. As a powerful computer vision open source library, OpenCV largely refers to the implementation details and style of MatLab. For example, after OpenCV2. All are exactly the same (such as imread, imshow, imwriter, etc.). This approach not only shortens the distance between product development and academic research, but also greatly improves the R&D efficiency of developers. It has to be said that Intel is really a great company. In computer memory, digital images are stored and operated in the form of matrices. For example, in MatLab, an image corresponds to a matrix after being read, and the same is true in OpenCV. In the early versions of OpenCV1.x, image processing was implemented through the structure of IplImage (the name comes from Intel Image Processing Library, another open source library of Intel, abbreviated as IplImage). The early OpenCV was written in C language, so the interface provided is also a C language interface, and its source code is completely C programming style. The IplImage structure is the basic data structure for OpenCV matrix operations. Up to the OpenCV2.x version, the OpenCV open source library introduced the idea of object-oriented programming, and a large number of source codes were rewritten in C++. The Mat class (abbreviation for Matrix) is a package class introduced by OpenCV for image processing. In terms of function, the Mat class is further enhanced on the basis of the IplImage structure, and, due to the introduction of C++ advanced programming features, the scalability of the Mat class is greatly improved, and the content of the Mat class is continuously enriched in later versions. By viewing the Mat class Definition, you will find that its design and implementation is very comprehensive and specific, basically covering the basic requirements of computer vision for image processing.
OpenCV中已经包含如下应用领域功能:二维和三维特征工具箱、运动估算、人脸识别系统、姿势识别、人机交互、移动机器人、运动理解、对象鉴别、分割与识别、立体视觉、运动跟踪、增强现实(AR技术)。OpenCV already includes the following application domain functions: 2D and 3D feature toolbox, motion estimation, face recognition system, pose recognition, human-computer interaction, mobile robot, motion understanding, object identification, segmentation and recognition, stereo vision, motion tracking , augmented reality (AR technology).
基于上述功能实现需要,OpenCV中还包括以下基于统计学机器学习库:Boosting算法、Decision Tree(决策树)学习、Gradient Boosting算法、EM算法(期望最大化)、KNN算法、朴素贝叶斯分类、人工神经网络、随机森林、支掌向量机。Based on the above function realization needs, OpenCV also includes the following statistical machine learning libraries: Boosting algorithm, Decision Tree (decision tree) learning, Gradient Boosting algorithm, EM algorithm (expectation maximization), KNN algorithm, naive Bayesian classification, Artificial neural network, random forest, palm vector machine.
OpenCV中多数模块是基于C++实现,其中有少部分是基于C语言实现,当前OpenCV提供的SDK已经支持C++、Java、Python等语言应用开发。当前OpenCV本身新开发的算法和模块接口都是基于C++产生。OpenCV支持几乎所有主流的OS系统上应用开发,包括Windows、Mac、Linux、FreeBSD、OpenBSD等。移动平台支持Android、IOS、BlackBerray等平台。用户可以从OpenCV官方获取相关SDK下载,开发文档和环境配置信息。OpenCV自从1.0版本发布以来,立刻吸引许多公司目光,被广泛应用在许多领域的产品研发与创新上,相关应用包括卫星地图与电子地图拼接、医学中图像噪声处理、对象检测、安防监控领域安全与入侵检测、自动监视报警、制造业与工业中的产品质量检测、摄像机标定。军事领域的无人机飞行、无人驾驶与水下机器人等众多领域。Most of the modules in OpenCV are implemented based on C++, and a small part of them are implemented based on C language. The current SDK provided by OpenCV already supports C++, Java, Python and other language application development. Currently, the newly developed algorithms and module interfaces of OpenCV itself are all based on C++. OpenCV supports application development on almost all mainstream OS systems, including Windows, Mac, Linux, FreeBSD, OpenBSD, etc. The mobile platform supports Android, IOS, BlackBerray and other platforms. Users can obtain related SDK downloads, development documents and environment configuration information from the OpenCV official website. Since the release of version 1.0, OpenCV has immediately attracted the attention of many companies and has been widely used in product development and innovation in many fields. Related applications include satellite map and electronic map stitching, image noise processing in medicine, object detection, and security monitoring in the field of safety and security. Intrusion detection, automatic monitoring and alarm, product quality inspection in manufacturing and industry, camera calibration. UAV flight in the military field, unmanned driving and underwater robots and many other fields.
再进一步的技术方案为:所述特征提取与对比模块设置有特征点提取工具,该特征点提取工具设置有视频读取功能块、取帧功能块、调用AKAZE算法功能块、特征点归一化功能块、写文件功能块;A further technical solution is: the feature extraction and comparison module is provided with a feature point extraction tool, and the feature point extraction tool is provided with a video reading function block, a frame fetching function block, an AKAZE algorithm function block, and feature point normalization Function block, write file function block;
所述特征提取与对比模块内还设置有视频特征参数,其中视频特征参数包括所述相似度阈值、视频输入路径、视频输出路径。The feature extraction and comparison module is also provided with video feature parameters, wherein the video feature parameters include the similarity threshold, video input path, and video output path.
通过上述方案,特征提取与对比模块用于实现视频读取、帧读取、帧图片发送、特征接收等功能,并且通过路径实现源视频查找。Through the above solution, the feature extraction and comparison module is used to realize functions such as video reading, frame reading, frame picture sending, and feature receiving, and realize source video search through paths.
再进一步的,所述帧特征序列中帧特征的独特特征属性包括特征横坐标、特征纵坐标、特征角度、特征尺寸、特征权重、特征扩展、特征金字塔层;根据独特特征属性,可以将不同拍摄角度、不同放大尺寸等属性的相同特征提取出来,特征提取不易丢失,提取完整性好,方便不同角度。Furthermore, the unique feature attributes of frame features in the frame feature sequence include feature abscissa, feature ordinate, feature angle, feature size, feature weight, feature extension, and feature pyramid layer; The same features of attributes such as angles and different enlarged sizes are extracted. The feature extraction is not easy to lose, the extraction integrity is good, and it is convenient for different angles.
再进一步的,所述视频特征点集合中的所有基准帧特征序列按照保存的先后顺序依次排列。当待对比帧特征序列与基准帧特征序列对比后,相似度小于设定的相似度阈值,将该待对比帧特征序列保存至特征序列存储模块内,并排列在前一个基准帧特征序列后面。当在使用最终保存好的视频特征点集合后,通过任意一个未知位置的特征序列,对比后就可得到该特征序列出现的大致位置,进行定位。Still further, all reference frame feature sequences in the video feature point set are arranged in sequence according to the order of preservation. When the feature sequence of the frame to be compared is compared with the feature sequence of the reference frame, and the similarity is less than the set similarity threshold, the feature sequence of the frame to be compared is stored in the feature sequence storage module and arranged behind the feature sequence of the previous reference frame. After using the final saved video feature point set, through any feature sequence of unknown position, after comparison, the approximate position where the feature sequence appears can be obtained for positioning.
再进一步的,任一帧特征序列或者待查询图片特征序列均包括依次连接的帧序号、特征序列开始标号、特征序列内容、特征序列结束标号;为了对每一帧进行标记,设置帧序号、特征序列开始标号、特征序列结束标号,通过标记可以对每一帧进行区分,对其起始和结束进行数字化识别和标记。Furthermore, any frame feature sequence or image feature sequence to be queried all includes sequentially connected frame number, feature sequence start label, feature sequence content, and feature sequence end label; in order to mark each frame, set frame number, feature Sequence start number, feature sequence end number, each frame can be distinguished by marking, and its start and end can be digitally identified and marked.
再进一步的,所述特征序列开始标号由X个字节的整数组成;所述特征序列结束标号由Y个字节的整数组成;所述视频特征点集合起点连接有起始标号;两两所述基准帧特征序列经特征连接符连接。Still further, the start label of the feature sequence is composed of an integer of X bytes; the end label of the feature sequence is composed of an integer of Y bytes; the starting point of the video feature point set is connected with a start label; The reference frame feature sequences are connected by feature connectors.
即对于视频A1,其视频特征点集合起始标号为:视频A1特征序列。That is, for video A1, the starting label of its video feature point set is: video A1 feature sequence.
一种基于图像特征提取的以图搜影系统,其关键技术在于:包括视频获取端和图片获取端;其中,所述视频获取端与待提取存储模块连接,所述待提取存储模块与特征提取与对比模块连接,在该特征提取与对比模块内设置有帧发送单元和特征序列接收单元,在特征提取与对比模块上设置有第三方视觉库连接端,该第三方视觉库连接端用于所述帧发送单元、特征序列接收单元分别与第三方视觉库连接,在所述特征提取与对比模块上还连接有特征序列存储模块,该特征序列存储模块与压缩视频数据库连接,所述压缩视频数据库与以图搜影模块连接;A system for searching video by image based on image feature extraction, its key technology is: comprising a video acquisition terminal and a picture acquisition terminal; wherein, the video acquisition terminal is connected to a storage module to be extracted, and the storage module to be extracted is connected to the Connected with the comparison module, a frame sending unit and a feature sequence receiving unit are arranged in the feature extraction and comparison module, and a third-party visual library connection terminal is provided on the feature extraction and comparison module, and the third-party visual library connection terminal is used for all The frame sending unit and the feature sequence receiving unit are respectively connected with a third-party visual library, and a feature sequence storage module is also connected to the feature extraction and comparison module, and the feature sequence storage module is connected with a compressed video database, and the compressed video database Connect with the image search module;
所述图片获取端经图片特征提取模块与所述压缩视频数据库连接,在所述图片特征提取模块上设置有第三方视觉库连接端,用于提取待查询图片特征序列,所述图片获取端还与所述以图搜影模块连接,所述以图搜影模块与源视频库连接。The picture acquisition terminal is connected with the compressed video database through the picture feature extraction module, and a third-party visual library connection terminal is arranged on the picture feature extraction module, which is used to extract the picture feature sequence to be queried, and the picture acquisition terminal also It is connected with the image search module, and the image search module is connected with the source video library.
通过上述系统,实现对源视频的获取、保存、生成特征序列、对视频帧进行提取、视频帧删除功能,并且输出最为精简的视频特征点集合,并且通过压缩视频数据库形成视频特征的聚类簇群,结合该聚类簇群得到M个视频特征点的簇序列集合。并且通过上述系统实现图片获取、图片特征提取和归类,通过与M个视频特征点的簇序列集合进行对比后,归类对比结果,得到相似度高的图片,结合待查询图片和帧图片进行实际图片对比后,对搜影结果进一步确定,从而结合压缩视频数据库内的源视频索引获取对应的视频或者视频保存的位置,实现快速准确的视频搜索。Through the above system, the acquisition, preservation, generation of feature sequences, video frame extraction, and video frame deletion functions of the source video are realized, and the most streamlined set of video feature points is output, and the clustering of video features is formed by compressing the video database. group, combined with the clustering cluster group to obtain a cluster sequence set of M video feature points. And through the above system to achieve picture acquisition, picture feature extraction and classification, after comparing with the cluster sequence set of M video feature points, classify and compare the results, obtain pictures with high similarity, combine the pictures to be queried and frame pictures After the actual pictures are compared, the video search results are further determined, so as to obtain the corresponding video or the location where the video is saved in combination with the source video index in the compressed video database, so as to realize fast and accurate video search.
再进一步的,所述压缩视频数据库内设置有聚类单元和数据单元;Still further, the compressed video database is provided with a clustering unit and a data unit;
所述聚类单元用于M个视频特征点集合进行聚类操作得到聚类簇群和M个视频特征点的簇序列集合;并对所述待查询图片特征序列中的所有特征进行聚类,得到待查询图片簇序列;The clustering unit is used to perform a clustering operation on M sets of video feature points to obtain a cluster sequence set of cluster clusters and M video feature points; and cluster all the features in the feature sequence of the picture to be queried, Obtain the image cluster sequence to be queried;
所述数据单元中保存有M个视频特征点集合、M个视频特征点的簇序列集合、每个视频的源视频索引以及任一待查询图片特征序列以及待查询图片簇序列;所述以图搜影模块中设置有聚类对比单元、图片获取单元、帧图片获取单元、源视频获取单元;M video feature point sets, cluster sequence sets of M video feature points, source video index of each video, any picture feature sequence to be queried and picture cluster sequence to be queried are stored in the data unit; The video search module is provided with a clustering comparison unit, a picture acquisition unit, a frame picture acquisition unit, and a source video acquisition unit;
所述聚类对比单元用于将所述待查询图片簇序列与M个视频特征点的簇序列集合进行对比,得到序列相似度;所述图片获取单元用于获取待查询图片;所述帧图片获取单元用于获取序列相似度最高对应的帧图片;The cluster comparison unit is used to compare the cluster sequence of the picture to be queried with the set of cluster sequences of M video feature points to obtain sequence similarity; the picture acquisition unit is used to obtain the picture to be queried; the frame picture The obtaining unit is used to obtain the frame picture corresponding to the highest sequence similarity;
所述源视频获取单元用于获取序列相似度最高对应的源视频。The source video obtaining unit is used to obtain the source video corresponding to the highest sequence similarity.
通过上述方案,聚类单元对M个视频中的每个特征进行了聚类操作,对每个特征进行归类。数据单元用于保存所有视频的的特征数据、特征归类数据以及源视频中每个视频对应的索引。在以图搜影时,得到的通过聚类对比单元实现对获取的图片进行归类对比,从M个视频特征点的簇序列集合中寻找出最为相似的序列。Through the above solution, the clustering unit performs a clustering operation on each feature in the M videos, and classifies each feature. The data unit is used to save the feature data, feature classification data of all videos and the corresponding index of each video in the source video. When searching for images by images, the obtained pictures are classified and compared through the clustering comparison unit, and the most similar sequences are found from the cluster sequence sets of M video feature points.
本发明的有益效果:利用图像特征提取技术,对视频中的每一帧的进行特征提取的相互对比,并剔除重复的出现的帧特征序列。实现低占用内存,由于视频压缩后,数据量减少。在进行以图搜影时,基于聚类算法得到的聚类簇群,对获取的图片进行归类,通过归类结果,进行相似度确定。对比过程块,并可以对图片出现在视频的大致帧数进行定位。相对于现有技术,本发明可以加快检索速度,并且定位精确。The beneficial effect of the present invention is to use the image feature extraction technology to perform feature extraction and comparison of each frame in the video, and to eliminate repeated frame feature sequences. Realize low memory usage, due to video compression, the amount of data is reduced. When performing image search, based on the clustering clusters obtained by the clustering algorithm, the acquired pictures are classified, and the similarity is determined through the classification results. Compare the process blocks, and locate the approximate frame number where the picture appears in the video. Compared with the prior art, the invention can speed up the retrieval speed and has accurate positioning.
附图说明Description of drawings
图1是本发明的系统结构框图;Fig. 1 is a system structure block diagram of the present invention;
图2是本发明的以图搜影方法流程图;Fig. 2 is a flow chart of the method for image search with pictures of the present invention;
图3是本发明的等效视频压缩存储方法流程图;Fig. 3 is the flow chart of the equivalent video compression storage method of the present invention;
图4是MADlib开源机器学习库架构图;Figure 4 is an architecture diagram of the MADlib open source machine learning library;
图5是检测视频列表图;Figure 5 is a list of detected videos;
图6是等效视频压缩运算瞬时状态图。Fig. 6 is an instantaneous state diagram of an equivalent video compression operation.
图7是待查询图片;Figure 7 is a picture to be queried;
图8是原始帧图片搜影结果示意图;Fig. 8 is a schematic diagram of the original frame image search results;
图9是裁剪尺寸1的帧图片搜影结果示意图;Fig. 9 is a schematic diagram of the image search result of the frame image with cropping size 1;
图10是裁剪尺寸2的帧图片搜影结果示意图。FIG. 10 is a schematic diagram of image search results for frame images with a cropping size of 2. FIG.
具体实施方式Detailed ways
下面结合附图对本发明的具体实施方式以及工作原理作进一步详细说明。The specific implementation manner and working principle of the present invention will be further described in detail below in conjunction with the accompanying drawings.
结合图1和图2可以看出,一种基于图像特征提取的以图搜影方法,具体步骤包括:步骤1:对M个视频中的每一帧图片进行特征提取、压缩后,得到M个视频特征点集合;步骤2:将M个视频特征点集合送入压缩视频数据库,经聚类操作得到聚类簇群和M个视频特征点的簇序列集合;步骤3:图片获取端对待查询图片进行获取,经图片特征提取模块后得到待查询图片特征序列;并送入压缩视频数据库进行聚类,得到待查询图片簇序列;步骤4:将待查询图片簇序列与M个视频特征点的簇序列集合进行对比,得到序列相似度;并提取与待查询图片簇序列的序列相似度最高对应的视频特征点集合以及帧特征序列;步骤5:根据视频特征点集合以及帧特征序列的索引,提取对应的视频中的帧图片,并将该帧图片与待查询图片进行对比,得到图片相似度;Combining Figure 1 and Figure 2, it can be seen that an image search method based on image feature extraction, the specific steps include: Step 1: After performing feature extraction and compression on each frame of M videos, M videos are obtained. Video feature point set; Step 2: Send M video feature point sets into the compressed video database, and obtain cluster clusters and M video feature point cluster sequence sets through clustering operations; Step 3: The image acquisition terminal treats the query image Acquisition, after the image feature extraction module, the feature sequence of the picture to be queried is obtained; and sent to the compressed video database for clustering, to obtain the cluster sequence of the picture to be queried; step 4: the cluster sequence of the picture to be queried and the cluster of M video feature points Compare the sequence sets to obtain the sequence similarity; and extract the video feature point set and frame feature sequence corresponding to the highest sequence similarity of the picture cluster sequence to be queried; Step 5: According to the index of the video feature point set and frame feature sequence, extract A frame picture in the corresponding video, and compare the frame picture with the picture to be queried to obtain the picture similarity;
步骤6:若图片相似度大于图片相似度阈值,则输出对应视频或者对应视频的索引。Step 6: If the picture similarity is greater than the picture similarity threshold, output the corresponding video or the index of the corresponding video.
在本实施例中,还设置有序列相似度阈值,该序列相似度阈值为80%,则当最高序列相似度低于该阈值,则不进行下一步,显示没有检索到相关内容。当最高序列相似度大于等于该阈值,则提取与待查询图片簇序列的序列相似度最高对应的视频特征点集合以及帧特征序列。在本实施例中,图片相似度阈值为90%。结合图3可以看出,步骤1中,实现对M个视频中的每一帧图片进行特征提取、压缩后,得到M个视频特征点集合的具体步骤为:In this embodiment, a sequence similarity threshold is also set, and the sequence similarity threshold is 80%. If the highest sequence similarity is lower than the threshold, the next step will not be performed and it will be displayed that no relevant content has been retrieved. When the highest sequence similarity is greater than or equal to the threshold, extract the video feature point set and frame feature sequence corresponding to the highest sequence similarity of the image cluster sequence to be queried. In this embodiment, the picture similarity threshold is 90%. It can be seen from Fig. 3 that in step 1, after feature extraction and compression of each frame of pictures in M videos, the specific steps to obtain M video feature point sets are as follows:
步骤11:视频获取端依次从M个视频特征点获取一段视频Ax,并保存在待提取存储模块内;步骤12:待提取存储模块得到视频Ax的总帧数N,并将视频Ax中的每一帧依次发送至特征提取与对比模块;步骤13:特征提取与对比模块用于指定基准帧,并根据基准帧对视频Ax进行预处理、图像特征提取、对比、筛选、删除、压缩后,得到视频特征点集合,具体内容为:Step 11: The video acquisition terminal obtains a section of video Ax from M video feature points in turn, and stores it in the storage module to be extracted; Step 12: The storage module to be extracted obtains the total number of frames N of the video Ax, and saves each video Ax in the video Ax One frame is sent to the feature extraction and comparison module in turn; Step 13: The feature extraction and comparison module is used to specify the reference frame, and preprocess the video Ax according to the reference frame, extract image features, compare, filter, delete, and compress to obtain A collection of video feature points, the specific content is:
S131:特征提取与对比模块获取第一帧,令该第一帧为基准帧,提取第一帧图像特征后对提取的特征进行排序,得到基准帧特征序列,并将该基准帧特征序列保存在特征序列存储模块内;S132:特征提取与对比模块获取下一帧,并令该帧为待对比帧,提取待对比帧图像特征后对提取的特征进行排序,得到待对比帧特征序列;S131: The feature extraction and comparison module obtains the first frame, makes the first frame a reference frame, sorts the extracted features after extracting the image features of the first frame, obtains a reference frame feature sequence, and saves the reference frame feature sequence in In the feature sequence storage module; S132: the feature extraction and comparison module obtains the next frame, and makes the frame a frame to be compared, extracts the image features of the frame to be compared, and sorts the extracted features to obtain a feature sequence of the frame to be compared;
S133:将待对比帧特征序列与所述基准帧特征序列进行对比;S133: Compare the feature sequence of the frame to be compared with the feature sequence of the reference frame;
若二者相似度大于等于设定的相似度阈值,则抛弃该待对比帧特征序列,进入步骤S134;若相似度小于设定的相似度阈值,则令该待对比帧特征序列为新的基准帧特征序列,即:令对应的待对比帧为基准帧;并将新的基准帧特征序列依次保存在特征序列存储模块内,进入步骤S134;If the similarity between the two is greater than or equal to the set similarity threshold, then discard the feature sequence of the frame to be compared, and enter step S134; if the similarity is less than the set similarity threshold, then make the feature sequence of the frame to be compared a new benchmark Frame feature sequence, that is: let the corresponding frame to be compared be the reference frame; and store the new reference frame feature sequence in the feature sequence storage module in turn, and enter step S134;
S134:判断视频Ax中的第N帧是否已经对比完毕;若是,特征序列存储模块将保存的所有基准帧特征序列组成视频特征点集合,并输出该视频Ax的视频特征点集合;否则返回步骤S132;S134: Determine whether the Nth frame in the video Ax has been compared; if so, the feature sequence storage module forms a video feature point set with all reference frame feature sequences preserved, and outputs the video feature point set of the video Ax; otherwise, return to step S132 ;
步骤14:判断M个视频是否全部压缩完毕,若是结束步骤1,否则返回步骤11。Step 14: Determine whether all the M videos have been compressed, if so, end step 1, otherwise return to step 11.
在本实施例中,对视频Ax中每一帧图片进行预处理时,图片大小重置处理后,将图片重置成400X300的大小,并且灰度值设置为50%。在本实施例中,所述特征提取与对比模块与第三方视觉库OpenCV连接;在本实施例中,OpenCV视觉库内设置有AKAZE算法。In this embodiment, when preprocessing each frame of the picture in the video Ax, after the picture size reset process, the picture is reset to a size of 400×300, and the gray value is set to 50%. In this embodiment, the feature extraction and comparison module is connected with the third-party vision library OpenCV; in this embodiment, the OpenCV vision library is provided with the AKAZE algorithm.
在本实施例中,步骤2中,M个视频特征点集合经聚类操作得到聚类簇群的具体步骤为:In the present embodiment, in
步骤21a:从M个视频特征点集合任意选择k个特征作为初始聚类簇的聚类中心;步骤21b:分别计算M个视频特征点集合中所有特征到k个特征作为聚类簇中心的聚类距离,并设定最小聚类距离对特征进行划分;Step 21a: Select k features arbitrarily from the M video feature point sets as the cluster center of the initial cluster; Step 21b: Calculate the clustering of all features to k features in the M video feature point sets as the cluster center Class distance, and set the minimum clustering distance to divide the features;
步骤21c:选择出聚类中心发生变化的聚类簇,并计算对应聚类簇的聚类距离均值,根据聚类距离均值确定该聚类簇的聚类中心,Step 21c: Select the cluster whose cluster center changes, and calculate the mean value of the cluster distance of the corresponding cluster, and determine the cluster center of the cluster according to the mean value of the cluster distance,
步骤21d:若聚类中心不再发生变化,则终止聚类,输出k聚类簇形成的聚类簇群以及对应的聚类中心,否则返回步骤21b;Step 21d: If the cluster center does not change any more, then terminate the clustering, output the cluster cluster formed by k clusters and the corresponding cluster center, otherwise return to step 21b;
步骤2中,M个视频特征点集合经聚类操作得到M个视频特征点的簇序列集合的具体步骤为:对所述聚类簇群中的k个聚类簇进行编号;In
依次计算M个视频特征点集合中,每个特征到k个聚类簇的聚类距离;Calculate the clustering distances from each feature to k clusters in the M video feature point sets in turn;
对每个特征的k个聚类距离进行排序,并将该特征归类到最小聚类距离对应聚类簇,并获取该特征对应的聚类簇编号,直至每一个视频特征点集合中所有的特征均归类完毕后,得到M个簇序列集合。Sort the k clustering distances of each feature, and classify the feature into the cluster corresponding to the minimum clustering distance, and obtain the cluster number corresponding to the feature, until all the video feature points in each video feature point set After the features are classified, M cluster sequence sets are obtained.
在本实施例中,通过与MADlib开源机器学习库连接,实现图片特征聚类,采用的聚类方法为k-means算法。In this embodiment, image feature clustering is realized by connecting with the MADlib open source machine learning library, and the clustering method adopted is the k-means algorithm.
MADlib是Pivotal公司与伯克利大学合作开发的一个开源机器学习库,提供了多种数据转换、数据探索、统计、数据挖掘和机器学习方法,使用它能够简易地对结构化数据进行分析和挖掘。用户可以非常方便地将MADlib加载到数据库中,扩展数据库的分析功能。2015年7月MADlib成为Apache软件基金会的孵化器项目,经过两年的发展,于2017年8月毕业成为Apache顶级项目。其当前最新版本为MADlib 1.15,可以与PostgreSQL、Greenplum和HAWQ等数据库系统无缝集成。在本实施例中,在预研中采用的是集成到PostgreSQL中的插件。MADlib is an open source machine learning library jointly developed by Pivotal and the University of Berkeley. It provides a variety of data conversion, data exploration, statistics, data mining and machine learning methods. It can easily analyze and mine structured data. Users can easily load MADlib into the database to expand the analysis function of the database. In July 2015, MADlib became an incubator project of the Apache Software Foundation. After two years of development, it graduated in August 2017 and became a top Apache project. Its current latest version is MADlib 1.15, which can seamlessly integrate with database systems such as PostgreSQL, Greenplum and HAWQ. In this embodiment, a plug-in integrated into PostgreSQL is used in the pre-research.
从图4可以看出,MADlib开源机器学习库架构,处于架构最上面一层是用户接口。如前所述,用户只需通过在SQL查询语句中调用MADlib提供的函数来完成数据挖掘的工作。这里的SQL语法要与特定数据库管理系统相匹配。最底层则是Greenplum、PostgreSQL等数据库管理系统,最终由它们处理查询请求。As can be seen from Figure 4, the MADlib open source machine learning library architecture, the top layer of the architecture is the user interface. As mentioned earlier, users only need to call the functions provided by MADlib in the SQL query statement to complete the data mining work. The SQL syntax here should match the specific database management system. The bottom layer is the database management system such as Greenplum, PostgreSQL, etc., and finally they process the query request.
从图4还可以看出,MADlib系统架构自上至下由以下四个主要组件构成:Python调用SQL模板实现的驱动函数、Python实现的高级抽象层、C++实现的核心函数、C++实现的低级数据库抽象层。It can also be seen from Figure 4 that the MADlib system architecture consists of the following four main components from top to bottom: the driver function implemented by Python calling the SQL template, the high-level abstraction layer implemented by Python, the core function implemented by C++, and the low-level database implemented by C++ abstraction layer.
驱动MADlib架构的主要设计思想与Hadoop是一致的,体现为:The main design idea driving the MADlib architecture is consistent with Hadoop, reflected in:
操作数据库内的本地数据,不在多个运行时环境中进行不必要的数据移动。Manipulate local data within the database without unnecessary data movement across multiple runtime environments.
充分利用数据库引擎功能,但将数据挖掘逻辑从特定数据库的实现细节中分离出来。利用MPP无共享技术提供的并行性和可扩展性,如Greenplum或HAWQ数据库系统。执行的维护活动对Apache社区和正在进行的学术研究开放。Take full advantage of database engine capabilities, but separate data mining logic from database-specific implementation details. Take advantage of the parallelism and scalability provided by MPP shared-nothing technologies, such as Greenplum or HAWQ database systems. The maintenance activities performed are open to the Apache community and ongoing academic research.
其中,k-means算法基本思想为:k-means聚类划分方法的基本思想是:将一个给定的有N个数据记录的集合,划分到K个分组中,每一个分组就代表一个簇,K<N。而且这K个分组满足下列条件:每一个分组至少包含一个数据记录。每一个数据记录属于且仅属于一个分组。Among them, the basic idea of the k-means algorithm is: the basic idea of the k-means clustering method is: divide a given set of N data records into K groups, and each group represents a cluster. K<N. And these K groups meet the following conditions: each group contains at least one data record. Each data record belongs to one and only one group.
算法首先给出一个初始的分组,以后通过反复迭代的方法改变分组,使得每一次改进之后的分组方案都较前一次好,而所谓好的标准就是:同一分组中对象的距离越近越好(已经收敛,反复迭代至组内数据几乎无差异),而不同分组中对象的距离越远越好。The algorithm first gives an initial grouping, and then changes the grouping through repeated iterations, so that the grouping scheme after each improvement is better than the previous one, and the so-called good standard is: the closer the objects in the same grouping, the better ( has converged, repeated iterations until there is almost no difference in the data within the group), and the farther the distance between objects in different groups, the better.
k-means算法的工作原理是:The working principle of the k-means algorithm is:
首先随机从数据集中选取K个点,每个点初始地代表每个簇的中心,然后计算剩余各个样本到中心点的距离,将它赋给最近的簇,接着重新计算每一簇的平均值作为新的中心点,整个过程不断重复,如果相邻两次调整没有明显变化,说明数据聚类形成的簇已经收敛。本算法的一个特点是在每次迭代中都要考察每个样本的分类是否正确。若不正确,就要调整,在全部样本调整完后,再修改中心点,进入下一次迭代。这个过程将不断重复直到满足某个终止条件,终止条件可以是以下任何一个:没有对象被重新分配给不同的聚类。聚类中心不再发生变化。First randomly select K points from the data set, each point initially represents the center of each cluster, then calculate the distance from the remaining samples to the center point, assign it to the nearest cluster, and then recalculate the average value of each cluster As a new center point, the whole process is repeated continuously. If there is no significant change in the two adjacent adjustments, it means that the cluster formed by data clustering has converged. A characteristic of this algorithm is to examine whether the classification of each sample is correct in each iteration. If it is not correct, it needs to be adjusted. After all the samples are adjusted, the center point is modified to enter the next iteration. This process is repeated until a certain termination condition is met, which can be any of the following: No objects are reassigned to a different cluster. The cluster centers no longer change.
误差平方和局部最小。The sum of squared errors is a local minimum.
k-means算法是很典型的基于距离的聚类算法,采用距离作为相似性的评价指标,即认为两个对象的距离越近,其相似度就越大。该算法认为簇是由距离靠近的对象组成,因此把得到紧凑且独立的簇作为最终目标。k-means算法的输入是聚类个数k,以及n个数据对象,输出是满足误差最小标准的k个聚簇。其处理流程为:从n个数据对象中任意选择k个对象作为初始中心。The k-means algorithm is a typical distance-based clustering algorithm, which uses distance as the evaluation index of similarity, that is, the closer the distance between two objects, the greater the similarity. The algorithm considers that clusters are composed of close objects, so the final goal is to obtain compact and independent clusters. The input of the k-means algorithm is the number of clusters k and n data objects, and the output is k clusters that meet the minimum error standard. The processing flow is as follows: randomly select k objects from n data objects as initial centers.
计算每个对象与这些中心对象的距离,并根据最小距离对相应的对象进行划分。重新计算每个有变化聚类的均值作为新的中心。Calculate the distance of each object from these central objects and divide the corresponding objects according to the minimum distance. Recalculate the mean of each changed cluster as the new center.
循环2、3直到每个聚类不再发生变化为止。终止条件一般为最小化对象到其聚类中心的距离的平方和:收集类别大于不限于以下几类电视节目:
实施过程中,首先收集分门别类的影片,使得族心划分尽量标准客观合理,对后面生产的族心划分,尽量能够更好的区分。During the implementation process, firstly, we collect classified films, so that the classification of ethnic groups can be as standard, objective and reasonable as possible, and the classification of ethnic groups produced later can be better distinguished as much as possible.
收集类别大于不限于以下几类电视节目:新闻类,综艺类,影视剧类,赛事类,风景类,动物类,国外大片,国产大片,战争片,室内剧……The collection categories are greater than but not limited to the following types of TV programs: news, variety shows, film and television dramas, sports events, scenery, animals, foreign blockbusters, domestic blockbusters, war films, indoor dramas...
在本实施例中,步骤3中,图片获取端对待查询图片进行获取,经特征提取后得到待查询图片特征序列;并送入压缩视频数据库进行聚类,得到待查询图片簇序列的具体步骤为:In this embodiment, in
步骤31:图片获取端获取待查询图片后,将待查询图片发送至图片特征提取模块;Step 31: After the image acquisition terminal obtains the image to be queried, it sends the image to be queried to the image feature extraction module;
步骤32:图片特征提取模块结合第三方视觉库,提取到待查询图片特征序列后送入所述压缩视频数据库;Step 32: the picture feature extraction module combines the third-party visual library to extract the picture feature sequence to be queried and send it to the compressed video database;
步骤33:计算待查询图片特征序列中的所有特征聚类距离,并将该特征归类到最小聚类距离对应聚类簇,得到待查询图片特征序列对应的待查询图片簇序列。Step 33: Calculate the clustering distances of all the features in the feature sequence of the picture to be queried, and classify the feature into the cluster corresponding to the minimum clustering distance, and obtain the cluster sequence of the picture to be queried corresponding to the feature sequence of the picture to be queried.
进一步的技术方案为:所述特征提取与对比模块、图片特征提取模块均与第三方视觉库连接;特征提取与对比模块至少设置有帧发送单元和特征序列接收单元,该帧发送单元用于将所述待对比帧发送至所述第三方视觉库进行待对比帧图像特征提取,并将得到的待对比帧特征序列反馈给特征序列接收单元;A further technical solution is: the feature extraction and comparison module and the picture feature extraction module are all connected to a third-party visual library; the feature extraction and comparison module is at least provided with a frame sending unit and a feature sequence receiving unit, and the frame sending unit is used to The frame to be compared is sent to the third-party visual library for image feature extraction of the frame to be compared, and the obtained frame to be compared feature sequence is fed back to the feature sequence receiving unit;
所述图片特征提取模块设置有图片发送单元和待查询图片特征序列接收单元,该图片发送单元用于将待查询图片发送至所述第三方视觉库进行特征提取,并将得到的待查询图片特征序列反馈给待查询图片特征序列接收单元。The picture feature extraction module is provided with a picture sending unit and a picture feature sequence receiving unit to be queried, and the picture sending unit is used to send the picture to be queried to the third-party visual library for feature extraction, and obtain the feature of the picture to be queried The sequence is fed back to the image feature sequence receiving unit to be queried.
在本实施例中,所述特征提取与对比模块设置有特征点提取工具,该特征点提取工具设置有视频读取功能块、取帧功能块、调用AKAZE算法功能块、特征点归一化功能块、写文件功能块;In this embodiment, the feature extraction and comparison module is provided with a feature point extraction tool, and the feature point extraction tool is provided with a video reading function block, a frame-taking function block, an AKAZE algorithm function block, and a feature point normalization function block, write file function block;
所述特征提取与对比模块内还设置有视频特征参数,其中视频特征参数包括所述相似度阈值、视频输入路径、视频输出路径。其中,采用编写的shell脚本来依次调用特征点提取工具,根据特征点提取工具结合相似度阈值为90%、视频输入路径、视频输出路径进行特征对比和输入输出。其中特征提取时,运行环境在docker容器,并结合编写的shell脚本依次调用特征点提取工具开始特征提取。The feature extraction and comparison module is also provided with video feature parameters, wherein the video feature parameters include the similarity threshold, video input path, and video output path. Among them, the written shell script is used to call the feature point extraction tool in turn, and the feature point comparison and input and output are performed according to the feature point extraction tool combined with a similarity threshold of 90%, video input path, and video output path. During the feature extraction, the operating environment is in the docker container, and combined with the written shell script, the feature point extraction tool is sequentially invoked to start feature extraction.
在本实施例中,所述第三方视觉库为OpenCV视觉库。在本实施例中,OpenCV视觉库内设置有AKAZE算法。在本实施例中,将特征点提取后输出为指定格式的csv文件;然后打包成jar文件。In this embodiment, the third-party vision library is the OpenCV vision library. In this embodiment, the AKAZE algorithm is set in the OpenCV vision library. In this embodiment, the extracted feature points are output as a csv file in a specified format; and then packaged into a jar file.
在本实施例中,所述帧特征序列中帧特征的独特特征属性包括特征横坐标、特征纵坐标、特征角度、特征尺寸、特征权重、特征扩展、特征金字塔层;In this embodiment, the unique feature attributes of frame features in the frame feature sequence include feature abscissa, feature ordinate, feature angle, feature size, feature weight, feature extension, and feature pyramid layer;
视频特征点集合中的所有基准帧特征序列按照保存的先后顺序依次排列;All reference frame feature sequences in the video feature point set are arranged in sequence according to the order of preservation;
任一帧特征序列或者待查询图片特征序列均包括依次连接的帧序号、特征序列开始标号、特征序列内容、特征序列结束标号;Any frame feature sequence or image feature sequence to be queried includes sequentially connected frame number, feature sequence start label, feature sequence content, and feature sequence end label;
特征序列开始标号由X个字节的整数组成;所述特征序列结束标号由Y个字节的整数组成;在本实施例中,X=Y=4。The start number of the characteristic sequence is composed of an integer of X bytes; the end number of the characteristic sequence is composed of an integer of Y bytes; in this embodiment, X=Y=4.
所述视频特征点集合起点连接有起始标号;两两所述基准帧特征序列经特征连接符连接。The starting points of the set of video feature points are connected with a start label; two pairs of the reference frame feature sequences are connected by feature connectors.
则视频Ax形成的特征序列为:Then the feature sequence formed by the video Ax is:
视频Ax特征序列+特征连接符+第1帧的帧序号+第1帧特征序列开始标号+第1帧特征序列内容+第1帧特征序列结束标号+特征连接符+第i1帧的帧序号+第i1帧特征序列开始标号+第i1帧特征序列内容+第i1帧特征序列结束标号+特征连接符+第i2帧的帧序号+第i2帧特征序列开始标号+第i2帧特征序列内容+第i2帧特征序列结束标号+……其中i1、i2为大于1的整数,且i1>i2。Video Ax feature sequence + feature connector + frame number of the first frame + start label of the first frame feature sequence + content of the first frame feature sequence + end label of the first frame feature sequence + feature connector + frame number of the i1th frame + The start label of the i1th frame feature sequence + the i1th frame feature sequence content + the i1th frame feature sequence end label + feature connector + the frame number of the i2th frame + the i2th frame feature sequence start label + the i2th frame feature sequence content + the i2th frame feature sequence content + the i2th frame i2 frame feature sequence end label + ... where i1 and i2 are integers greater than 1, and i1>i2.
在本实施例中,所述视频特征点压缩包为经过gzip压缩行程的二进制序列的存储文件。结合图1可以看出,一种基于图像特征提取的以图搜影系统,包括视频获取端和图片获取端;In this embodiment, the video feature point compression package is a storage file of a binary sequence that has undergone a gzip compression process. Combining with Figure 1, it can be seen that an image search system based on image feature extraction includes a video acquisition terminal and an image acquisition terminal;
其中,所述视频获取端与待提取存储模块连接,所述待提取存储模块与特征提取与对比模块连接,在该特征提取与对比模块内设置有帧发送单元和特征序列接收单元,在特征提取与对比模块上设置有第三方视觉库连接端,该第三方视觉库连接端用于所述帧发送单元、特征序列接收单元分别与第三方视觉库连接,在所述特征提取与对比模块上还连接有特征序列存储模块,该特征序列存储模块与压缩视频数据库连接,所述压缩视频数据库与以图搜影模块连接;Wherein, the video acquisition terminal is connected with the storage module to be extracted, the storage module to be extracted is connected with the feature extraction and comparison module, and a frame sending unit and a feature sequence receiving unit are arranged in the feature extraction and comparison module, and the feature extraction The comparison module is provided with a third-party visual library connection end, which is used for the frame sending unit and the feature sequence receiving unit to connect with the third-party visual library respectively, and on the feature extraction and comparison module. A feature sequence storage module is connected, and the feature sequence storage module is connected with a compressed video database, and the compressed video database is connected with a picture search module;
所述图片获取端经图片特征提取模块与压缩视频数据库连接,在图片特征提取模块上设置有第三方视觉库连接端,用于提取待查询图片特征序列,所述图片获取端还与以图搜影模块连接,以图搜影模块与源视频库连接。The picture acquisition terminal is connected with the compressed video database through the picture feature extraction module, and a third-party visual library connection terminal is arranged on the picture feature extraction module, which is used to extract the picture feature sequence to be queried. The video module is connected, and the image search module is connected with the source video library.
结合图1还可以看出,所述压缩视频数据库内设置有聚类单元和数据单元;It can also be seen in conjunction with Fig. 1 that a clustering unit and a data unit are arranged in the compressed video database;
所述聚类单元用于M个视频特征点集合进行聚类操作得到聚类簇群和M个视频特征点的簇序列集合;并对所述待查询图片特征序列中的所有特征进行聚类,得到待查询图片簇序列;The clustering unit is used to perform a clustering operation on M sets of video feature points to obtain a cluster sequence set of cluster clusters and M video feature points; and cluster all the features in the feature sequence of the picture to be queried, Obtain the image cluster sequence to be queried;
数据单元中保存有M个视频特征点集合、M个视频特征点的簇序列集合、每个视频的源视频索引以及任一待查询图片特征序列以及待查询图片簇序列;In the data unit, there are M video feature point sets, M video feature point cluster sequence sets, the source video index of each video, any picture feature sequence to be queried, and the picture cluster sequence to be queried;
所述以图搜影模块中设置有聚类对比单元、图片获取单元、帧图片获取单元、源视频获取单元;所述聚类对比单元用于将所述待查询图片簇序列与M个视频特征点的簇序列集合进行对比,得到序列相似度;The described image search module is provided with a cluster comparison unit, a picture acquisition unit, a frame picture acquisition unit, and a source video acquisition unit; the cluster comparison unit is used to compare the picture cluster sequence to be queried with M video features The cluster sequence sets of points are compared to obtain the sequence similarity;
所述图片获取单元用于获取待查询图片;The picture acquisition unit is used to acquire the picture to be queried;
所述帧图片获取单元用于获取序列相似度最高对应的帧图片;The frame picture acquisition unit is used to acquire the frame picture corresponding to the highest sequence similarity;
所述源视频获取单元用于获取序列相似度最高对应的源视频。The source video obtaining unit is used to obtain the source video corresponding to the highest sequence similarity.
为进一步说明本实施方式,采用获取电视栏目《大声说出来》的18个视频和随机获取的任意两个小视频作为源视频。To further illustrate this embodiment, 18 videos from the TV program "Speak Out Loud" and any two small videos obtained randomly are used as source videos.
其中,《大声说出来》的18个视频下载地址为:http://qjcq.cbg.cn/dsscl/1.shtml。Among them, the 18 video download addresses of "Speak Out Loud" are: http://qjcq.cbg.cn/dsscl/1.shtml.
利用爬虫技术,采集下载上述网页中第一页中的部分视频,共18个。其中,大部分视频30分钟左右,有一个45分钟,还有一个1小时时长。另外两个视频大概在3-5分钟左右。共计20个视频。视频列表图详见附图5。Using crawler technology to collect and download some of the videos on the first page of the above-mentioned webpages, a total of 18 videos. Among them, most of the videos are about 30 minutes long, one is 45 minutes long, and the other is 1 hour long. The other two videos are about 3-5 minutes long. There are 20 videos in total. See attached
结合图3,命名说明:其中170101001和170101002为3-5分钟短视频,其余名称为《大声说出来》某期视频的播出时间+编号组成。如170524001表示的是17年5月24日播出的按照自然顺序的一期节目《猜疑》。其余命名皆符合上述规则。上述视频文件总大小为:2.49GBCombined with Figure 3, naming instructions: 170101001 and 170101002 are short videos of 3-5 minutes, and the rest of the names are composed of the broadcast time + number of a certain episode of "Speak Out Loud". For example, 170524001 represents a program "Suspicion" broadcast on May 24, 2017 according to the natural order. The rest of the names are in accordance with the above rules. The total size of the above video files is: 2.49GB
将等效视频压缩存储系统安装在机器配置为2U、4核、共16线程、32G内存的上位机内,采用cpu进行的提取,算力有限,运算过程中平均负载在30%+。运算瞬时状态详见图6。对整个视频进行特征点提取后,得到的csv文件所占用磁盘空间为1.4GB。再打包成tar.gz文件,最终得到的视频特征点压缩包所占用磁盘空间为480MB。The equivalent video compression storage system is installed in a host computer with a machine configuration of 2U, 4 cores, 16 threads in total, and 32G memory. The CPU is used for extraction, and the computing power is limited, and the average load during the operation is 30%+. See Figure 6 for details on the instantaneous state of the operation. After extracting feature points from the entire video, the resulting csv file takes up 1.4GB of disk space. It is then packaged into a tar.gz file, and the disk space occupied by the finally obtained video feature point compression package is 480MB.
从图7可以看出,是待检查图片示意图,其中根据现实中侵权视频的侵权特征:大多数都是将源视频进行部分裁剪,甚至添加自己的logo信息。而将源视频信息作为侵权视频画面的一部分的情况还是比较少的。在本市实施例中,测试过程主要是对视频中的某一帧进行原始帧截取:test01;裁剪尺寸1:test02;,裁剪尺寸2:test05三种情况进行验证。As can be seen from Figure 7, it is a schematic diagram of pictures to be inspected. According to the infringement characteristics of the infringing videos in reality: most of them cut part of the source video and even add their own logo information. However, it is relatively rare to use the source video information as a part of the infringing video screen. In the embodiment of this city, the test process is mainly to perform verification on a certain frame in the video: original frame interception: test01; cropping size 1: test02; cropping size 2: test05.
验证结果详见图8-10,经过上述的测试我们可以看到,随着裁剪尺寸的加大,能搜索出来的结果会越来越少。因为我们实现计算特征点的时候考虑了侵权视频的侵权特征,所以对源视频也进行了裁剪,然后再提取特征点。目前上述展示所用的版本是,3分钟*2段*4种缩放尺度*80%帧开发并测试出炉(共4W帧),最大兼容裁剪上下左右尺寸为10%10%20%。The verification results are shown in Figure 8-10. After the above tests, we can see that as the cropping size increases, fewer and fewer results can be found. Because we considered the infringement features of the infringing video when calculating the feature points, we also cropped the source video and then extracted the feature points. The current version used for the above display is 3 minutes * 2 segments * 4 zoom scales * 80% frames developed and tested (total 4W frames), the maximum compatible cropping top, bottom, left, and right dimensions are 10% 10% 20%.
应当指出的是,上述说明并非是对本发明的限制,本发明也并不仅限于上述举例,本技术领域的普通技术人员在本发明的实质范围内所做出的变化、改性、添加或替换,也应属于本发明的保护范围。It should be noted that the above description is not intended to limit the present invention, and the present invention is not limited to the above-mentioned examples. Those skilled in the art may make changes, modifications, additions or replacements within the scope of the present invention. It should also belong to the protection scope of the present invention.
Claims (9)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910095245.XA CN111506772B (en) | 2019-01-31 | 2019-01-31 | Image searching method and system based on image feature extraction |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201910095245.XA CN111506772B (en) | 2019-01-31 | 2019-01-31 | Image searching method and system based on image feature extraction |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| CN111506772A CN111506772A (en) | 2020-08-07 |
| CN111506772B true CN111506772B (en) | 2023-05-09 |
Family
ID=71874021
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201910095245.XA Active CN111506772B (en) | 2019-01-31 | 2019-01-31 | Image searching method and system based on image feature extraction |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN111506772B (en) |
Families Citing this family (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN112100412B (en) * | 2020-09-01 | 2024-04-12 | 深圳市欢太科技有限公司 | Image retrieval method, device, computer equipment and storage medium |
| CN112861875B (en) * | 2021-01-20 | 2022-10-04 | 西南林业大学 | Method for distinguishing different wood products |
| CN116128533B (en) * | 2023-03-06 | 2023-07-28 | 广西螺霸王食品科技有限公司 | Food sales data management system |
| CN116881485B (en) * | 2023-06-19 | 2024-07-12 | 北京百度网讯科技有限公司 | Method and device for generating image retrieval index, electronic equipment and medium |
Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2010015286A (en) * | 2008-07-02 | 2010-01-21 | Olympus Corp | Server, terminal device, program, information storage medium, image retrieval method, and information presentation method |
| CN102821323A (en) * | 2012-08-01 | 2012-12-12 | 成都理想境界科技有限公司 | Video playing method, video playing system and mobile terminal based on augmented reality technique |
| CN103294813A (en) * | 2013-06-07 | 2013-09-11 | 北京捷成世纪科技股份有限公司 | Sensitive image search method and device |
| CN105653700A (en) * | 2015-03-13 | 2016-06-08 | Tcl集团股份有限公司 | Video search method and system |
| CN105917359A (en) * | 2013-10-21 | 2016-08-31 | 微软技术许可有限责任公司 | Mobile video search |
| CN106021321A (en) * | 2016-05-06 | 2016-10-12 | 成都索贝数码科技股份有限公司 | Picture based online real-time video search method |
| CN107220585A (en) * | 2017-03-31 | 2017-09-29 | 南京邮电大学 | A kind of video key frame extracting method based on multiple features fusion clustering shots |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP3311077B2 (en) * | 1993-05-06 | 2002-08-05 | 松下電器産業株式会社 | Image retrieval device |
| FR2825817B1 (en) * | 2001-06-07 | 2003-09-19 | Commissariat Energie Atomique | IMAGE PROCESSING METHOD FOR THE AUTOMATIC EXTRACTION OF SEMANTIC ELEMENTS |
| FR2835940B1 (en) * | 2002-02-08 | 2004-04-23 | Thomson Licensing Sa | METHOD FOR SEARCHING FOR THE NEAREST NEIGHBORS OF A REQUESTED VECTOR |
| US9201958B2 (en) * | 2013-10-24 | 2015-12-01 | TCL Research America Inc. | Video object retrieval system and method |
| GB201405045D0 (en) * | 2014-03-21 | 2014-05-07 | Secr Defence | Recognition of objects within a video |
| JP6395506B2 (en) * | 2014-08-22 | 2018-09-26 | キヤノン株式会社 | Image processing apparatus and method, program, and imaging apparatus |
| CN105678244B (en) * | 2015-12-31 | 2018-12-18 | 北京理工大学 | A kind of near video search method based on improved edit-distance |
| CN106503633A (en) * | 2016-10-10 | 2017-03-15 | 上海电机学院 | The method for building up in face characteristic storehouse in a kind of video image |
| CN107392238B (en) * | 2017-07-12 | 2021-05-04 | 华中师范大学 | Outdoor plant knowledge expansion learning system based on mobile visual search |
-
2019
- 2019-01-31 CN CN201910095245.XA patent/CN111506772B/en active Active
Patent Citations (7)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP2010015286A (en) * | 2008-07-02 | 2010-01-21 | Olympus Corp | Server, terminal device, program, information storage medium, image retrieval method, and information presentation method |
| CN102821323A (en) * | 2012-08-01 | 2012-12-12 | 成都理想境界科技有限公司 | Video playing method, video playing system and mobile terminal based on augmented reality technique |
| CN103294813A (en) * | 2013-06-07 | 2013-09-11 | 北京捷成世纪科技股份有限公司 | Sensitive image search method and device |
| CN105917359A (en) * | 2013-10-21 | 2016-08-31 | 微软技术许可有限责任公司 | Mobile video search |
| CN105653700A (en) * | 2015-03-13 | 2016-06-08 | Tcl集团股份有限公司 | Video search method and system |
| CN106021321A (en) * | 2016-05-06 | 2016-10-12 | 成都索贝数码科技股份有限公司 | Picture based online real-time video search method |
| CN107220585A (en) * | 2017-03-31 | 2017-09-29 | 南京邮电大学 | A kind of video key frame extracting method based on multiple features fusion clustering shots |
Also Published As
| Publication number | Publication date |
|---|---|
| CN111506772A (en) | 2020-08-07 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Suprem et al. | Odin: Automated drift detection and recovery in video analytics | |
| CN111506772B (en) | Image searching method and system based on image feature extraction | |
| JP6005837B2 (en) | Image analysis apparatus, image analysis system, and image analysis method | |
| US11157550B2 (en) | Image search based on feature values | |
| US9053386B2 (en) | Method and apparatus of identifying similar images | |
| CN110059807B (en) | Image processing method, device and storage medium | |
| US10482122B2 (en) | System and method for multiresolution and multitemporal image search | |
| AU2015220463B2 (en) | Systems and methods for image-Feature-Based Recognition | |
| CN113177518B (en) | Vehicle re-identification method based on weak supervision area recommendation | |
| Qian et al. | Image location estimation by salient region matching | |
| US11354936B1 (en) | Incremental clustering for face recognition systems | |
| CN114419408A (en) | Target re-identification method, terminal device and computer-readable storage medium | |
| CN111178455B (en) | Image clustering method, system, device and medium | |
| CN112257801A (en) | Incremental clustering method and device for images, electronic equipment and storage medium | |
| TWI769665B (en) | Target data updating method, electronic equipment and computer readable storage medium | |
| CN112084812B (en) | Image processing method, device, computer equipment and storage medium | |
| Yao et al. | Spatio-temporal information for human action recognition | |
| CN108780457A (en) | Multiple queries are executed in steady video search and search mechanism | |
| CN115115825A (en) | Method and device for detecting object in image, computer equipment and storage medium | |
| CN112069331A (en) | Data processing method, data retrieval method, data processing device, data retrieval device, data processing equipment and storage medium | |
| CN107609104B (en) | Method and system for searching related video according to video image material | |
| CN111510724A (en) | Equivalent video compression storage method and system based on image feature extraction | |
| Chen et al. | Photo retrieval based on spatial layout with hardware acceleration for mobile devices | |
| CN115146103A (en) | Image retrieval method, image retrieval apparatus, computer device, storage medium, and program product | |
| CN115455083A (en) | Duplicate checking method and device, electronic equipment and computer storage medium |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| GR01 | Patent grant | ||
| GR01 | Patent grant | ||
| TR01 | Transfer of patent right |
Effective date of registration: 20231010 Address after: No. 12-4, No. 789 Jingwei Avenue, Yuzhong District, Chongqing, 400000 Patentee after: CHONGQING XIAOXI INTELLIGENT TECHNOLOGY CO.,LTD. Address before: Room 02B-387, Building B (2nd Floor), No. A28 Information Road, Haidian District, Beijing, 100089 Patentee before: Beijing Xiaoxi Intelligent Technology Center (limited partnership) |
|
| TR01 | Transfer of patent right |