CN116721370A - Capture and identification method of interesting animal behavior clips for long-term video surveillance - Google Patents
Capture and identification method of interesting animal behavior clips for long-term video surveillance Download PDFInfo
- Publication number
- CN116721370A CN116721370A CN202310510516.XA CN202310510516A CN116721370A CN 116721370 A CN116721370 A CN 116721370A CN 202310510516 A CN202310510516 A CN 202310510516A CN 116721370 A CN116721370 A CN 116721370A
- Authority
- CN
- China
- Prior art keywords
- video
- motion
- frame
- information
- long
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/41—Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/82—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/40—Scenes; Scene-specific elements in video content
- G06V20/46—Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/70—Labelling scene content, e.g. deriving syntactic or semantic representations
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A40/00—Adaptation technologies in agriculture, forestry, livestock or agroalimentary production
- Y02A40/70—Adaptation technologies in agriculture, forestry, livestock or agroalimentary production in livestock or poultry
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Evolutionary Computation (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Databases & Information Systems (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Image Analysis (AREA)
Abstract
Description
技术领域Technical field
本发明属于人工智能技术领域,涉及一种视频内容分析方法,具体涉及一种针对长时视频监控的感兴趣动物行为片段捕获与识别方法。The invention belongs to the field of artificial intelligence technology, relates to a video content analysis method, and specifically relates to a method for capturing and identifying interesting animal behavior segments for long-term video monitoring.
背景技术Background technique
随着视频处理技术的不断发展,以及各种直播媒体流的涌现,视频内容分析方法已经引起了学术界和工业界的兴趣。目前比较成熟的产品已成功应用于各类监控系统,实现自动检测和自动报警等功能,然而对于动物监控视频,由于其数据稀少且种类繁多,目前主流的动物媒体节目或直播仍需要人工分析剪辑,使用人工智能算法进行事件检测仍是领域内具有挑战的问题,具有以下技术难点:With the continuous development of video processing technology and the emergence of various live media streams, video content analysis methods have attracted interest from academia and industry. At present, relatively mature products have been successfully used in various monitoring systems to realize functions such as automatic detection and automatic alarm. However, for animal monitoring videos, due to the scarcity of data and the wide variety of data, current mainstream animal media programs or live broadcasts still require manual analysis and editing. , using artificial intelligence algorithms for event detection is still a challenging problem in the field, with the following technical difficulties:
1、对于24小时直播信号,监控视频存在较长时间的冗余片段,且采集出的画面存在光线、阴影、遮挡等影响,视频画面质量参差不齐。这限制了算法在复杂日常视频中的部署与应用。1. For 24-hour live broadcast signals, surveillance videos have redundant segments for a long time, and the collected images are affected by light, shadow, occlusion, etc., and the video image quality is uneven. This limits the deployment and application of the algorithm in complex daily videos.
2、目前国内外对于运动分析与理解研究目标主要基于人类,对于动物的运动分析尚缺乏成熟的技术与数据支撑。2. At present, the research goals of motion analysis and understanding at home and abroad are mainly based on humans, and there is still a lack of mature technology and data support for animal motion analysis.
3、同一种动作在不同动物种类中的表现形式不同,如“老虎跑动”与“袋鼠跑动”,导致动物的动作被单独划分为一类,使得需要检测的动作总类别数目剧增,大大增加了动作的分类的难度。3. The same action has different manifestations in different animal species, such as "tiger running" and "kangaroo running", which results in the animal's actions being divided into a separate category, resulting in a sharp increase in the total number of action categories that need to be detected. This greatly increases the difficulty of classifying actions.
发明内容Contents of the invention
为了克服现有技术存在的上述不足,本发明提供了一种针对长时视频监控的感兴趣动物行为片段捕获与识别方法。该方法能够帮助媒体从业人员获取动物运动素材,减少媒体从业人员对大量视频进行二次编辑的工作量。In order to overcome the above-mentioned shortcomings of the existing technology, the present invention provides a method for capturing and identifying animal behavior fragments of interest for long-term video surveillance. This method can help media practitioners obtain animal movement materials and reduce the workload of media practitioners in secondary editing of a large number of videos.
本发明的目的是通过以下技术方案实现的:The purpose of the present invention is achieved through the following technical solutions:
一种针对长时视频监控的感兴趣动物行为片段捕获与识别方法,包括如下步骤:A method for capturing and identifying animal behavior clips of interest for long-term video surveillance, including the following steps:
步骤S1:获取视频直播流视频数据;Step S1: Obtain live video streaming video data;
步骤S2:提出基于运动区域关注的动作敏感检测器,利用动作敏感检测器发现具有意义的视频片段与素材,同时将语义表达显著的目标所在镜头中位置进行定位;Step S2: Propose a motion-sensitive detector based on motion area attention, use the motion-sensitive detector to discover meaningful video clips and materials, and at the same time locate the location of the target with significant semantic expression in the shot;
步骤S3:输出处理后的运动区域列表,写入日志;Step S3: Output the processed motion area list and write it into the log;
步骤S4:构建SlowFast的语义分析模型,所述SlowFast的语义分析模型包括慢通道和快通道,其中:Step S4: Construct a SlowFast semantic analysis model. The SlowFast semantic analysis model includes slow channel and fast channel, where:
所述慢通道用于提取空间语义信息,其输入是低帧率的采样视频帧;The slow channel is used to extract spatial semantic information, and its input is a low frame rate sampled video frame;
所述快通道用于提取时序语义信息,其输入是高帧率的采样视频帧;The fast channel is used to extract temporal semantic information, and its input is high frame rate sampled video frames;
所述快通道的输出结果通过侧向连接送入慢通道,实现两个通道的信息交互,产生更有效的视频语义分析结果;The output results of the fast channel are sent to the slow channel through the lateral connection, realizing information interaction between the two channels and producing more effective video semantic analysis results;
步骤S5:对步骤S1中的视频数据进行分割获取一个或多个视频片段,并记录所述一个或多个视频片段的时间戳信息;Step S5: Segment the video data in step S1 to obtain one or more video clips, and record the timestamp information of the one or more video clips;
步骤S6:对所述一个或多个视频片段进行抽帧处理,获取与单个视频片段对应的视频帧组;Step S6: Perform frame extraction processing on the one or more video clips to obtain a video frame group corresponding to a single video clip;
步骤S7:将所述一个或多个视频片段对应的视频帧组输入到步骤S4构建好的SlowFast模型中,以获取各个视频片段的语义;Step S7: Input the video frame group corresponding to the one or more video clips into the SlowFast model constructed in step S4 to obtain the semantics of each video clip;
步骤S8:通过语义信息,将步骤S3中捕获到的运动画面分类为背景以及各种素材类别,输出发生运动的不同时间戳信息,写入日志,并且得到动物动作剪辑,提交信息至后端服务器,以便于后期处理与调度;Step S8: Use semantic information to classify the moving pictures captured in step S3 into background and various material categories, output different timestamp information of the movement, write it into the log, obtain animal action clips, and submit the information to the back-end server. , to facilitate post-processing and scheduling;
步骤S9:根据步骤S8的运动类别结果驱动摄像头转动。Step S9: Drive the camera to rotate according to the motion category result of step S8.
相比于现有技术,本发明具有如下优点:Compared with the existing technology, the present invention has the following advantages:
1、本发明提出的运动敏感检测器能够高效地识别视频语义,过滤其他冗余信息,并且具有较快的运算速度与实时性。对于捕获的动作片段,运动敏感检测器能达到95%以上的召回率。1. The motion-sensitive detector proposed by the present invention can efficiently identify video semantics, filter other redundant information, and has fast computing speed and real-time performance. For captured action clips, the motion-sensitive detector can achieve a recall rate of more than 95%.
2、本发明提出的基于SlowFast的深度视频语义分析模型的Top-5精度能够达到90%以上。2. The Top-5 accuracy of the deep video semantic analysis model based on SlowFast proposed by this invention can reach more than 90%.
3、本发明能够帮助媒体从业人员获取动物运动素材,减少媒体从业人员对大量视频进行二次编辑的工作量,为后续动物运动捕获与行为识别提供一种新的思路。3. The present invention can help media practitioners obtain animal motion materials, reduce the workload of media practitioners in secondary editing of a large number of videos, and provide a new idea for subsequent animal motion capture and behavior recognition.
附图说明Description of the drawings
图1是本发明针对长时视频监控的感兴趣动物行为片段捕获与识别方法的流程缩略图;Figure 1 is a process thumbnail of the method for capturing and identifying animal behavior clips of interest for long-term video monitoring according to the present invention;
图2是基于运动区域关注的动作敏感检测器结构图;Figure 2 is a structural diagram of a motion-sensitive detector based on motion area attention;
图3是基于SlowFast的视频语义分析网络结构示意图;Figure 3 is a schematic diagram of the video semantic analysis network structure based on SlowFast;
图4是噪声抑制与运动区域关联示意图;Figure 4 is a schematic diagram of the relationship between noise suppression and motion areas;
图5是感兴趣动物行为片段捕获与识别结果示意图。Figure 5 is a schematic diagram of the capture and recognition results of behavioral fragments of animals of interest.
具体实施方式Detailed ways
下面结合附图对本发明的技术方案作进一步的说明,但并不局限于此,凡是对本发明技术方案进行修改或者等同替换,而不脱离本发明技术方案的精神和范围,均应涵盖在本发明的保护范围中。The technical solution of the present invention will be further described below in conjunction with the accompanying drawings, but it is not limited thereto. Any modification or equivalent replacement of the technical solution of the present invention without departing from the spirit and scope of the technical solution of the present invention shall be covered by the present invention. within the scope of protection.
本发明提供了一种针对长时视频监控的感兴趣动物行为片段捕获与识别方法,完成对视频信息的采集、传输分发、生产制作等环节的人工智能关键技术研究和开发,通过图像处理算法与深度学习算法构建智能化的视频语义分析系统。该系统对海量的视频数据,智能地获取视频中动物运动元素,并准确定位其中的动物目标。本系统得到的运动目标位置信息以及运动语义,可以应用于其他下游任务,如驱动监控摄像头移动、智能剪辑推送等,节省媒体从业者二次编辑的工作量。如图1所示,所述方法包括如下步骤:The present invention provides a method for capturing and identifying animal behavior fragments of interest for long-term video monitoring, and completes the research and development of key artificial intelligence technologies in the collection, transmission and distribution, production and production of video information. Through image processing algorithms and Deep learning algorithms build intelligent video semantic analysis systems. The system intelligently obtains animal movement elements from massive video data and accurately locates animal targets in the video. The moving target position information and motion semantics obtained by this system can be applied to other downstream tasks, such as driving surveillance camera movement, intelligent clip push, etc., saving media practitioners the workload of secondary editing. As shown in Figure 1, the method includes the following steps:
步骤S1:获取视频直播流视频数据。Step S1: Obtain live video streaming video data.
本步骤中,通过调用提供的接口获取直播视频流。In this step, obtain the live video stream by calling the provided interface.
步骤S2:提出基于运动区域关注的动作敏感检测器,发现具有意义的视频片段与素材,同时将语义表达显著的目标所在镜头中位置进行定位。如图2所示,具体步骤如下:Step S2: Propose a motion-sensitive detector based on motion area attention to discover meaningful video clips and materials, and at the same time locate the location of the target with significant semantic expression in the shot. As shown in Figure 2, the specific steps are as follows:
步骤S21:进行灰度图像处理。Step S21: Perform grayscale image processing.
本步骤中,使用OpenCV-python库进行图像灰度化处理。In this step, the OpenCV-python library is used for image grayscale processing.
步骤S22:使用基于现有的KNN模型的前景检测,去除小变化噪声。具体步骤如下:Step S22: Use foreground detection based on the existing KNN model to remove small change noise. Specific steps are as follows:
步骤S221:对于图像某个位置的新像素值,与该像素值历史信息(包括前几帧的像素值和像素点是前景还是背景的判断)比较,如果像素值之间的差别在设定阈值(需要根据场景与分辨率进行调试)内,则认为新像素值与该历史信息是匹配的,是“潜在的”一类;反之,则不是“潜在的”一类;Step S221: Compare the new pixel value at a certain position of the image with the historical information of the pixel value (including the pixel values of previous frames and the determination of whether the pixel is foreground or background). If the difference between the pixel values is within the set threshold (needs to be debugged according to the scene and resolution), the new pixel value is considered to match the historical information and is a "potential" category; otherwise, it is not a "potential" category;
步骤S222:在步骤S221中的所有历史信息比较完毕后,如果与历史信息匹配的次数超过了设定阈值,那么新像素点被归为“潜在背景点”,如果被匹配的历史信息中属于背景的点个数超过设定阈值,那么新的像素点就被归为背景点;Step S222: After all historical information in step S221 is compared, if the number of matches with historical information exceeds the set threshold, then the new pixel is classified as a "potential background point". If the matched historical information belongs to the background The number of points exceeds the set threshold, then the new pixel points are classified as background points;
步骤S223:将新像素点更新至历史信息中。Step S223: Update new pixels into historical information.
步骤S23:使用膨胀腐蚀算子、区域绘制等方式使得临近运动区域连接在一起,并保存检测出运动的区域列表。Step S23: Use dilation corrosion operator, area drawing, etc. to connect adjacent motion areas together, and save a list of areas where motion is detected.
本步骤中,在使用KNN寻找到前景后,使用图像边界提取算法找到前景图像的边界,由于光线等扰动影响,提取的边界不仅包含目标物体边界,同时也包含噪声边界,因此将噪声边界抑制,并将真正的运动区域关联是重要的。本步骤主要通过膨胀腐蚀算子进行去噪、并使用区域绘制等方式使得运动区域产生关联,可视化效果如图4所示。图4中,左上为初步绘制运动区域,右上为第一次去噪,左下为运动区域关联结果,右下为检测框在图片上的结果。In this step, after using KNN to find the foreground, use the image boundary extraction algorithm to find the boundary of the foreground image. Due to the influence of light and other disturbances, the extracted boundary not only contains the target object boundary, but also contains the noise boundary, so the noise boundary is suppressed. And correlating the true areas of motion is important. This step mainly uses the expansion corrosion operator to denoise, and uses methods such as area drawing to correlate the moving areas. The visualization effect is shown in Figure 4. In Figure 4, the upper left is the preliminary drawing of the motion area, the upper right is the first denoising, the lower left is the correlation result of the motion area, and the lower right is the result of the detection frame on the picture.
步骤S24:为防止由于光照剧烈变化引起的异常帧,敏感运动检测器通过判断关联帧之间的运动梯度、关联帧面积变化斜率、帧运动面积与关联帧平均运动区域面积的比值以及帧区域数目变化幅度来检测异常变化帧。Step S24: In order to prevent abnormal frames caused by drastic changes in illumination, the sensitive motion detector determines the motion gradient between associated frames, the associated frame area change slope, the ratio of the frame motion area to the associated frame average motion area area, and the number of frame areas. Change amplitude to detect abnormally changing frames.
步骤S25:如果S24中出现异常变化帧,则认为该帧为非动作帧;反之,则认为是正常帧,该帧是否为非动作帧取决于S23中得到的检测结果。Step S25: If an abnormally changing frame appears in S24, the frame is considered a non-action frame; otherwise, the frame is considered a normal frame. Whether the frame is a non-action frame depends on the detection result obtained in S23.
步骤S3:输出处理后的运动区域列表,写入日志。Step S3: Output the processed motion area list and write it into the log.
本步骤中,对于判断为正常运动的帧:直接使用KNN检测结果;对于判断为运动区域巨大变化的帧:在前后帧运动区域使用现有的交集运算,并根据关联比例阈值判断该帧是否受到了遮挡、光照等其他因素影响。In this step, for frames judged to be in normal motion: directly use the KNN detection results; for frames judged to have huge changes in the motion area: use the existing intersection operation in the motion areas of the previous and next frames, and determine whether the frame is affected by the correlation ratio threshold. It is affected by other factors such as occlusion and lighting.
步骤S4:构建SlowFast的语义分析模型。Step S4: Build SlowFast’s semantic analysis model.
如图3所示,SlowFast的语义分析模型结构上面的稀疏帧采样分支是SlowFast模型中的“慢通道”(Slow Pathway),“慢通道”主要关注于提取空间语义信息,其输入是低帧率的采样视频帧(实验里设置为每次采样跳过了16帧,如果按视频30fps算,即每秒只采样2帧)。低帧率采样意味着这个“慢通道”使用了较大时序跨度的视频帧序列,因此“慢通道”对视频中的时序变化并不敏感,而更擅长捕获视频帧的空间信息,这是由于“慢通道”具有更多的3D卷积核。As shown in Figure 3, the sparse frame sampling branch above the SlowFast semantic analysis model structure is the "Slow Pathway" in the SlowFast model. The "Slow Pathway" mainly focuses on extracting spatial semantic information, and its input is a low frame rate sampling video frames (in the experiment, it was set to skip 16 frames for each sampling. If calculated based on the video 30fps, that is, only 2 frames are sampled per second). Low frame rate sampling means that this "slow channel" uses a sequence of video frames with a larger temporal span, so the "slow channel" is not sensitive to temporal changes in the video, but is better at capturing the spatial information of video frames. This is due to "Slow channel" has more 3D convolution kernels.
图3下面的分支是Slow-Fast模型中的“快通道”(Fast Pathway),“快通道”主要关注于提取时序语义信息,其输入是高帧率的采样视频帧(实验设置为慢通道的α倍,通常设为8)。 在使用更高的时序分辨率作为输入时,也保持了同样的高时序分辨率作为输出,没用采用时序下采样。同时,为了让这条通道专注于时序信息的提取,“快通道”模型在设计时减少在视频空间上的运算量,其卷积通道(卷积核的个数决定)通常设置为慢通道的β倍,通常设置为1/8。The branch below Figure 3 is the "Fast Pathway" in the Slow-Fast model. The "Fast Pathway" mainly focuses on extracting temporal semantic information, and its input is a high frame rate sampled video frame (the experiment is set to Slow Pathway α times, usually set to 8). When using a higher timing resolution as the input, the same high timing resolution is maintained as the output without using timing downsampling. At the same time, in order to allow this channel to focus on extracting temporal information, the "fast channel" model is designed to reduce the amount of calculations in the video space. Its convolution channel (determined by the number of convolution kernels) is usually set to the slow channel. β times, usually set to 1/8.
为了实现“快通道”与“慢通道”的信息交互,本发明使用了侧向连接,这是在计算机视觉领域一种常见算法,比如在2D图像物体检测的特征金字塔,把不同层的特征进行融合;在视频检测的一个主流方法Two Stream,也采取了类似的融合策略。在这里,如结构图3所示,快通道的输出结果通过侧向连接送入慢通道,实现了两个通道的信息交互,产生更有效的视频语义分析结果。In order to realize the information interaction between "fast channel" and "slow channel", the present invention uses lateral connection, which is a common algorithm in the field of computer vision. For example, in the feature pyramid of 2D image object detection, the features of different layers are connected. Fusion; Two Stream, a mainstream method for video detection, also adopts a similar fusion strategy. Here, as shown in the structure diagram 3, the output results of the fast channel are sent to the slow channel through lateral connections, realizing information interaction between the two channels and producing more effective video semantic analysis results.
步骤S5:对步骤S1中的视频数据进行分割获取一个或多个视频片段,并记录所述一个或多个视频片段的时间戳信息。Step S5: Divide the video data in step S1 to obtain one or more video segments, and record the timestamp information of the one or more video segments.
本步骤中,将视频数据分割为若干个70s的视频片段,记录时间戳信息以便后续对视频片段进行拼装。In this step, the video data is divided into several 70s video clips, and the timestamp information is recorded for subsequent assembly of the video clips.
步骤S6:对所述一个或多个视频片段进行抽帧处理,获取与单个视频片段对应的视频帧组。Step S6: Perform frame extraction processing on the one or more video segments to obtain a video frame group corresponding to a single video segment.
本步骤中,对于动物视频来说,其以背景居多,前景较少,视频相邻帧之间的变化较少,做抽帧处理后能够降低模型计算量并且能凸显视频帧蕴含的信息量。In this step, for animal videos, there are mostly backgrounds, less foregrounds, and fewer changes between adjacent frames of the video. Frame extraction processing can reduce the amount of model calculations and highlight the amount of information contained in the video frames.
步骤S7:将所述一个或多个视频片段对应的视频帧组输入到训练好的SlowFast模型中,以获取各个视频片段的语义。Step S7: Input the video frame group corresponding to the one or more video clips into the trained SlowFast model to obtain the semantics of each video clip.
步骤S8:通过语义信息,将步骤S3中捕获到的运动画面分类为背景以及各种素材类别,输出发生运动的不同时间戳信息,写入日志,并且得到动物动作剪辑,识别结果如图5所示。并且,提交得到的日志与动物动作剪辑至后端服务器,以便于后期处理与调度。Step S8: Use semantic information to classify the moving pictures captured in step S3 into background and various material categories, output different timestamp information of the movement, write it into the log, and obtain animal action clips. The recognition results are shown in Figure 5. Show. Furthermore, the obtained logs and animal action clips are submitted to the back-end server for post-processing and scheduling.
步骤S9:根据步骤S8中的运动类别结果驱动摄像头转动。Step S9: Drive the camera to rotate according to the motion category result in step S8.
本步骤中,根据动作优先级以及目标位置,采用循环补正算法驱动摄像头转动,使目标动物处于镜头中间区域,以提升素材质量。具体实现流程如下:In this step, based on the action priority and target position, a loop correction algorithm is used to drive the camera to rotate so that the target animal is in the middle area of the lens to improve the quality of the material. The specific implementation process is as follows:
步骤S91:获取步骤S3中含捕获的运动目标坐标的json文件。Step S91: Obtain the json file containing the coordinates of the moving target captured in step S3.
步骤S92:根据坐标,设定球形摄像头在x、y轴偏转方向,并执行小幅度偏转,进行位置补正。Step S92: According to the coordinates, set the deflection direction of the spherical camera on the x and y axes, and perform a small deflection to perform position correction.
步骤S93:启用运动敏感检测器,再次检测运动目标位置,重复步骤S91和步骤S912,直至目标处于镜头中间区域。Step S93: Enable the motion-sensitive detector to detect the position of the moving target again, and repeat steps S91 and S912 until the target is in the middle area of the lens.
Claims (10)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310510516.XA CN116721370A (en) | 2023-05-08 | 2023-05-08 | Capture and identification method of interesting animal behavior clips for long-term video surveillance |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202310510516.XA CN116721370A (en) | 2023-05-08 | 2023-05-08 | Capture and identification method of interesting animal behavior clips for long-term video surveillance |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN116721370A true CN116721370A (en) | 2023-09-08 |
Family
ID=87872392
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN202310510516.XA Pending CN116721370A (en) | 2023-05-08 | 2023-05-08 | Capture and identification method of interesting animal behavior clips for long-term video surveillance |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN116721370A (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN117395451A (en) * | 2023-09-13 | 2024-01-12 | 重庆数智融合创新科技有限公司 | Video file access method and system based on cloud storage |
Citations (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20080181457A1 (en) * | 2007-01-31 | 2008-07-31 | Siemens Aktiengesellschaft | Video based monitoring system and method |
| CN109948474A (en) * | 2019-03-04 | 2019-06-28 | 成都理工大学 | AI thermal imaging all-weather intelligent monitoring method |
| CN110519566A (en) * | 2019-08-28 | 2019-11-29 | 山东科技大学 | A kind of belt movement state monitoring method based on video processing |
| CN111401308A (en) * | 2020-04-08 | 2020-07-10 | 蚌埠学院 | A fish behavior video recognition method based on optical flow effect |
| CN111611847A (en) * | 2020-04-01 | 2020-09-01 | 杭州电子科技大学 | Video action detection method based on scale attention hole convolutional network |
| CN113627410A (en) * | 2021-10-14 | 2021-11-09 | 江苏奥斯汀光电科技股份有限公司 | Method for recognizing and retrieving action semantics in video |
| CN113743306A (en) * | 2021-09-06 | 2021-12-03 | 浙江广厦建设职业技术大学 | Method for analyzing abnormal behaviors of real-time intelligent video monitoring based on slowfast double-frame rate |
| CN113963315A (en) * | 2021-11-16 | 2022-01-21 | 重庆邮电大学 | A method and system for real-time video multi-person behavior recognition in complex scenes |
| CN115063724A (en) * | 2022-06-22 | 2022-09-16 | 南京理工大学 | A kind of identification method and electronic device of fruit tree field ridge |
-
2023
- 2023-05-08 CN CN202310510516.XA patent/CN116721370A/en active Pending
Patent Citations (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20080181457A1 (en) * | 2007-01-31 | 2008-07-31 | Siemens Aktiengesellschaft | Video based monitoring system and method |
| CN109948474A (en) * | 2019-03-04 | 2019-06-28 | 成都理工大学 | AI thermal imaging all-weather intelligent monitoring method |
| CN110519566A (en) * | 2019-08-28 | 2019-11-29 | 山东科技大学 | A kind of belt movement state monitoring method based on video processing |
| CN111611847A (en) * | 2020-04-01 | 2020-09-01 | 杭州电子科技大学 | Video action detection method based on scale attention hole convolutional network |
| CN111401308A (en) * | 2020-04-08 | 2020-07-10 | 蚌埠学院 | A fish behavior video recognition method based on optical flow effect |
| CN113743306A (en) * | 2021-09-06 | 2021-12-03 | 浙江广厦建设职业技术大学 | Method for analyzing abnormal behaviors of real-time intelligent video monitoring based on slowfast double-frame rate |
| CN113627410A (en) * | 2021-10-14 | 2021-11-09 | 江苏奥斯汀光电科技股份有限公司 | Method for recognizing and retrieving action semantics in video |
| CN113963315A (en) * | 2021-11-16 | 2022-01-21 | 重庆邮电大学 | A method and system for real-time video multi-person behavior recognition in complex scenes |
| CN115063724A (en) * | 2022-06-22 | 2022-09-16 | 南京理工大学 | A kind of identification method and electronic device of fruit tree field ridge |
Non-Patent Citations (1)
| Title |
|---|
| 汪威: "基于深度学习的自动扶梯视频人体动作识别", 软件工, vol. 24, no. 9, 30 September 2021 (2021-09-30) * |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN117395451A (en) * | 2023-09-13 | 2024-01-12 | 重庆数智融合创新科技有限公司 | Video file access method and system based on cloud storage |
| CN117395451B (en) * | 2023-09-13 | 2024-10-18 | 重庆数智融合创新科技有限公司 | Video file access method and system based on cloud storage |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Gygli | Ridiculously fast shot boundary detection with fully convolutional neural networks | |
| CN106856577B (en) | Video abstract generation method capable of solving multi-target collision and shielding problems | |
| CN104573111B (en) | A structured storage and pre-retrieval method for pedestrian data in surveillance video | |
| CN101902617B (en) | Device and method for realizing video structural description by using DSP and FPGA | |
| CN105844659B (en) | Method and device for tracking moving parts | |
| CN110826447A (en) | Restaurant kitchen staff behavior identification method based on attention mechanism | |
| CN109063667B (en) | Scene-based video identification mode optimization and pushing method | |
| CN111553231B (en) | Face snapshot and deduplication system, method, terminal and medium based on information fusion | |
| CN110021034A (en) | A kind of tracking recording broadcasting method and system based on head and shoulder detection | |
| CN113688741A (en) | A sports training evaluation system and method based on the coordination of event camera and visual camera | |
| Senadeera et al. | Cue-net: violence detection video analytics with spatial cropping enhanced uniformerv2 and modified efficient additive attention | |
| CN111160203A (en) | A method for loitering behavior analysis based on head and shoulders model and IOU tracking | |
| CN114140656B (en) | Marine ship target identification method based on event camera | |
| CN116721370A (en) | Capture and identification method of interesting animal behavior clips for long-term video surveillance | |
| CN116012789A (en) | Method and system for real-time monitoring of kitchen behavior based on improved target detection model | |
| Zhao et al. | Recognizing High-Speed Moving Objects with Spike Camera | |
| CN112750145A (en) | Target detection and tracking method, device and system | |
| CN117710853A (en) | A video scene detection method based on meta-event dependencies | |
| CN103957472A (en) | Timing-sequence-keeping video summary generation method and system based on optimal reconstruction of events | |
| Quenot et al. | Rushes summarization by IRIM consortium: redundancy removal and multi-feature fusion | |
| Thiruthanigesan et al. | A systematic approach for object detection in unconstrained environments | |
| Verma et al. | STEP: Simultaneous Tracking and Estimation of Pose for Animals and Humans | |
| CN114630104B (en) | Artificial intelligence multi-model alarm processing method and device for broadcast television | |
| CN114998391A (en) | Method for rapidly filtering redundant information under condition of detachable yolov 5-based image | |
| Höferlin et al. | Video visual analytics of tracked moving objects |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| SE01 | Entry into force of request for substantive examination | ||
| SE01 | Entry into force of request for substantive examination |