[go: up one dir, main page]

CN111310668A - A gait recognition method based on skeleton information - Google Patents

A gait recognition method based on skeleton information Download PDF

Info

Publication number
CN111310668A
CN111310668A CN202010100136.5A CN202010100136A CN111310668A CN 111310668 A CN111310668 A CN 111310668A CN 202010100136 A CN202010100136 A CN 202010100136A CN 111310668 A CN111310668 A CN 111310668A
Authority
CN
China
Prior art keywords
gait
features
sequence
sample
samples
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010100136.5A
Other languages
Chinese (zh)
Other versions
CN111310668B (en
Inventor
刘晓凯
尤昭阳
毕胜
刘祥
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dalian Maritime University
Original Assignee
Dalian Maritime University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dalian Maritime University filed Critical Dalian Maritime University
Priority to CN202010100136.5A priority Critical patent/CN111310668B/en
Publication of CN111310668A publication Critical patent/CN111310668A/en
Application granted granted Critical
Publication of CN111310668B publication Critical patent/CN111310668B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/103Static body considered as a whole, e.g. static pedestrian or occupant recognition
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

本发明提供一种基于骨架信息的步态识别方法,包括:采集步态视频序列;采用OpenPose对步态视频序列进行姿态估计,得到步态关键点序列;构建时空骨架序列;将邻接矩阵与步态关键点序列输入到多尺度时空图卷积网络进行训练;训练完成后,使用训练好的模型进行测试,提取步态特征,进行特征匹配。本发明主要采用人体关键点形式,引入针对图结构的图卷积神经网络,并改进连接方式以及划分策略,网络采用孪生机制,将交叉熵损失和对比损失结合,并融合网络的浅层特征、中层特征与深层特征,从一定程度上提升了步态识别的鲁棒性。

Figure 202010100136

The present invention provides a gait recognition method based on skeleton information, including: collecting gait video sequences; using OpenPose to perform posture estimation on the gait video sequences to obtain gait key point sequences; constructing spatiotemporal skeleton sequences; inputting the adjacency matrix and the gait key point sequence into a multi-scale spatiotemporal graph convolutional network for training; after the training is completed, using the trained model for testing, extracting gait features, and performing feature matching. The present invention mainly adopts the form of human key points, introduces a graph convolutional neural network for graph structure, and improves the connection method and partitioning strategy. The network adopts a twin mechanism, combines cross entropy loss and contrast loss, and integrates the shallow features, middle features and deep features of the network, which improves the robustness of gait recognition to a certain extent.

Figure 202010100136

Description

一种基于骨架信息的步态识别方法A gait recognition method based on skeleton information

技术领域technical field

本发明涉及模式识别技术领域,具体而言,尤其涉及一种基于骨架信息的步态识别方法。The present invention relates to the technical field of pattern recognition, in particular, to a gait recognition method based on skeleton information.

背景技术Background technique

步态识别是一种新兴的生物特征识别技术,旨在通过视频序列中人们走路姿态进行身份识别,与其他生物识别技术相比,步态识别具有非接触、远距离和不易伪装的优点。在安防、智能监控领域更具优势,到目前为止,在实际复杂环境中,步态识别应用仍然存在一些问题。Gait recognition is an emerging biometric recognition technology, which aims to identify people by walking gestures in video sequences. Compared with other biometric technologies, gait recognition has the advantages of non-contact, long-distance and not easy to camouflage. It has more advantages in the field of security and intelligent monitoring. So far, there are still some problems in the application of gait recognition in the actual complex environment.

近几年来,国内外学术科研机构越来越重视步态识别技术。现有技术主要分为以下两种:In recent years, domestic and foreign academic research institutions have paid more and more attention to gait recognition technology. The existing technologies are mainly divided into the following two types:

1、基于模型的方法。主要是将身体切分成多块或者获取身体关节点,通过关节点或者身体运动的部分运动轨迹进行拟合。这类方法主要依赖身体的各块静态特征以及关节点的运动轨迹进行建模,包括二维模型和三维模型。1. Model-based approach. It is mainly to divide the body into multiple pieces or obtain body joint points, and fit the joint points or part of the motion trajectory of the body movement. This type of method mainly relies on the static characteristics of each block of the body and the motion trajectories of joint points for modeling, including two-dimensional models and three-dimensional models.

2、基于非模型的方法。主要是采集人类走路的外形还有其中特征参数等步态特征来实现,不需要重建行走步态的模型。研究对象大致有三种:步态能量图、轮廓图序列和人体关键点序列。2. Non-model-based methods. It is mainly realized by collecting the shape of human walking and gait features such as characteristic parameters, and does not need to reconstruct the model of walking gait. There are roughly three kinds of research objects: gait energy map, contour map sequence and human body key point sequence.

近年来,随着深度学习技术在计算机视觉各领域取得很大的进步,涌现了大量基于深度学习的步态识别方法。例如:采用步态能量图描述步态序列,利用深度卷积神经网络训练匹配模型,从而匹配人的身份。重复选取所有视角中任意两个视角对卷积神经网络进行匹配训练,该方法能够获得较好的多视角步态识别性能,存在的不足之处是,在人体行走视角变化范围较大时,所提取的多视角步态特征表征能力不够,同时对于服装以及携带物等鲁棒性较低。再者,采用姿态信息进行步态特征表示和匹配。利用姿态估计算法从步态视频序列中获取人体关键点坐标,利用卷积神经网络和长短时记忆网络训练步态关键点序列,同时引入手工特征进行步态识别。但是现有步态识别技术仍然存在以下不足:In recent years, with the great progress of deep learning technology in various fields of computer vision, a large number of gait recognition methods based on deep learning have emerged. For example, gait energy maps are used to describe gait sequences, and deep convolutional neural networks are used to train matching models to match human identities. Repeatedly select any two perspectives from all perspectives to match and train the convolutional neural network. This method can obtain better multi-view gait recognition performance. The disadvantage is that when the human walking perspective changes in a large range, all the The extracted multi-view gait features have insufficient representation ability, and are less robust to clothing and carrying objects. Furthermore, gesture information is used for gait feature representation and matching. The pose estimation algorithm is used to obtain the coordinates of human body key points from the gait video sequence, and the gait key point sequence is trained by using convolutional neural network and long-short-term memory network. At the same time, manual features are introduced for gait recognition. However, the existing gait recognition technology still has the following shortcomings:

1、通过步态轮廓图或步态能量图进行步态识别,对轮廓质量和背景要求较高,受光照条件以及复杂背景影响较大,往往导致轮廓图提取不完整。1. Gait recognition through gait contour map or gait energy map requires high contour quality and background, and is greatly affected by lighting conditions and complex backgrounds, which often leads to incomplete contour map extraction.

2、受服装以及携带物等协变量的影响,会导致协变量无法与人体本身分离,导致识别精度降低。2. Affected by covariates such as clothing and carried objects, the covariates cannot be separated from the human body itself, resulting in reduced recognition accuracy.

发明内容SUMMARY OF THE INVENTION

根据上述提出的技术问题,而提供一种基于骨架信息的步态识别方法。本发明主要采用人体关键点形式,引入针对图结构的图卷积神经网络,并改进连接方式以及划分策略,网络采用孪生机制,将交叉熵损失和对比损失结合,并融合网络的浅层特征、中层特征与深层特征,从一定程度上提升了步态识别的鲁棒性。According to the technical problem raised above, a gait recognition method based on skeleton information is provided. The present invention mainly adopts the form of human key points, introduces a graph convolutional neural network for graph structure, and improves the connection method and division strategy. The middle-level features and deep-level features improve the robustness of gait recognition to a certain extent.

本发明采用的技术手段如下:The technical means adopted in the present invention are as follows:

一种基于骨架信息的步态识别方法,包括如下步骤:A gait recognition method based on skeleton information, comprising the following steps:

S1、采集步态视频序列;S1. Collect gait video sequences;

S2、采用OpenPose对上述步态视频序列进行姿态估计,得到步态关键点序列;S2. Use OpenPose to perform pose estimation on the above gait video sequence to obtain a sequence of gait key points;

S3、构建时空骨架序列;S3, constructing a spatiotemporal skeleton sequence;

S4、将邻接矩阵与步态关键点序列输入到多尺度时空图卷积网络进行训练;S4. Input the adjacency matrix and the gait keypoint sequence into the multi-scale spatiotemporal graph convolutional network for training;

S5、训练完成后,使用训练好的模型进行测试,提取步态特征,进行特征匹配。S5. After the training is completed, use the trained model for testing, extract gait features, and perform feature matching.

进一步地,所述步骤S3具体为:Further, the step S3 is specifically:

S31、将所述步态关键点序列在空间上进行人体自然连接;同时引入对称性将对称的关节点进行连接(只连接腿部对称关键点,因为在携带物条件下胳膊之间对称性缺失);在时间上,将帧与帧之间相同的关键点进行连接;S31. Perform the natural connection of the gait key point sequence in space; at the same time, introduce symmetry to connect the symmetrical joint points (only connect the symmetrical key points of the legs, because the symmetry between the arms is missing under the condition of carrying objects) ); in time, connect the same key points between frames;

S32、定义采样函数,一个节点vti的邻域集被定义为:S32. Define the sampling function, and the neighborhood set of a node v ti is defined as:

B(vti)={vtj|d(vtj,vti)≤D},B(v ti )={v tj |d(v tj ,v ti )≤D},

其中,B为节点vti的邻域集合;v表示节点;D表示距离;d(vtj,vti)表示两个节点之间的最短路径,通常取D=1;因此,采样函数被定义为:p(vtj,vti)=vtjAmong them, B is the neighborhood set of the node v ti ; v represents the node; D represents the distance; d(v tj , v ti ) represents the shortest path between two nodes, usually D=1; therefore, the sampling function is defined is: p(v tj , v ti )=v tj ;

S33、选择划分策略,将邻域集划分成四个子集,即该节点本身为第一个子集,非对称节点中,距离重力中心比该节点本身近的为第二个子集,距离重力重心比该节点本身远的为第三个子集,对称节点定义为第四个子集,即:S33. Select a division strategy, and divide the neighborhood set into four subsets, that is, the node itself is the first subset, and among the asymmetric nodes, the second subset is closer to the center of gravity than the node itself, and the distance to the center of gravity is the second subset. The third subset is farther than the node itself, and the symmetric node is defined as the fourth subset, namely:

Figure BDA0002386658570000031
Figure BDA0002386658570000031

S34、定义权重函数,邻域集划分为四个子集后,每一个子集有一个数字标签,采用一个映射函数lti将每一个节点映射到它的子集标签,映射函数定义为:B(vti)→{0,……,K-1},K=4;权重函数定义为:w(vti,vtj)=w’(lti(vtj));S34. Define a weight function. After the neighborhood set is divided into four subsets, each subset has a digital label, and a mapping function l ti is used to map each node to its subset label. The mapping function is defined as: B( v ti )→{0,...,K-1}, K=4; the weight function is defined as: w(v ti ,v tj )=w'(l ti (v tj ));

S35、将空间图卷积扩展到空时域,将空时邻域集定义为:B(vti)={vqj|d(vtj,vti)≤K,|q–t|≤Γ}},B为节点vti的邻域集合;v表示节点;K表示距离;Γ控制包括在邻域内的图的范围,即时间卷积核。S35. Extend the spatial graph convolution to the space-time domain, and define the space-time neighborhood set as: B(v ti )={v qj |d(v tj ,v ti )≤K, |q–t|≤Γ }}, B is the neighborhood set of the node v ti ; v represents the node; K represents the distance; Γ controls the range of the graph included in the neighborhood, that is, the temporal convolution kernel.

进一步地,所述训练多尺度时空图卷积神经网络的过程具体如下:Further, the process of training the multi-scale spatiotemporal graph convolutional neural network is as follows:

S41、选择样本后,从ID与所选择的样本相同的所有样本中随机选取一个样本作为正样本,从ID与所选择的样本不同的所有样本中随机选取一个样本作为负样本;S41, after selecting the sample, randomly select a sample from all samples whose ID is the same as the selected sample as a positive sample, and randomly select a sample from all samples whose ID is different from the selected sample as a negative sample;

S42、采用孪生机制,在一次迭代中,将所述选择的样本输入支路1,将正样本和负样本依次输入支路2,支路1和支路2共享参数;S42, adopting the twinning mechanism, in one iteration, the selected sample is input into branch 1, the positive sample and the negative sample are input into branch 2 in turn, and branch 1 and branch 2 share parameters;

S43、采用SoftMax和交叉熵损失函数对支路1中所述选择的样本特征进行分类;S43, using SoftMax and cross entropy loss function to classify the selected sample features in branch 1;

S44、采用对比损失函数对比所述选择的样本与正样本的特征,以及所述选择的样本与负样本的特征;来自于同一个ID的样本,标签为1,来自不同ID的样本,标签为0。S44, using a contrast loss function to compare the features of the selected samples and the positive samples, as well as the features of the selected samples and the negative samples; samples from the same ID have a label of 1, and samples from different IDs have a label of 1 0.

S45、两部分损失加和,总损失为:S45, the sum of the two losses, the total loss is:

Loss=Lid+0.5*[Lc(sample,pos,1)+Lc(sample,neg,0)],Loss=Lid+0.5*[Lc(sample,pos,1)+Lc(sample,neg,0)],

其中,Lid为交叉熵损失,Lc为对比损失,再进行反向传播,更新网络。Among them, Lid is the cross entropy loss, Lc is the contrast loss, and then backpropagation is performed to update the network.

进一步地,所述步骤S4中在训练多尺度时空图卷积神经网络时还包括如下设置过程:Further, in the step S4, the following setting process is also included when training the multi-scale spatiotemporal graph convolutional neural network:

步骤1、输入步态序列,维度为[3,100,18],其中3为输入关键点特征有3个通道,分别为X,Y坐标和置信度C,100为时间维度有100帧,18为每帧共有18个关键点;Step 1. Input the gait sequence, the dimension is [3, 100, 18], of which 3 is the input key point feature with 3 channels, namely X, Y coordinates and confidence C, 100 is the time dimension has 100 frames, 18 is each There are 18 key points in the frame;

步骤2、前三层输出64通道,卷积核大小为(9,3),9为时间卷积核尺寸,3为空间卷积核尺寸,输出维度为[64,100,18];Step 2. The first three layers output 64 channels, the convolution kernel size is (9, 3), 9 is the time convolution kernel size, 3 is the spatial convolution kernel size, and the output dimension is [64, 100, 18];

步骤3、中间三层输出128通道,卷积核大小为(9,3),输出维度为[128,50,18],在第四层,时间维度卷积步长为2;Step 3. The middle three layers output 128 channels, the convolution kernel size is (9,3), the output dimension is [128,50,18], and in the fourth layer, the time dimension convolution step size is 2;

步骤4、后三层输出256通道,卷积核大小为(9,3),输出维度为[256,25,18],在第七层,时间维度卷积步长为2;Step 4. The last three layers output 256 channels, the convolution kernel size is (9,3), the output dimension is [256,25,18], and in the seventh layer, the time dimension convolution step size is 2;

步骤5、进行全局平均池化,池化后,特征维度变成256维;Step 5. Perform global average pooling. After pooling, the feature dimension becomes 256 dimensions;

步骤6、将第一层输出的特征[64,100,18],进行维度交换后,做平均池化,变成18维特征;Step 6. After the features [64, 100, 18] output by the first layer are exchanged for dimensions, average pooling is performed to turn them into 18-dimensional features;

步骤7、将第五层的输出特征[128,50,18],进行维度交换后,做平均池化,变成18维特征;Step 7. After the output features of the fifth layer [128, 50, 18] are exchanged for dimensions, average pooling is performed to turn them into 18-dimensional features;

步骤8、采用将浅层特征与深层特征融合的方式表示步态特征,将第一层的18维特征、第五层的18维特征与最后一层的256维特征进行拼接,变成292维特征;Step 8. The gait feature is represented by the fusion of shallow features and deep features, and the 18-dimensional features of the first layer, the 18-dimensional features of the fifth layer and the 256-dimensional features of the last layer are spliced into 292-dimensional features. feature;

步骤9、采用SoftMax分类器将292维特征进行分类。Step 9. Use the SoftMax classifier to classify the 292-dimensional features.

本发明还提供了一种基于骨架信息的步态识别方法的测试方法,包括如下步骤:The present invention also provides a test method for the gait recognition method based on skeleton information, comprising the following steps:

步骤Ⅰ:输入待测试的步态关键点序列;Step 1: Input the sequence of gait key points to be tested;

步骤Ⅱ:利用训练好的网络提取步态特征,并对该特征进行二范数归一化;Step II: Use the trained network to extract gait features, and perform two-norm normalization on the features;

步骤Ⅲ:对样本库中的样本进行步骤Ⅰ和步骤Ⅱ的操作,则可用特征向量来代表待检索行人步态序列和检索库中行人步态序列;Step III: Perform the operations of Step I and Step II on the samples in the sample database, and the feature vector can be used to represent the pedestrian gait sequence to be retrieved and the pedestrian gait sequence in the retrieval database;

步骤Ⅳ:计算待检索行人步态序列与检索库中行人步态序列之间的距离,即针对一个待检索行人步态序列,计算其特征与检索库中所有行人步态序列之间的距离;Step IV: Calculate the distance between the pedestrian gait sequence to be retrieved and the pedestrian gait sequence in the retrieval database, that is, for a pedestrian gait sequence to be retrieved, calculate the distance between its features and all the pedestrian gait sequences in the retrieval database;

步骤Ⅴ:按照上述计算的距离由小到大将检索库中样本进行相似性排序,越靠前则越有可能与待检索行人ID一致。Step V: Sort the similarity of the samples in the retrieval database from small to large according to the distance calculated above. The higher the distance, the more likely it is consistent with the ID of the pedestrian to be retrieved.

较现有技术相比,本发明具有以下优点:Compared with the prior art, the present invention has the following advantages:

1、本发明提供的步态识别方法,针对背包以及服装等协变量影响,采用关键点序列来表征步态的方式,解决了用轮廓图以及能量图表示的步态特征在协变量条件下鲁棒性较低问题,提升了在协变量影响下识别的准确率。1. The gait recognition method provided by the present invention, aiming at the influence of covariates such as backpacks and clothing, adopts the method of characterizing the gait by the sequence of key points, and solves the problem that the gait characteristics represented by the contour map and the energy map are ineffective under the condition of covariates. The problem of low stickiness improves the accuracy of identification under the influence of covariates.

2、本发明提供的步态识别方法,针对步态特有的对称性特点,在构建时空骨架序列时引入对称性,将人体腿部对称的关节点信息加入邻接矩阵,增强了相关节点的关联度,同时减小了因关节估计不准确带来的噪声,提升了识别的准确率。2. The gait recognition method provided by the present invention, aiming at the unique symmetry characteristics of the gait, introduces symmetry when constructing the space-time skeleton sequence, and adds the symmetrical joint point information of the human body to the adjacency matrix, which enhances the correlation degree of the relevant nodes. , while reducing the noise caused by inaccurate joint estimation and improving the recognition accuracy.

3、由于深度卷积神经网络提取的是高层特征,单一表示高层语义信息,无法描述静态特征,因此采用将浅层特征、中层特征及深层特征融合的多尺度方式,丰富了步态特征的表达形式,提升了识别的准确率。3. Since the deep convolutional neural network extracts high-level features, it can only represent high-level semantic information and cannot describe static features. Therefore, a multi-scale method that combines shallow features, mid-level features and deep features is adopted to enrich the expression of gait features. form, which improves the accuracy of recognition.

基于上述理由本发明可在模式识别等领域广泛推广。Based on the above reasons, the present invention can be widely promoted in the fields of pattern recognition and the like.

附图说明Description of drawings

为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图做以简单地介绍,显而易见地,下面描述中的附图是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to illustrate the embodiments of the present invention or the technical solutions in the prior art more clearly, the following briefly introduces the accompanying drawings that need to be used in the description of the embodiments or the prior art. Obviously, the accompanying drawings in the following description These are some embodiments of the present invention, and for those of ordinary skill in the art, other drawings can also be obtained from these drawings without any creative effort.

图1为本发明方法流程图。Fig. 1 is the flow chart of the method of the present invention.

图2为本发明多尺度时空图卷积神经网络具体设置示意图。FIG. 2 is a schematic diagram of a specific setting of a multi-scale spatiotemporal graph convolutional neural network according to the present invention.

图3为本发明测试方法示意图。Figure 3 is a schematic diagram of the testing method of the present invention.

具体实施方式Detailed ways

为了使本技术领域的人员更好地理解本发明方案,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分的实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都应当属于本发明保护的范围。In order to make those skilled in the art better understand the solutions of the present invention, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only Embodiments are part of the present invention, but not all embodiments. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without creative efforts shall fall within the protection scope of the present invention.

需要说明的是,本发明的说明书和权利要求书及上述附图中的术语“第一”、“第二”等是用于区别类似的对象,而不必用于描述特定的顺序或先后次序。应该理解这样使用的数据在适当情况下可以互换,以便这里描述的本发明的实施例能够以除了在这里图示或描述的那些以外的顺序实施。此外,术语“包括”和“具有”以及他们的任何变形,意图在于覆盖不排他的包含,例如,包含了一系列步骤或单元的过程、方法、系统、产品或设备不必限于清楚地列出的那些步骤或单元,而是可包括没有清楚地列出的或对于这些过程、方法、产品或设备固有的其它步骤或单元。It should be noted that the terms "first", "second" and the like in the description and claims of the present invention and the above drawings are used to distinguish similar objects, and are not necessarily used to describe a specific sequence or sequence. It is to be understood that the data so used may be interchanged under appropriate circumstances such that the embodiments of the invention described herein can be practiced in sequences other than those illustrated or described herein. Furthermore, the terms "comprising" and "having" and any variations thereof, are intended to cover non-exclusive inclusion, for example, a process, method, system, product or device comprising a series of steps or units is not necessarily limited to those expressly listed Rather, those steps or units may include other steps or units not expressly listed or inherent to these processes, methods, products or devices.

如图1所示,本发明提供了一种基于骨架信息的步态识别方法,包括如下步骤:As shown in Figure 1, the present invention provides a gait recognition method based on skeleton information, comprising the following steps:

S1、采集步态视频序列;S1. Collect gait video sequences;

S2、采用OpenPose对上述步态视频序列进行姿态估计,得到步态关键点序列;S2. Use OpenPose to perform pose estimation on the above gait video sequence to obtain a sequence of gait key points;

S3、构建时空骨架序列;S3, constructing a spatiotemporal skeleton sequence;

进一步地,作为本发明优选的实施方式,所述步骤S3具体为:Further, as a preferred embodiment of the present invention, the step S3 is specifically:

S31、将所述步态关键点序列在空间上进行人体自然连接;因为人走路姿态具有对称性的特点,所以同时引入对称性将对称的关节点进行连接(只连接腿部对称关键点,因为在携带物条件下胳膊之间对称性缺失);在时间上,将帧与帧之间相同的关键点进行连接;S31. Perform the natural connection of the gait key point sequence on the human body in space; because the walking posture of a person has the characteristics of symmetry, symmetry is introduced to connect the symmetrical joint points at the same time (only the symmetrical key points of the legs are connected, because Loss of symmetry between the arms in the carrying condition); temporally, connect the same key points from frame to frame;

S32、定义采样函数,一个节点vti的邻域集被定义为:S32. Define the sampling function, and the neighborhood set of a node v ti is defined as:

B(vti)={vtj|d(vtj,vti)≤D},B(v ti )={v tj |d(v tj ,v ti )≤D},

其中,B为节点vti的邻域集合;v表示节点;D表示距离;d(vtj,vti)表示两个节点之间的最短路径,通常取D=1;因此,采样函数被定义为:p(vtj,vti)=vtjAmong them, B is the neighborhood set of the node v ti ; v represents the node; D represents the distance; d(v tj , v ti ) represents the shortest path between two nodes, usually D=1; therefore, the sampling function is defined is: p(v tj , v ti )=v tj ;

S33、选择划分策略,将邻域集划分成四个子集,即该节点本身为第一个子集,非对称节点中,距离重力中心比该节点本身近的为第二个子集,距离重力重心比该节点本身远的为第三个子集,对称节点定义为第四个子集,即:S33. Select a division strategy, and divide the neighborhood set into four subsets, that is, the node itself is the first subset, and among the asymmetric nodes, the second subset is closer to the center of gravity than the node itself, and the distance to the center of gravity is the second subset. The third subset is farther than the node itself, and the symmetric node is defined as the fourth subset, namely:

Figure BDA0002386658570000071
Figure BDA0002386658570000071

S34、定义权重函数,邻域集划分为四个子集后,每一个子集有一个数字标签,采用一个映射函数lti将每一个节点映射到它的子集标签,映射函数定义为:B(vti)→{0,……,K-1},K=4;权重函数定义为:w(vti,vtj)=w’(lti(vtj));S34. Define a weight function. After the neighborhood set is divided into four subsets, each subset has a digital label, and a mapping function l ti is used to map each node to its subset label. The mapping function is defined as: B( v ti )→{0,...,K-1}, K=4; the weight function is defined as: w(v ti ,v tj )=w'(l ti (v tj ));

S35、将空间图卷积扩展到空时域,将空时邻域集定义为:B(vti)={vqj|d(vtj,vti)≤K,|q–t|≤Γ}},B为节点vti的邻域集合;v表示节点;K表示距离;Γ控制包括在邻域内的图的范围,即时间卷积核。S35. Extend the spatial graph convolution to the space-time domain, and define the space-time neighborhood set as: B(v ti )={v qj |d(v tj ,v ti )≤K, |q–t|≤Γ }}, B is the neighborhood set of the node v ti ; v represents the node; K represents the distance; Γ controls the range of the graph included in the neighborhood, that is, the temporal convolution kernel.

S4、将邻接矩阵与步态关键点序列输入到多尺度时空图卷积网络进行训练;S4. Input the adjacency matrix and the gait keypoint sequence into the multi-scale spatiotemporal graph convolutional network for training;

进一步地,作为本发明优选的实施方式,如图2所示,所述步骤S4中在训练多尺度时空图卷积神经网络之前还包括如下设置过程:Further, as a preferred embodiment of the present invention, as shown in FIG. 2 , the step S4 further includes the following setting process before training the multi-scale spatiotemporal graph convolutional neural network:

步骤1、输入步态序列,维度为[3,100,18],其中3为输入关键点特征有3个通道,分别为X,Y坐标和置信度C,100为时间维度有100帧,18为每帧共有18个关键点;Step 1. Input the gait sequence, the dimension is [3, 100, 18], of which 3 is the input key point feature with 3 channels, namely X, Y coordinates and confidence C, 100 is the time dimension has 100 frames, 18 is each There are 18 key points in the frame;

步骤2、前三层输出64通道,卷积核大小为(9,3),9为时间卷积核尺寸,3为空间卷积核尺寸,输出维度为[64,100,18];Step 2. The first three layers output 64 channels, the convolution kernel size is (9, 3), 9 is the time convolution kernel size, 3 is the spatial convolution kernel size, and the output dimension is [64, 100, 18];

步骤3、中间三层输出128通道,卷积核大小为(9,3),输出维度为[128,50,18],在第四层,时间维度卷积步长为2;Step 3. The middle three layers output 128 channels, the convolution kernel size is (9,3), the output dimension is [128,50,18], and in the fourth layer, the time dimension convolution step size is 2;

步骤4、后三层输出256通道,卷积核大小为(9,3),输出维度为[256,25,18],在第七层,时间维度卷积步长为2;Step 4. The last three layers output 256 channels, the convolution kernel size is (9,3), the output dimension is [256,25,18], and in the seventh layer, the time dimension convolution step size is 2;

步骤5、进行全局平均池化,池化后,特征维度变成256维;Step 5. Perform global average pooling. After pooling, the feature dimension becomes 256 dimensions;

步骤6、将第一层输出的特征[64,100,18],进行维度交换后,做平均池化,变成18维特征;Step 6. After the features [64, 100, 18] output by the first layer are exchanged for dimensions, average pooling is performed to turn them into 18-dimensional features;

步骤7、将第五层的输出特征[128,50,18],进行维度交换后,做平均池化,变成18维特征;Step 7. After the output features of the fifth layer [128, 50, 18] are exchanged for dimensions, average pooling is performed to turn them into 18-dimensional features;

步骤8、由于深度卷积神经网络提取的是高层特征,单一表示高层语义信息,无法描述静态特征,因此采用将浅层特征与深层特征融合的方式表示步态特征,将第一层的18维特征、第五层的18维特征与最后一层的256维特征进行拼接,变成292维特征;Step 8. Since the deep convolutional neural network extracts high-level features, it represents high-level semantic information alone and cannot describe static features. Therefore, the gait features are represented by fusing shallow features and deep features. The features, the 18-dimensional features of the fifth layer and the 256-dimensional features of the last layer are spliced into 292-dimensional features;

步骤9、采用SoftMax分类器将292维特征进行分类。Step 9. Use the SoftMax classifier to classify the 292-dimensional features.

本实施例中,采用CASIA-B数据集,NM:正常行走条件,BG:携带物条件,CL:穿大衣条件。如下表所示:In this embodiment, the CASIA-B dataset is used, NM: normal walking condition, BG: carrying condition, CL: wearing coat condition. As shown in the table below:

Figure BDA0002386658570000081
Figure BDA0002386658570000081

所述训练多尺度时空图卷积神经网络的过程具体如下:The specific process of training a multi-scale spatiotemporal graph convolutional neural network is as follows:

在训练阶段,目的是训练网络使其能提取到能代表行人的特征,故网络以分类的形式进行训练,具体步骤为:In the training phase, the purpose is to train the network so that it can extract features that can represent pedestrians. Therefore, the network is trained in the form of classification. The specific steps are:

S41、选择样本后,从ID与所选择的样本相同的所有样本中随机选取一个样本作为正样本,从ID与所选择的样本不同的所有样本中随机选取一个样本作为负样本;S41, after selecting the sample, randomly select a sample from all samples whose ID is the same as the selected sample as a positive sample, and randomly select a sample from all samples whose ID is different from the selected sample as a negative sample;

S42、采用孪生机制,在一次迭代中,将所述选择的样本输入支路1,将正样本和负样本依次输入支路2,支路1和支路2共享参数;S42, adopting the twinning mechanism, in one iteration, the selected sample is input into branch 1, the positive sample and the negative sample are input into branch 2 in turn, and branch 1 and branch 2 share parameters;

S43、采用SoftMax和交叉熵损失函数对支路1中所述选择的样本特征进行分类;S43, using SoftMax and cross entropy loss function to classify the selected sample features in branch 1;

S44、采用对比损失函数对比所述选择的样本与正样本的特征,以及所述选择的样本与负样本的特征;将来自于同一个ID的样本,标签为1,来自不同ID的样本,标签为0。S44, using a contrast loss function to compare the features of the selected samples and the positive samples, and the features of the selected samples and the negative samples; the samples from the same ID have a label of 1, and the samples from different IDs have a label of 1. is 0.

S45、两部分损失加和,总损失为:S45, the sum of the two losses, the total loss is:

Loss=Lid+0.5*[Lc(sample,pos,1)+Lc(sample,neg,0)],Loss=Lid+0.5*[Lc(sample,pos,1)+Lc(sample,neg,0)],

其中,Lid为交叉熵损失,Lc为对比损失,再进行反向传播,更新网络。Among them, Lid is the cross entropy loss, Lc is the contrast loss, and then backpropagation is performed to update the network.

S5、训练完成后,使用训练好的模型进行测试,提取步态特征,进行特征匹配。S5. After the training is completed, use the trained model for testing, extract gait features, and perform feature matching.

如图3所示,本发明还提供了一种基于骨架信息的步态识别方法的测试方法,包括如下步骤:As shown in Figure 3, the present invention also provides a test method for a gait recognition method based on skeleton information, comprising the following steps:

步骤Ⅰ:输入待测试的步态关键点序列;Step 1: Input the sequence of gait key points to be tested;

步骤Ⅱ:利用训练好的网络提取步态特征,并对该特征进行二范数归一化;Step II: Use the trained network to extract gait features, and perform two-norm normalization on the features;

步骤Ⅲ:对样本库中的样本进行步骤Ⅰ和步骤Ⅱ的操作,则可用特征向量来代表待检索行人步态序列和检索库中行人步态序列;Step III: Perform the operations of Step I and Step II on the samples in the sample database, and the feature vector can be used to represent the pedestrian gait sequence to be retrieved and the pedestrian gait sequence in the retrieval database;

步骤Ⅳ:计算待检索行人步态序列与检索库中行人步态序列之间的距离,即针对一个待检索行人步态序列,计算其特征与检索库中所有行人步态序列之间的距离;Step IV: Calculate the distance between the pedestrian gait sequence to be retrieved and the pedestrian gait sequence in the retrieval database, that is, for a pedestrian gait sequence to be retrieved, calculate the distance between its features and all the pedestrian gait sequences in the retrieval database;

步骤Ⅴ:按照上述计算的距离由小到大将检索库中样本进行相似性排序,越靠前则越有可能与待检索行人ID一致。Step V: Sort the similarity of the samples in the retrieval database from small to large according to the distance calculated above. The higher the distance, the more likely it is consistent with the ID of the pedestrian to be retrieved.

最后应说明的是:以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述实施例所记载的技术方案进行修改,或者对其中部分或者全部技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明实施例技术方案的范围。Finally, it should be noted that the above embodiments are only used to illustrate the technical solutions of the present invention, but not to limit them; although the present invention has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that it can still be The technical solutions described in the foregoing embodiments are modified, or some or all of the technical features thereof are equivalently replaced; and these modifications or replacements do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present invention.

Claims (5)

1. A gait recognition method based on skeleton information is characterized by comprising the following steps:
s1, acquiring a gait video sequence;
s2, carrying out attitude estimation on the gait video sequence by adopting OpenPose to obtain a gait key point sequence;
s3, constructing a spatio-temporal skeleton sequence;
s4, inputting the adjacency matrix and the gait key point sequence into a multi-scale space-time graph convolution network for training;
and S5, after the training is finished, testing by using the trained model, extracting gait characteristics and carrying out characteristic matching.
2. The method for recognizing gait based on skeleton information according to claim 1, characterized in that the step S3 is specifically:
s31, carrying out human body natural connection on the gait key point sequence in space; meanwhile, symmetry is introduced to connect symmetrical joint points (only connecting symmetrical key points of the legs because of lack of symmetry between arms under the condition of carrying objects); connecting the same key points from frame to frame in time;
s32, defining a sampling function, a node vtiIs defined as:
B(vti)={vtj|d(vtj,vti)≤D},
wherein B is node vtiA neighborhood set of (c); v represents a node; d represents a distance; d (v)tj,vti) Representing the shortest path between two nodes, usually taking D ═ 1; thus, the sampling function is defined as: p (v)tj,vti)=vtj
S33, selecting a partitioning strategy, partitioning the neighborhood set into four subsets, that is, the node itself is the first subset, in the asymmetric node, a second subset closer to the gravity center than the node itself, and a third subset farther from the gravity center than the node itself, where the symmetric node is defined as the fourth subset, that is:
Figure FDA0002386658560000011
s34, defining a weight function, dividing the neighborhood set into four subsets, each subset having a digital label, and adopting a mapping function ltiMapping each node to its subset label, the mapping function being defined as: b (v)ti) → {0, … …, K-1}, K ═ 4; the weight function is defined as: w (v)ti,vtj)=w’(lti(vtj));
S35, expanding the space graph convolution to a space-time domain, and defining a space-time neighborhood set as: b (v)ti)={vqj|d(vtj,vti) K is less than or equal to, q-t is less than or equal to gamma, and B is a node vtiA neighborhood set of (c); v represents a node; k represents a distance; Γ controls the extent of the graph, i.e., the temporal convolution kernel, that is included in the neighborhood.
3. The method for gait recognition based on the skeleton information as claimed in claim 1, wherein the process of training the multi-scale space-time graph convolutional neural network is as follows:
s41, after the samples are selected, randomly selecting one sample from all samples with the same ID as the selected samples as a positive sample, and randomly selecting one sample from all samples with different IDs from the selected samples as a negative sample;
s42, inputting the selected sample into a branch 1, sequentially inputting a positive sample and a negative sample into a branch 2 in one iteration by adopting a twin mechanism, wherein the branch 1 and the branch 2 share parameters;
s43, classifying the selected sample characteristics in the branch 1 by adopting SoftMax and a cross entropy loss function;
s44, comparing the characteristics of the selected sample and the positive sample and the characteristics of the selected sample and the negative sample by using a contrast loss function; samples from the same ID are labeled 1, and samples from different IDs are labeled 0.
S45, adding the two part losses, wherein the total loss is as follows:
Loss=Lid+0.5*[Lc(sample,pos,1)+Lc(sample,neg,0)],
and performing back propagation to update the network, wherein Lid is cross entropy loss, and Lc is contrast loss.
4. A gait recognition method based on skeleton information according to claim 3, characterized in that said step S4 further includes the following setting process when training the multi-scale space-time graph convolutional neural network:
step 1, inputting a gait sequence, wherein the dimensionality is [3,100,18], 3 is that input key point characteristics have 3 channels which are respectively X, Y coordinates and confidence C, 100 is that the time dimensionality has 100 frames, and 18 is that each frame has 18 key points;
step 2, outputting 64 channels from the first three layers, wherein the convolution kernel size is (9,3), 9 is the time convolution kernel size, 3 is the space convolution kernel size, and the output dimension is [64,100,18 ];
step 3, outputting 128 channels in the middle three layers, wherein the convolution kernel size is (9,3), the output dimensionality is [128,50,18], and in the fourth layer, the time dimension convolution step size is 2;
step 4, outputting 256 channels in the last three layers, wherein the convolution kernel size is (9,3), the output dimensionality is [256,25,18], and in the seventh layer, the time dimension convolution step size is 2;
step 5, performing global average pooling, wherein after the pooling is performed, the characteristic dimension is changed into 256 dimensions;
step 6, carrying out dimension exchange on the features [64,100 and 18] output by the first layer, and then carrying out average pooling to obtain 18-dimensional features;
step 7, carrying out dimension exchange on the output features [128,50 and 18] of the fifth layer, and then carrying out average pooling to obtain 18-dimensional features;
step 8, representing the gait characteristics by fusing the shallow layer characteristics, the middle layer characteristics and the deep layer characteristics, and splicing the 18-dimensional characteristics of the first layer, the 18-dimensional characteristics of the fifth layer and the 256-dimensional characteristics of the last layer to obtain 292-dimensional characteristics;
and 9, classifying the 292-dimensional features by adopting a SoftMax classifier.
5. A gait recognition method testing method based on skeleton information is characterized by comprising the following steps:
step I: inputting a gait key point sequence to be tested;
step II: extracting gait features by using the trained network, and carrying out two-norm normalization on the features;
step III: carrying out the operations of the step I and the step II on the samples in the sample library, and representing the gait sequence of the pedestrian to be searched and the gait sequence of the pedestrian in the search library by using the characteristic vector;
step IV: calculating the distance between the pedestrian gait sequence to be retrieved and the pedestrian gait sequence in the retrieval library, namely calculating the distance between the characteristics of the pedestrian gait sequence to be retrieved and all the pedestrian gait sequences in the retrieval library aiming at one pedestrian gait sequence to be retrieved;
step V: and performing similarity sorting on the samples in the search library according to the calculated distance from small to large, wherein the more front the samples are, the more likely the samples are consistent with the ID of the pedestrian to be searched.
CN202010100136.5A 2020-02-18 2020-02-18 Gait recognition method based on skeleton information Active CN111310668B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010100136.5A CN111310668B (en) 2020-02-18 2020-02-18 Gait recognition method based on skeleton information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010100136.5A CN111310668B (en) 2020-02-18 2020-02-18 Gait recognition method based on skeleton information

Publications (2)

Publication Number Publication Date
CN111310668A true CN111310668A (en) 2020-06-19
CN111310668B CN111310668B (en) 2023-06-23

Family

ID=71147331

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010100136.5A Active CN111310668B (en) 2020-02-18 2020-02-18 Gait recognition method based on skeleton information

Country Status (1)

Country Link
CN (1) CN111310668B (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860291A (en) * 2020-07-16 2020-10-30 上海交通大学 Multimodal pedestrian identification method and system based on pedestrian appearance and gait information
CN112101176A (en) * 2020-09-09 2020-12-18 元神科技(杭州)有限公司 User identity recognition method and system combining user gait information
CN112434655A (en) * 2020-12-07 2021-03-02 安徽大学 Gait recognition method based on adaptive confidence map convolution network
CN112633222A (en) * 2020-12-30 2021-04-09 民航成都电子技术有限责任公司 Gait recognition method, device, equipment and medium based on confrontation network
CN112906599A (en) * 2021-03-04 2021-06-04 杭州海康威视数字技术股份有限公司 Gait-based personnel identity identification method and device and electronic equipment
CN112926522A (en) * 2021-03-30 2021-06-08 广东省科学院智能制造研究所 Behavior identification method based on skeleton attitude and space-time diagram convolutional network
CN113191230A (en) * 2021-04-20 2021-07-30 内蒙古工业大学 Gait recognition method based on gait space-time characteristic decomposition
CN114052726A (en) * 2021-11-25 2022-02-18 湖南中科助英智能科技研究院有限公司 A thermal infrared human gait recognition method and device in a dark environment
CN114359962A (en) * 2021-12-27 2022-04-15 浙江大华技术股份有限公司 Posture recognition method and device and readable storage medium
CN114783050A (en) * 2022-03-21 2022-07-22 上海交通大学 Gait identity recognition method and system based on dynamic vision sensor
CN115240269A (en) * 2022-06-30 2022-10-25 浙江大华技术股份有限公司 Gait recognition method and device based on body type transformation and storage medium
CN115346149A (en) * 2022-06-21 2022-11-15 浙江大沩人工智能科技有限公司 Rope skipping counting method and system based on space-time diagram convolution network
CN119785413A (en) * 2024-09-10 2025-04-08 杭州电子科技大学 A gait recognition method based on multimodal data input

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090232353A1 (en) * 2006-11-10 2009-09-17 University Of Maryland Method and system for markerless motion capture using multiple cameras
CN110096950A (en) * 2019-03-20 2019-08-06 西北大学 A kind of multiple features fusion Activity recognition method based on key frame

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090232353A1 (en) * 2006-11-10 2009-09-17 University Of Maryland Method and system for markerless motion capture using multiple cameras
CN110096950A (en) * 2019-03-20 2019-08-06 西北大学 A kind of multiple features fusion Activity recognition method based on key frame

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
ZHENSHEN QU 等: ""Design of Real-Time Measurement System with Vision/IMU for Close-Range Semi-Physical Rendezvous and Docking Simulation"", 《PROCEEDINGS OF 2016 IEEE CHINESE GUIDANCE, NAVIGATION AND CONTROL CONFERENCE (IEEE CGNCC2016)》 *
刘晓凯: ""智能监控系统中行人重识别方法研究"", 《中国博士学位论文全文数据库 信息科技辑》 *
宋宪 等: ""基于时空图卷积网络的人体运动状态识别研究"", 《新材料、新技术、新工艺在竞技体育的研发与应用(二)》 *
李瑞: ""图像和深度图中的动作识别与手势姿态估计"", 《中国博士学位论文全文数据库 信息科技辑》 *
董安 等: ""基于图卷积的骨架行为识别"", 《图形图像》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111860291A (en) * 2020-07-16 2020-10-30 上海交通大学 Multimodal pedestrian identification method and system based on pedestrian appearance and gait information
CN112101176A (en) * 2020-09-09 2020-12-18 元神科技(杭州)有限公司 User identity recognition method and system combining user gait information
CN112101176B (en) * 2020-09-09 2024-04-05 元神科技(杭州)有限公司 User identity recognition method and system combining user gait information
CN112434655A (en) * 2020-12-07 2021-03-02 安徽大学 Gait recognition method based on adaptive confidence map convolution network
CN112434655B (en) * 2020-12-07 2022-11-08 安徽大学 A Gait Recognition Method Based on Adaptive Confidence Graph Convolutional Network
CN112633222B (en) * 2020-12-30 2023-04-28 民航成都电子技术有限责任公司 Gait recognition method, device, equipment and medium based on countermeasure network
CN112633222A (en) * 2020-12-30 2021-04-09 民航成都电子技术有限责任公司 Gait recognition method, device, equipment and medium based on confrontation network
CN112906599A (en) * 2021-03-04 2021-06-04 杭州海康威视数字技术股份有限公司 Gait-based personnel identity identification method and device and electronic equipment
CN112926522A (en) * 2021-03-30 2021-06-08 广东省科学院智能制造研究所 Behavior identification method based on skeleton attitude and space-time diagram convolutional network
CN112926522B (en) * 2021-03-30 2023-11-24 广东省科学院智能制造研究所 Behavior recognition method based on skeleton gesture and space-time diagram convolution network
CN113191230A (en) * 2021-04-20 2021-07-30 内蒙古工业大学 Gait recognition method based on gait space-time characteristic decomposition
CN114052726A (en) * 2021-11-25 2022-02-18 湖南中科助英智能科技研究院有限公司 A thermal infrared human gait recognition method and device in a dark environment
CN114359962A (en) * 2021-12-27 2022-04-15 浙江大华技术股份有限公司 Posture recognition method and device and readable storage medium
CN114783050A (en) * 2022-03-21 2022-07-22 上海交通大学 Gait identity recognition method and system based on dynamic vision sensor
CN115346149A (en) * 2022-06-21 2022-11-15 浙江大沩人工智能科技有限公司 Rope skipping counting method and system based on space-time diagram convolution network
CN115240269A (en) * 2022-06-30 2022-10-25 浙江大华技术股份有限公司 Gait recognition method and device based on body type transformation and storage medium
CN119785413A (en) * 2024-09-10 2025-04-08 杭州电子科技大学 A gait recognition method based on multimodal data input

Also Published As

Publication number Publication date
CN111310668B (en) 2023-06-23

Similar Documents

Publication Publication Date Title
CN111310668B (en) Gait recognition method based on skeleton information
CN109948425B (en) A pedestrian search method and device based on structure-aware self-attention and online instance aggregation and matching
CN113221625B (en) Method for re-identifying pedestrians by utilizing local features of deep learning
CN108596039B (en) Bimodal emotion recognition method and system based on 3D convolutional neural network
CN106951872B (en) A Pedestrian Re-identification Method Based on Unsupervised Deep Model and Hierarchical Attributes
Ming et al. Simple triplet loss based on intra/inter-class metric learning for face verification
CN111274922B (en) Pedestrian re-identification method and system based on multi-level deep learning network
CN111428658B (en) A gait recognition method based on modal fusion
CN104616316B (en) Personage's Activity recognition method based on threshold matrix and Fusion Features vision word
CN112199983B (en) Long-time large-range pedestrian re-identification method based on multi-level screening
CN112396027A (en) Vehicle weight recognition method based on graph convolution neural network
CN110263697A (en) Pedestrian based on unsupervised learning recognition methods, device and medium again
CN110163117B (en) Pedestrian re-identification method based on self-excitation discriminant feature learning
CN109961051A (en) A Pedestrian Re-identification Method Based on Clustering and Blocking Feature Extraction
Xiong et al. ASK: Adaptively selecting key local features for RGB-D scene recognition
CN109815826A (en) The generation method and device of face character model
CN104036255A (en) Facial expression recognition method
CN114419671A (en) An occluded pedestrian re-identification method based on hypergraph neural network
CN108960078A (en) A method of based on monocular vision, from action recognition identity
CN111353411A (en) Face-shielding identification method based on joint loss function
CN114973305B (en) Accurate human body analysis method for crowded people
CN111985367A (en) Pedestrian re-recognition feature extraction method based on multi-scale feature fusion
CN113627380B (en) Cross-vision pedestrian re-identification method and system for intelligent security and early warning
CN111339849A (en) A Pedestrian Re-identification Method Based on Pedestrian Attributes
CN111985332A (en) Gait recognition method for improving loss function based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
EE01 Entry into force of recordation of patent licensing contract

Application publication date: 20200619

Assignee: Dalian Sanying Technology Co.,Ltd.

Assignor: Dalian Maritime University

Contract record no.: X2024210000055

Denomination of invention: A gait recognition method based on skeleton information

Granted publication date: 20230623

License type: Common License

Record date: 20240830

EE01 Entry into force of recordation of patent licensing contract