[go: up one dir, main page]

CN112699956B - Neuromorphic visual target classification method based on improved impulse neural network - Google Patents

Neuromorphic visual target classification method based on improved impulse neural network Download PDF

Info

Publication number
CN112699956B
CN112699956B CN202110025987.2A CN202110025987A CN112699956B CN 112699956 B CN112699956 B CN 112699956B CN 202110025987 A CN202110025987 A CN 202110025987A CN 112699956 B CN112699956 B CN 112699956B
Authority
CN
China
Prior art keywords
neural network
event
improved
time
network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202110025987.2A
Other languages
Chinese (zh)
Other versions
CN112699956A (en
Inventor
赵广社
姚满
王鼎衡
刘美兰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian Jiaotong University
Original Assignee
Xian Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian Jiaotong University filed Critical Xian Jiaotong University
Priority to CN202110025987.2A priority Critical patent/CN112699956B/en
Publication of CN112699956A publication Critical patent/CN112699956A/en
Application granted granted Critical
Publication of CN112699956B publication Critical patent/CN112699956B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/04Inference or reasoning models
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种基于改进脉冲神经网络的神经形态视觉目标分类方法。所述方法包括:S1:获取神经形态视觉目标分类数据集;S2:脉冲事件流序列化聚合:将数据集中的时空脉冲事件流数据,按照设定的时间分辨率dt聚合成新的事件帧序列数据;S3:构建改进脉冲神经网络模型:改进泄露‑积累‑发射(Leaky Integrity and Fire,LIF)脉冲神经元在时间维度上的突触连接方式,基于改进LIF神经元层搭建改进脉冲神经网络;S4:对于脉冲事件流序列化聚合后的数据集,从序列中随机抽取样本作为输入,训练和测试所构建的改进脉冲神经网络;S5:保存训练好的改进脉冲神经网络结构和网络参数。本发明能够有效的提升神经形态视觉的目标识别与分类中的网络分类准确率问题。

The invention discloses a neuromorphic visual target classification method based on an improved impulse neural network. The method includes: S1: Obtain neuromorphic visual target classification data set; S2: Pulse event stream serialization aggregation: aggregate the spatiotemporal pulse event stream data in the data set into a new event frame sequence according to the set time resolution dt. Data; S3: Construct an improved spiking neural network model: improve the synaptic connection method of Leaky Integrity and Fire (LIF) spiking neurons in the time dimension, and build an improved spiking neural network based on the improved LIF neuron layer; S4: For the data set after serialization and aggregation of the impulse event stream, samples are randomly selected from the sequence as input to train and test the constructed improved impulse neural network; S5: Save the trained improved impulse neural network structure and network parameters. The present invention can effectively improve the network classification accuracy problem in target recognition and classification of neuromorphic vision.

Description

一种基于改进脉冲神经网络的神经形态视觉目标分类方法A neuromorphic visual target classification method based on improved spiking neural network

技术领域Technical field

本发明属于机器学习中的深度学习领域,具体涉及一种基于改进脉冲神经网络的神经形态视觉目标分类方法。The invention belongs to the field of deep learning in machine learning, and specifically relates to a neuromorphic visual target classification method based on an improved impulse neural network.

背景技术Background technique

近年来,以人工神经网络(Artificial Neural Networks,ANNs)为代表的深度学习在图像识别、自然语言处理等领域取得了巨大的成功,但由于人工神经网络模型运行机制与生物学中实际观察到的大脑运行机制有着根本不同,其难以实现真正的强人工智能。脉冲神经网络(Spiking Neural Networks,SNNs)以更具生物可解释性的脉冲神经元模型作为基本单元,与基于脉冲频率编码信息的人工神经网络相比,拥有低时延和低能耗等优势,可以模拟各种神经信号和任意的连续函数,是进行复杂时空信息处理的有效工具。In recent years, deep learning, represented by Artificial Neural Networks (ANNs), has achieved great success in fields such as image recognition and natural language processing. However, due to the differences between the operating mechanism of artificial neural network models and the actual observations in biology, The operating mechanism of the brain is fundamentally different, making it difficult to achieve true strong artificial intelligence. Spiking Neural Networks (SNNs) use a more biologically interpretable spiking neuron model as the basic unit. Compared with artificial neural networks based on pulse frequency encoding information, they have the advantages of low latency and low energy consumption, and can Simulating various neural signals and arbitrary continuous functions is an effective tool for complex spatiotemporal information processing.

以“事件(脉冲)”为基础的神经形态视觉事件流由基于仿生学原理设计的事件相机通过异步测量每个像素点的亮度变化而产生,事件流对亮度变化的时间、位置和亮度变化极性进行编码。与传统相机相比,事件相机提供了具有吸引力的性能:高时间分辨率(微秒级)、非常高的动态范围(140dB vs 60dB)、高像素带宽(kHz数量级),从而有效减少了运动模糊。由于SNNs对输入脉冲的处理具有超低时延特性,因此通常使用脉冲神经网络来处理视觉事件流,从而充分利用视觉事件流的高时间分辨率进行实时输出(微秒级时延输出)。然而现有的典型脉冲神经网络结构难以充分利用脉冲事件流数据的样本容量,故其视觉目标分类上的表现不如人工神经网络。The neuromorphic visual event stream based on "events (pulses)" is generated by an event camera designed based on bionics principles by asynchronously measuring the brightness change of each pixel. The event stream is extremely sensitive to the time, location and brightness changes of the brightness change. Sexuality is coded. Compared to traditional cameras, event cameras offer attractive performance: high temporal resolution (microsecond level), very high dynamic range (140dB vs 60dB), high pixel bandwidth (order of kHz), thus effectively reducing motion Vague. Since SNNs have ultra-low latency characteristics in processing input pulses, spiking neural networks are usually used to process the visual event stream, thereby fully utilizing the high time resolution of the visual event stream for real-time output (microsecond-level delay output). However, the existing typical spiking neural network structure cannot fully utilize the sample capacity of spiking event stream data, so its performance in visual target classification is not as good as that of artificial neural networks.

目标识别与分类是神经形态视觉的一个重要任务。目前,现有的神经形态视觉目标识别与分类学习算法主要有如下三种类型:Object recognition and classification is an important task in neuromorphic vision. Currently, there are three main types of existing neuromorphic visual target recognition and classification learning algorithms:

(1)基于时间编码的脉冲时间依赖可塑性的非监督式算法。时间编码方式使得这种学习算法可以有与输入几乎同步的输出,但时间依赖可塑性学习算法主要通过局部的神经元活动来改变突触权重,很难实现高性能。(1) Unsupervised algorithm for pulse time-dependent plasticity based on temporal coding. The temporal encoding method allows this learning algorithm to have an output that is almost synchronized with the input. However, the time-dependent plasticity learning algorithm mainly changes synaptic weights through local neuron activity, making it difficult to achieve high performance.

(2)基于频率编码的ANN转SNN非直接训练式监督算法。通过将训练好的ANNs模型映射为对应的SNNs,使得这种学习算法目前可以获取与ANNs相当的网络性能。但是,这种转换方法并没有利用事件流中的时间信息,且存在着诸多约束条件,例如激活函数只能使用ReLU等。(2) Non-direct training supervision algorithm from ANN to SNN based on frequency coding. By mapping the trained ANNs model to the corresponding SNNs, this learning algorithm can currently obtain network performance comparable to ANNs. However, this conversion method does not utilize the time information in the event stream, and there are many constraints. For example, the activation function can only use ReLU, etc.

(3)基于频率编码的误差反向传播直接训练式监督算法。通过训练误差在时间和空间上进行反向传播来进行突触的权重的学习,代表性算法为基于时间的反向传播算法(Backpropagation-Through-Time,BPTT)。但是,由于脉冲函数不可微分等原因,在深层SNNs在训练中存在梯度消失(爆炸)等问题,难以获取好的网络性能。(3) Error backpropagation direct training supervision algorithm based on frequency coding. Synaptic weights are learned by backpropagating training errors in time and space. The representative algorithm is the time-based backpropagation algorithm (Backpropagation-Through-Time, BPTT). However, due to reasons such as the non-differentiability of the impulse function, deep SNNs have problems such as gradient disappearance (explosion) during training, making it difficult to obtain good network performance.

现有的基于脉冲神经网络的神经形态视觉目标分类方法,无论是时间编码方式还是频率编码方式中的哪一种,脉冲神经网络结构(由于梯度问题,网络较浅)和训练方式(转换方法不能利用时间信息)存在的固有问题,使得神经形态视觉目标分类准确率的提升都是一个难题。现有技术中,基于LIF(Leaky Integrity and Fire)神经元模型的脉冲神经网络使用误差反向传播算法进行监督学习,这使得研究者们可以使用GPU进行加速训练,还可以使用在深度学习中十分成熟的Pytorch等训练工具。一般地,较深的网络结构是提升网络性能的必要条件,然而训练中存在的脉冲不可微分导致的梯度问题使得深度脉冲神经网络在训练中难以收敛。The existing neuromorphic visual target classification methods based on spiking neural networks, regardless of which one is the time encoding method or the frequency encoding method, the spiking neural network structure (the network is shallow due to the gradient problem) and the training method (the conversion method cannot The inherent problems in using temporal information make it difficult to improve the accuracy of neuromorphic visual target classification. In the existing technology, the spiking neural network based on the LIF (Leaky Integrity and Fire) neuron model uses the error backpropagation algorithm for supervised learning, which allows researchers to use GPUs for accelerated training and can also be used in deep learning. Mature Pytorch and other training tools. Generally, a deeper network structure is a necessary condition to improve network performance. However, the gradient problem caused by non-differentiable impulses in training makes it difficult for deep impulse neural networks to converge during training.

发明内容Contents of the invention

为克服上述现有技术的不足,本发明提供了一种基于改进脉冲神经网络的神经形态视觉目标分类方法。该方法通过对输入数据进行重新聚合,改进脉冲神经网络基本构造神经元模型,来实现高性能的神经形态视觉目标分类。本发明提出的方法首先将脉冲事件流数据聚合成事件帧序列,在训练网络时采用时间随机裁剪方法确定输入网络的数据,这相当于增加可供网络训练的数据量;接着改进了构建了基于改进LIF神经元的脉冲神经网络,通过加强网络在较长时间序列中提取时空特征的能力,避免了训练深度脉冲神经网络,使得浅层改进脉冲神经网络在神经形态视觉目标分类任务中也能得到较高的准确率。In order to overcome the above-mentioned shortcomings of the prior art, the present invention provides a neuromorphic visual target classification method based on an improved spiking neural network. This method achieves high-performance neuromorphic visual target classification by re-aggregating input data and improving the basic structure of the spiking neural network neuron model. The method proposed by the present invention first aggregates the pulse event stream data into an event frame sequence, and uses a time random clipping method to determine the data input to the network when training the network, which is equivalent to increasing the amount of data available for network training; then, an improved system based on Improving the spiking neural network of LIF neurons avoids training a deep spiking neural network by strengthening the network's ability to extract spatiotemporal features in longer time series, making the shallow improved spiking neural network also available in neuromorphic visual target classification tasks Higher accuracy.

为实现上述目的,本发明采用如下技术方案来实现的:In order to achieve the above objects, the present invention adopts the following technical solutions:

一种基于改进脉冲神经网络的神经形态视觉目标分类方法,该方法首先将脉冲事件流数据聚合成事件帧序列,在训练网络时采用时间随机裁剪方法确定输入网络的数据,这相当于增加可供网络训练的数据量;接着改进了构建了基于改进LIF神经元的脉冲神经网络,通过加强网络在较长时间序列中提取时空特征的能力,避免了训练深度脉冲神经网络,使得浅层改进脉冲神经网络在神经形态视觉目标分类任务中也能得到较高的准确率。A neuromorphic visual target classification method based on an improved spiking neural network. This method first aggregates the spiking event stream data into a sequence of event frames. When training the network, a time random cropping method is used to determine the data input to the network. This is equivalent to increasing the available data. The amount of data for network training; then we improved the construction of a spiking neural network based on improved LIF neurons. By strengthening the network's ability to extract spatiotemporal features in a longer time series, we avoided training a deep spiking neural network, making shallow improvements to spiking neural networks. The network can also achieve high accuracy in neuromorphic visual object classification tasks.

本发明进一步的改进在于,该方法具体包括以下步骤:A further improvement of the present invention is that the method specifically includes the following steps:

S1:获取神经形态视觉目标分类数据集;S1: Obtain neuromorphic visual object classification data set;

S2:脉冲事件流序列化聚合:将神经形态视觉目标分类数据集中的时空脉冲事件流数据,按照设定的时间分辨率dt聚合成新的事件帧序列数据;S2: Pulse event stream serialization aggregation: Aggregate the spatiotemporal pulse event stream data in the neuromorphic visual target classification data set into new event frame sequence data according to the set time resolution dt;

S3:构建改进脉冲神经网络模型:改进泄露-积累-发射LIF脉冲神经元在时间维度上的突触连接方式,基于改进LIF神经元层搭建改进脉冲神经网络;S3: Construct an improved spiking neural network model: improve the synaptic connection method of leakage-accumulation-emitting LIF spiking neurons in the time dimension, and build an improved spiking neural network based on the improved LIF neuron layer;

S4:对于脉冲事件流序列化聚合后的数据集,从序列中随机抽取样本作为输入,训练和测试所构建的改进脉冲神经网络;S4: For the data set after the serialization and aggregation of the pulse event stream, samples are randomly selected from the sequence as input, and the improved pulse neural network constructed is trained and tested;

S5:保存训练好的改进脉冲神经网络结构和网络参数,该网络参数即为对应的神经形态视觉目标分类所需要的参数。S5: Save the trained improved spiking neural network structure and network parameters. The network parameters are the parameters required for the corresponding neuromorphic visual target classification.

本发明进一步的改进在于,步骤S2具体分三步进行:A further improvement of the present invention is that step S2 is specifically carried out in three steps:

第一步,获取神经形态视觉目标分类数据集中时空脉冲事件流,由集合E={ei|ei=[xi,yi,t′i,pi]}确定;其中ei为脉冲事件流中的第i个脉冲事件,(xi,yi)为第i个脉冲事件的像素坐标,t′i为第i个脉冲事件在整个时间流中的时间戳,pi为第i个脉冲事件的光强变化极性,事件相机异步输脉冲事件流的时间分辨率为dt′,空间分辨率为H×W;The first step is to obtain the spatio-temporal pulse event stream in the neuromorphic visual target classification data set, which is determined by the set E={e i |e i =[x i ,y i ,t′ i ,pi ] }; where e i is the pulse The i-th pulse event in the event stream, (xi , y i ) is the pixel coordinate of the i-th pulse event, t′ i is the timestamp of the i-th pulse event in the entire time stream, p i is the i-th pulse event The light intensity change polarity of each pulse event, the time resolution of the asynchronous pulse event stream transmitted by the event camera is dt′, and the spatial resolution is H×W;

第二步,基于事件相机输出的脉冲事件流,聚合时间分辨率为dt′的事件帧,以t′时刻聚合为例,将t′时刻产生的若干个事件集合Et′组装成张量Xt′,其中,Et′={ei|ei=[xi,yi,t′,pi]},Xt′∈RH×W×2In the second step, based on the pulse event stream output by the event camera, event frames with a time resolution of dt′ are aggregated. Taking aggregation at time t′ as an example, several event sets E t′ generated at time t′ are assembled into a tensor X t′ , where, E t′ = {e i |e i = [x i ,y i ,t′, p i ]}, X t′ ∈R H×W×2 ;

第三步,基于时间分辨率为dt′的事件帧,Xt=f(X′t)生成t时刻的事件帧张量Xt∈RH×W×2,其中,dt=β×dt′,β为聚合时间因子;X′t={Xt′|t′∈[β×t,β×(t+1)-1]};f为累加操作等一般性计算操作。The third step is to generate the event frame tensor X t ∈R H×W×2 at time t based on the event frame with time resolution dt′, X t =f(X′ t ), where dt=β×dt′ , β is the aggregation time factor;

本发明进一步的改进在于,步骤S3中,改进LIF脉冲神经元层数学表达式为:A further improvement of the present invention is that in step S3, the improved mathematical expression of the LIF pulse neuron layer is:

其中,xt,n-1为t时刻第n-1层输入的事件帧,ht-1,n为t-1时刻第n层的内部状态量,ut,n为t时刻第n层的膜电势,xt,n为t时刻第n层传递到下一层的空间输出,ht,n为t时刻第n层传递到下一时刻的时间输出。Among them, x t,n-1 is the event frame input by the n-1th layer at time t, h t-1,n is the internal state quantity of the nth layer at time t-1, u t,n is the nth layer at time t The membrane potential of , x t,n is the spatial output from the nth layer at time t to the next layer, h t,n is the time output from the nth layer at time t to the next time.

本发明进一步的改进在于,RNNs_Model构成模型的网络为循环神经网络;A further improvement of the present invention is that the network forming the model of RNNs_Model is a recurrent neural network;

所述LIF_Model的数学表达式为:The mathematical expression of the LIF_Model is:

其中,函数g为阶跃函数, 表示Hadamard乘积;Among them, function g is a step function, Represents Hadamard product;

基于改进LIF脉冲神经元层,LIF神经元层和全连接层,构建改进脉冲神经网络。Based on the improved LIF spiking neuron layer, LIF neuron layer and fully connected layer, an improved spiking neural network is constructed.

本发明进一步的改进在于,步骤S4具体包括:A further improvement of the present invention is that step S4 specifically includes:

在训练改进脉冲神经网络时,若事件帧数据按照设定的时间分辨率dt一共生成的Ttotal个事件帧,从中随机抽取T个事件帧作为训练输入,T<TtotalWhen training the improved spiking neural network, if the event frame data generates a total of T total event frames according to the set time resolution dt, and T event frames are randomly selected from them as training input, T < T total ;

在测试改进脉冲神经网络时,使用滑动窗口方法在时间维度上对事件帧序列进行预处理,获得固定大小为T的预分割片段,作为测试数据;预分割片段设置为m个,对同一个数据预分割出来的m份测试数据输入网络得到m个预测结果,当有m/2个以上的结果预测正确时,判定这个测试数据预测正确。When testing the improved spiking neural network, the sliding window method is used to preprocess the event frame sequence in the time dimension, and pre-segmented fragments of a fixed size T are obtained as test data; the pre-segmented fragments are set to m, and the same data is The pre-segmented m test data are input into the network to obtain m prediction results. When more than m/2 results are predicted correctly, the test data is judged to be correctly predicted.

相对于现有技术,本发明至少具有如下有益的技术效果:Compared with the existing technology, the present invention at least has the following beneficial technical effects:

1、本发明所述学习系统中,通过将事件流数据重新聚合成事件帧数据(张量数据),使得脉冲神经网络学习和推理中可以使用GPU作为训练加速工具、使用Pytorch等深度学习平台语言工具。1. In the learning system of the present invention, by re-aggregating the event stream data into event frame data (tensor data), the GPU can be used as a training acceleration tool and deep learning platform languages such as Pytorch can be used in the learning and reasoning of the impulse neural network. tool.

2、本发明所述学习系统中,通过使用随机时间裁剪方法获取T个事件帧作为训练输入进行网络训练,这相当于增加了可供训练使用的数据量,因此能够有效提升所获取的网络模型的性能。2. In the learning system of the present invention, T event frames are obtained as training input by using the random time clipping method for network training. This is equivalent to increasing the amount of data available for training, and therefore can effectively improve the obtained network model. performance.

3、本发明提供了一类改进LIF神经元模型。使用改进LIF神经元模型组成全连接(卷积)脉冲神经网络,相对于基于LIF模型的脉冲神经网络能够大幅度提升神经形态视觉的目标识别与分类精确度。3. The present invention provides an improved LIF neuron model. The improved LIF neuron model is used to form a fully connected (convolutional) spiking neural network. Compared with the spiking neural network based on the LIF model, the target recognition and classification accuracy of neuromorphic vision can be greatly improved.

附图说明Description of the drawings

图1为本发明所提供的一种基于改进脉冲神经网络的神经形态视觉目标分类方法流程示意图。Figure 1 is a schematic flow chart of a neuromorphic visual target classification method based on an improved spiking neural network provided by the present invention.

图2为神经形态视觉数据集CIFAR10-DVS中的一个事件流数据(鱼类),经过脉冲事件流序列化聚合模块后的一种输出。即,将100,000微秒的脉冲事件流聚合成9个分辨率为10毫秒的事件帧。Figure 2 is an output of an event stream data (fish) in the neuromorphic vision data set CIFAR10-DVS after passing through the pulse event stream serialization and aggregation module. That is, a 100,000 microsecond pulse event stream is aggregated into 9 event frames with a resolution of 10 milliseconds.

图3为改进脉冲神经网络的一种网络构造方式。Figure 3 shows a network construction method for improving the spiking neural network.

图4为LIF脉冲神经元模型示意图。Figure 4 is a schematic diagram of the LIF spiking neuron model.

图5为改进LIF脉冲神经元(RNNs_LIF)模型示意图。Figure 5 is a schematic diagram of the improved LIF spiking neuron (RNNs_LIF) model.

图6为一种改进LIF脉冲神经元(RNN_LIF)模型示意图。Figure 6 is a schematic diagram of an improved LIF spiking neuron (RNN_LIF) model.

图7为一种改进LIF脉冲神经元(LSTM_LIF)模型示意图。Figure 7 is a schematic diagram of an improved LIF spiking neuron (LSTM_LIF) model.

图8为一种改进LIF脉冲神经元(GRU_LIF)模型示意图。Figure 8 is a schematic diagram of an improved LIF spiking neuron (GRU_LIF) model.

具体实施方式Detailed ways

下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some of the embodiments of the present invention, rather than all the embodiments. Based on the embodiments of the present invention, all other embodiments obtained by those of ordinary skill in the art without creative efforts fall within the scope of protection of the present invention.

本发明的目的是提供一种基于改进脉冲神经网络的神经形态视觉目标分类方法,有效的解决神经形态视觉的目标识别与分类问题。为使本发明的上述目的、特征和优点能够更加明显易懂,下面结合附图和具体实施方式对本发明作进一步详细的说明。The purpose of the present invention is to provide a neuromorphic visual target classification method based on an improved impulse neural network to effectively solve the problem of neuromorphic visual target recognition and classification. In order to make the above objects, features and advantages of the present invention more obvious and understandable, the present invention will be described in further detail below with reference to the accompanying drawings and specific embodiments.

图1为本发明所提供的一种基于改进脉冲神经网络的神经形态视觉目标分类方法流程示意图,包括:Figure 1 is a schematic flow chart of a neuromorphic visual target classification method based on an improved spiking neural network provided by the present invention, including:

S1:获取神经形态视觉目标分类数据集。S1: Obtain neuromorphic visual object classification data set.

S2:脉冲事件流序列化聚合,将数据集中的时空脉冲事件流数据,按照设定的时间分辨率dt聚合成新的事件帧序列数据,具体分三步进行:S2: Pulse event stream serialization aggregation. The spatio-temporal pulse event stream data in the data set is aggregated into new event frame sequence data according to the set time resolution dt. This is done in three steps:

第一步,获取神经形态视觉目标分类数据集中时空脉冲事件流。所述时空脉冲事件流包含了多个时空脉冲事件,采用地址事件表达协议进行描述。具体地,所述时空脉冲事件流由集合E={ei|ei=[xi,yi,t′i,pi]}确定;其中ei为脉冲事件流中的第i个脉冲事件,(xi,yi)为第i个脉冲事件的像素坐标,t′i为第i个脉冲事件在整个时间流中的时间戳,pi为第i个脉冲事件的光强变化极性。其中,事件相机异步输脉冲事件流的时间分辨率为dt′,空间分辨率为H×W。The first step is to obtain the spatiotemporal spike event stream in the neuromorphic visual object classification data set. The spatio-temporal pulse event stream contains multiple spatio-temporal pulse events and is described using an address event expression protocol. Specifically, the spatio-temporal pulse event stream is determined by the set E = {e i | e i = [xi , yi , t′ i , p i ]}; where ei is the i-th pulse in the pulse event stream event, (xi , y i ) is the pixel coordinate of the i-th pulse event, t′ i is the timestamp of the i-th pulse event in the entire time stream, p i is the light intensity change extreme of the i-th pulse event sex. Among them, the time resolution of the asynchronous pulse event stream transmitted by the event camera is dt′, and the spatial resolution is H×W.

第二步,基于事件相机输出的脉冲事件流,聚合时间分辨率为dt′的事件帧。所述事件帧序列采用张量进行描述。具体地,以t′时刻聚合为例,将t′时刻产生的若干个事件集合Et′组装成张量Xt′。其中,Et′={ei|ei=[xi,yi,t′,pi]},Xt′∈RH×W×2In the second step, based on the pulse event stream output by the event camera, event frames with a time resolution of dt′ are aggregated. The event frame sequence is described using tensors. Specifically, taking the aggregation at time t′ as an example, several event sets E t′ generated at time t′ are assembled into a tensor X t′ . Among them, E t′ ={e i |e i =[x i ,y i ,t′, pi ]}, X t′ ∈R H×W×2 .

第三步,基于时间分辨率为dt′的事件帧,生成时间分辨率为dt的事件帧。即,使用公式Xt=f(X′t)生成t时刻的事件帧张量Xt∈RH×w×2。其中,dt=β×dt′,β为聚合时间因子;X′t={Xt′|t′∈[β×t,β×(t+1)-1]};f为一般性聚合函数,可以是累加操作、加权累加操作或“与或非”等操作。The third step is to generate an event frame with time resolution dt based on the event frame with time resolution dt′. That is, the event frame tensor X t ∈R H×w×2 at time t is generated using the formula X t =f(X′ t ). Among them, dt=β×dt′, β is the aggregation time factor; X′ t ={X t′ |t′∈[β×t,β×(t+1)-1]}; f is a general aggregation function , which can be an accumulation operation, a weighted accumulation operation, or an "AND or NOT" operation.

当dt′=1μs,β=3时,“与”聚合方法的一个示例如下图所示:When dt′=1μs and β=3, an example of the “AND” aggregation method is shown in the figure below:

图2展示了,数据集CIFAR10-DVS数据集中的一个数据(鱼类),当输入为900,000微秒的脉冲事件流,聚合方法为“或”操作,聚合时间因子为10,000时,脉冲事件流序列化聚合模块的输出。即,图2中展示了事件流聚合而成的9个分辨率为10毫秒的事件帧。Figure 2 shows a piece of data (fish) in the CIFAR10-DVS data set. When the input is a pulse event stream of 900,000 microseconds, the aggregation method is the "OR" operation, and the aggregation time factor is 10,000, the pulse event stream sequence ize the output of the aggregation module. That is, Figure 2 shows the event stream aggregated into 9 event frames with a resolution of 10 milliseconds.

S3:构建改进脉冲神经网络模型:改进LIF脉冲神经元在时间维度上的突触连接方式,基于改进LIF神经元层搭建改进脉冲神经网络。通过以一定的网络连接方式组合改进LIF脉冲神经元层、LIF神经元层和全连接层,构建所需要训练和测试的改进脉冲神经网络。图3展示了一种改进LIF脉冲神经元层、LIF神经元层和全连接层组建的一种网络连接方式。S3: Construct an improved spiking neural network model: improve the synaptic connection method of LIF spiking neurons in the time dimension, and build an improved spiking neural network based on the improved LIF neuron layer. By combining the improved LIF spiking neuron layer, LIF neuron layer and fully connected layer in a certain network connection method, the improved spiking neural network required for training and testing is constructed. Figure 3 shows a network connection method that improves the formation of LIF spiking neuron layer, LIF neuron layer and fully connected layer.

如图4所示,构造经典脉冲神经网络的LIF神经元可以通过如下公式描述:As shown in Figure 4, the LIF neurons constructing the classic spiking neural network can be described by the following formula:

其中,函数g为阶跃函数, 表示Hadamard乘积,xt,n-1为t时刻第n-1层输入的事件帧,ht-1,n为t-1时刻第n层的内部状态量;ut,n为膜电势;将膜电势与神经元阈值uth进行比较,并将xt,n作为空间输出传递到下一层,ht,n作为LIF神经元的时间输出传递到下一时刻。Among them, function g is a step function, Represents the Hadamard product, x t,n-1 is the event frame input to the n-1th layer at time t, h t-1,n is the internal state quantity of the nth layer at time t-1; u t,n is the membrane potential; The membrane potential is compared to the neuron threshold u th , and x t,n is passed to the next layer as the spatial output, and h t,n is passed to the next moment as the temporal output of the LIF neuron.

LIF神经元通过公式来聚合空间输入xt,n-1和神经元上一时刻内部状态ht-1,n。聚合后的结果ut,n为LIF神经元的膜电势。LIF神经元的这种时空信息聚合方式,实际上是一种时间上的直接连接,其缺点在于,在聚合单个LIF神经元的膜电势时,只聚合了此神经元在上一个时刻的中间状态,并没有考虑同一层中其他神经元中间状态对神经元膜电势的影响。LIF neurons pass the formula To aggregate the spatial input x t,n-1 and the internal state of the neuron at the previous moment h t-1,n . The aggregated result u t,n is the membrane potential of the LIF neuron. This spatiotemporal information aggregation method of LIF neurons is actually a direct connection in time. Its disadvantage is that when aggregating the membrane potential of a single LIF neuron, only the intermediate state of this neuron at the previous moment is aggregated. , and does not consider the impact of the intermediate states of other neurons in the same layer on the neuron membrane potential.

在本发明中,通过改进LIF神经元中膜电势的获取方式,获得了一类改进LIF神经元模型。如图5所示,本发明中使用处理序列数据的循环神经网络(Recurrent NeuralNetworks,RNNs)来聚合空间输入xt,n-1和神经元上一时刻内部状态ht-1,n。所述改进LIF神经元层模型数学表达式为:In the present invention, by improving the acquisition method of membrane potential in LIF neurons, a type of improved LIF neuron model is obtained. As shown in Figure 5, the present invention uses Recurrent Neural Networks (RNNs) that process sequence data to aggregate the spatial input x t,n-1 and the internal state of the neuron at the previous moment h t-1,n . The mathematical expression of the improved LIF neuron layer model is:

其中,xt,n-1为t时刻第n-1层输入的事件帧,ht-1,n为t-1时刻第n层的内部状态量;将xt,n-1和ht-1,n分别作为RNNs_Model的空间输入和时间输入,选定RNNs_Model的一个输出作为膜电势ut,n;将膜电势与神经元阈值进行比较,并将xt,n作为空间输出传递到下一层,ht ,n作为LIF神经元的时间输出传递到下一时刻。Among them, x t,n-1 is the event frame input by the n-1th layer at time t, h t-1,n is the internal state quantity of the nth layer at time t-1; x t,n-1 and h t -1,n are used as the spatial input and time input of RNNs_Model respectively, and one output of RNNs_Model is selected as the membrane potential u t,n ; the membrane potential is compared with the neuron threshold, and x t,n is passed to the next as the spatial output. One layer, h t ,n is passed to the next moment as the temporal output of LIF neurons.

所述RNNs_Model可以是循环神经网络(Recurrent Neural Network,RNN)、长短期记忆网络(LSTM,Long Short-Term Memory)、双向循环神经网络(Bidirectional RNN,Bi-RNN)和门控循环单元网络(Gated Recurrent Unit networks,GRU)等循环神经网络模型。The RNNs_Model can be a recurrent neural network (Recurrent Neural Network, RNN), a long short-term memory network (LSTM, Long Short-Term Memory), a bidirectional recurrent neural network (Bidirectional RNN, Bi-RNN) and a gated recurrent unit network (Gated Recurrent Unit networks (GRU) and other recurrent neural network models.

所述发射或泄露机制(LIF_Model)表达式为:The expression of the emission or leakage mechanism (LIF_Model) is:

其中,函数g为阶跃函数, 表示Hadamard乘积。Among them, function g is a step function, Represents the Hadamard product.

如图6所示,一种RNNs_Model,使用RNN模型聚合xt,n-1和ht-1,n获得膜电势ut,n,可以得到RNN_LIF神经元模型:As shown in Figure 6, an RNNs_Model uses the RNN model to aggregate x t,n-1 and h t-1,n to obtain the membrane potential u t,n , and the RNN_LIF neuron model can be obtained:

如图7所示,一种RNNs_Model,使用LSTM模型聚合xt,n-1和ht-1,n获得膜电势ut,n,可以得到LSTM_LIF神经元模型:As shown in Figure 7, a RNNs_Model uses the LSTM model to aggregate x t,n-1 and h t-1,n to obtain the membrane potential u t,n , and the LSTM_LIF neuron model can be obtained:

如图8所示,一种RNNs_Model,使用GRU模型聚合xt,n-1和ht-1,n获得膜电势ut,n,可以得到GRU_LIF神经元模型:As shown in Figure 8, a RNNs_Model uses the GRU model to aggregate x t,n-1 and h t-1,n to obtain the membrane potential u t,n , and the GRU_LIF neuron model can be obtained:

在上文所述模型中,权重矩阵W和输入中间状态向量h或空间输入向量之间的运算为矩阵乘法,此时网络模型为改进脉冲神经网络。当权重矩阵W和输入中间状态向量h或空间输入向量之间的运算为卷积运算时,则网络模型变为改进卷积脉冲神经网络。In the model mentioned above, the operation between the weight matrix W and the input intermediate state vector h or the spatial input vector is matrix multiplication. At this time, the network model is an improved spiking neural network. When the operation between the weight matrix W and the input intermediate state vector h or spatial input vector is a convolution operation, the network model becomes an improved convolutional impulse neural network.

S4:对于脉冲事件流序列化聚合后的数据集训练和测试所构建的改进脉冲神经网络。S4: Training and testing of the improved spiking neural network constructed on the dataset after serialization and aggregation of spiking event streams.

在训练改进脉冲神经网络时,使用随机时空裁剪方法从所有事件帧数据中选出T个事件帧作为神经网络模块训练输入,所述T个事件帧采用张量进行描述。即,在单个时空脉冲事件流数据中,按照设定的时间分辨率dt一共可以生成的Ttotal个事件帧,从中随机抽取T(T<Ttotal)个事件帧作为训练输入。When training the improved spiking neural network, a random spatiotemporal clipping method is used to select T event frames from all event frame data as neural network module training inputs, and the T event frames are described by tensors. That is, in a single spatiotemporal pulse event stream data, a total of T total event frames can be generated according to the set time resolution dt, from which T (T < T total ) event frames are randomly selected as training input.

在测试改进脉冲神经网络时,使用滑动窗口方法在时间维度上对事件帧序列进行预处理,获得固定大小为T的预分割片段,作为测试数据。预分割片段设置为m个,对同一个数据预分割出来的m份测试数据输入网络可以得到m个预测结果,当有m/2个以上的结果预测正确时,判定这个测试数据预测正确。When testing the improved spiking neural network, the sliding window method is used to preprocess the event frame sequence in the time dimension, and pre-segmented fragments of a fixed size T are obtained as test data. The number of pre-segmentation fragments is set to m. If m pieces of test data pre-segmented from the same data are input into the network, m prediction results can be obtained. When more than m/2 results are predicted correctly, the test data prediction is judged to be correct.

S5:保存训练好的改进脉冲神经网络结构和网络参数,该网络参数即为对应的神经形态视觉目标分类所需要的参数。S5: Save the trained improved spiking neural network structure and network parameters. The network parameters are the parameters required for the corresponding neuromorphic visual target classification.

为了更好地说明本发明的有益效果,下面给出本发明所述方法在神经形态视觉目标分类数据集DVS128 Gesture上的实验结果。In order to better illustrate the beneficial effects of the present invention, the experimental results of the method of the present invention on the neuromorphic visual target classification data set DVS128 Gesture are given below.

在DVS128 Gesture数据集上,我们设置了不同时间分辨率下,事件帧序列T=60时,在全连接脉冲神经网络和卷积脉冲神经网络中的实验结果。由上面两个表格可以看出,无论是基于全连接的脉冲神经网络,还是基于卷积的脉冲神经网络,LIF-LSTM相比较于经典SNN的网络性能在所有时间分辨率上都有所提升。此外,在dt较小时,LIF-LSTM的提升效果更加明显,这说明LSTM-LIF相较于经典SNN在具有较低时延的同时能够保持高性能。On the DVS128 Gesture data set, we set experimental results in fully connected spiking neural networks and convolutional spiking neural networks under different time resolutions and when the event frame sequence T=60. As can be seen from the above two tables, whether it is based on a fully connected spiking neural network or a convolution-based spiking neural network, the network performance of LIF-LSTM has improved at all time resolutions compared with the classic SNN. In addition, when dt is small, the improvement effect of LIF-LSTM is more obvious, which shows that LSTM-LIF can maintain high performance while having lower latency than classic SNN.

以上列举的仅是本发明的具体实施例。显然,本发明不限于以上实施例,还可以有许多变形。本领域的普通技术人员能从本发明公开的内容直接导出或联想到的所有变形,均应认为是本发明的保护范围。What are listed above are only specific embodiments of the present invention. Obviously, the present invention is not limited to the above embodiments, and many modifications are possible. All modifications that a person of ordinary skill in the art can directly derive or associate from the disclosure of the present invention should be considered to be within the protection scope of the present invention.

Claims (2)

1.一种基于改进脉冲神经网络的神经形态视觉目标分类方法,其特征在于,该方法首先将脉冲事件流数据聚合成事件帧序列,在训练网络时采用时间随机裁剪方法确定输入网络的数据,这相当于增加可供网络训练的数据量;接着改进了构建了基于改进LIF神经元的脉冲神经网络,通过加强网络在较长时间序列中提取时空特征的能力,避免了训练深度脉冲神经网络,使得浅层改进脉冲神经网络在神经形态视觉目标分类任务中也能得到较高的准确率;该方法具体包括以下步骤:1. A neuromorphic visual target classification method based on an improved spiking neural network, which is characterized in that the method first aggregates the spiking event stream data into an event frame sequence, and uses a time random clipping method to determine the data input to the network when training the network, This is equivalent to increasing the amount of data available for network training; then we improved the construction of a spiking neural network based on improved LIF neurons, which avoids training a deep spiking neural network by strengthening the network's ability to extract spatiotemporal features in longer time series. This enables the shallow improved spiking neural network to achieve higher accuracy in neuromorphic visual target classification tasks; the method specifically includes the following steps: S1:获取神经形态视觉目标分类数据集;S1: Obtain neuromorphic visual object classification data set; S2:脉冲事件流序列化聚合:将神经形态视觉目标分类数据集中的时空脉冲事件流数据,按照设定的时间分辨率dt聚合成新的事件帧序列数据;具体分三步进行:S2: Pulse event stream serialization aggregation: Aggregate the spatio-temporal pulse event stream data in the neuromorphic visual target classification data set into new event frame sequence data according to the set time resolution dt; this is done in three steps: 第一步,获取神经形态视觉目标分类数据集中时空脉冲事件流,由集合E={ei|ei=[xi,yi,ti ,pi]}确定;其中ei为脉冲事件流中的第i个脉冲事件,(xi,yi)为第i个脉冲事件的像素坐标,ti 为第i个脉冲事件在整个时间流中的时间戳,pi为第i个脉冲事件的光强变化极性,事件相机异步输脉冲事件流的时间分辨率为dt,空间分辨率为H×W;The first step is to obtain the spatio-temporal pulse event stream in the neuromorphic visual target classification data set, which is determined by the set E={e i |e i =[x i ,y i ,t i ,pi ] }; where e i is the pulse For the i-th pulse event in the event stream, (xi , y i ) is the pixel coordinate of the i-th pulse event, ti ′ is the timestamp of the i-th pulse event in the entire time stream, and p i is the i-th pulse event. The light intensity change polarity of each pulse event, the time resolution of the asynchronous pulse event stream transmitted by the event camera is dt , and the spatial resolution is H×W; 第二步,基于事件相机输出的脉冲事件流,聚合时间分辨率为dt的事件帧,以t时刻聚合为例,将t时刻产生的若干个事件集合Et′组装成张量Xt′,其中,Et′={ei|ei=[xi,yi,t,pi]},Xt′∈RH×W×2The second step is to aggregate event frames with a time resolution of dt based on the pulse event stream output by the event camera. Taking the aggregation at time t as an example, assemble several event sets E t′ generated at time t ′ into a tensor X t′ , where E t′ ={e i |e i =[x i ,y i ,t ,p i ]}, X t′ ∈R H×W×2 ; 第三步,基于时间分辨率为dt的事件帧,Xt=f(X t)生成t时刻的事件帧张量Xt∈RH ×W×2,其中,dt=β×dt,β为聚合时间因子;X t={Xt′|t′∈[β×t,β×(t+1)-1]};f为累加操作一般性计算操作;The third step is to generate the event frame tensor X t ∈R H ×W×2 at time t based on the event frame with time resolution dt , X t =f(X t ), where dt=β×dt , β is the aggregation time factor ; S3:构建改进脉冲神经网络模型:改进泄露-积累-发射LIF脉冲神经元在时间维度上的突触连接方式,基于改进LIF神经元层搭建改进脉冲神经网络;改进LIF脉冲神经元层数学表达式为:S3: Construct an improved spiking neural network model: improve the synaptic connection method of leakage-accumulation-emitting LIF spiking neurons in the time dimension, build an improved spiking neural network based on the improved LIF neuron layer; improve the mathematical expression of the LIF spiking neuron layer for: 其中,xt,n-1为t时刻第n-1层输入的事件帧,ht-1,n为t-1时刻第n层的内部状态量,ut,n为t时刻第n层的膜电势,xt,n为t时刻第n层传递到下一层的空间输出,ht,n为t时刻第n层传递到下一时刻的时间输出;Among them, x t,n-1 is the event frame input by the n-1th layer at time t, h t-1,n is the internal state quantity of the nth layer at time t-1, u t,n is the nth layer at time t The membrane potential of RNNs_Model构成模型的网络为循环神经网络;The network forming the model of RNNs_Model is a recurrent neural network; 所述LIF_Model的数学表达式为:The mathematical expression of the LIF_Model is: 其中,函数g为阶跃函数, 表示Hadamard乘积;Among them, function g is a step function, Represents Hadamard product; 基于改进LIF脉冲神经元层,LIF神经元层和全连接层,构建改进脉冲神经网络;Based on the improved LIF spiking neuron layer, LIF neuron layer and fully connected layer, an improved spiking neural network is constructed; S4:对于脉冲事件流序列化聚合后的数据集,从序列中随机抽取样本作为输入,训练和测试所构建的改进脉冲神经网络;S4: For the data set after the serialization and aggregation of the pulse event stream, samples are randomly selected from the sequence as input, and the improved pulse neural network constructed is trained and tested; S5:保存训练好的改进脉冲神经网络结构和网络参数,该网络参数即为对应的神经形态视觉目标分类所需要的参数。S5: Save the trained improved spiking neural network structure and network parameters. The network parameters are the parameters required for the corresponding neuromorphic visual target classification. 2.根据权利要求1所述的一种基于改进脉冲神经网络的神经形态视觉目标分类方法,其特征在于,步骤S4具体包括:2. A neuromorphic visual target classification method based on an improved spiking neural network according to claim 1, characterized in that step S4 specifically includes: 在训练改进脉冲神经网络时,若事件帧数据按照设定的时间分辨率dt一共生成的Ttotal个事件帧,从中随机抽取T个事件帧作为训练输入,T<TtotalWhen training the improved spiking neural network, if the event frame data generates a total of T total event frames according to the set time resolution dt, and T event frames are randomly selected from them as training input, T < T total ; 在测试改进脉冲神经网络时,使用滑动窗口方法在时间维度上对事件帧序列进行预处理,获得固定大小为T的预分割片段,作为测试数据;预分割片段设置为m个,对同一个数据预分割出来的m份测试数据输入网络得到m个预测结果,当有m/2个以上的结果预测正确时,判定这个测试数据预测正确。When testing the improved spiking neural network, the sliding window method is used to preprocess the event frame sequence in the time dimension, and pre-segmented fragments of a fixed size T are obtained as test data; the pre-segmented fragments are set to m, and the same data is The pre-segmented m test data are input into the network to obtain m prediction results. When more than m/2 results are predicted correctly, the test data is judged to be correctly predicted.
CN202110025987.2A 2021-01-08 2021-01-08 Neuromorphic visual target classification method based on improved impulse neural network Active CN112699956B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110025987.2A CN112699956B (en) 2021-01-08 2021-01-08 Neuromorphic visual target classification method based on improved impulse neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110025987.2A CN112699956B (en) 2021-01-08 2021-01-08 Neuromorphic visual target classification method based on improved impulse neural network

Publications (2)

Publication Number Publication Date
CN112699956A CN112699956A (en) 2021-04-23
CN112699956B true CN112699956B (en) 2023-09-22

Family

ID=75513519

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110025987.2A Active CN112699956B (en) 2021-01-08 2021-01-08 Neuromorphic visual target classification method based on improved impulse neural network

Country Status (1)

Country Link
CN (1) CN112699956B (en)

Families Citing this family (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113205048B (en) * 2021-05-06 2022-09-09 浙江大学 Gesture recognition method and recognition system
CN113255905B (en) * 2021-07-16 2021-11-02 成都时识科技有限公司 Signal processing method of neurons in impulse neural network and network training method
CN113515468B (en) * 2021-09-14 2021-11-30 成都时识科技有限公司 Data updating method, storage space setting method, device, chip and equipment
CN114065806B (en) * 2021-10-28 2022-12-20 贵州大学 Manipulator touch data classification method based on impulse neural network
CN114282647B (en) * 2021-12-09 2024-02-02 上海应用技术大学 Pulse neural network-based target detection method for neuromorphic vision sensor
CN114387277B (en) * 2021-12-17 2025-04-22 中国电子科技南湖研究院 Method and device for realizing target segmentation based on event sequence training neural network
CN114358068B (en) * 2021-12-24 2024-10-18 厦门大学 Space potential measurement method and system for neural network small sample learning
CN113988276B (en) * 2021-12-27 2022-08-30 中科南京智能技术研究院 Target identification method and system
CN114332545B (en) * 2022-03-17 2022-08-05 之江实验室 Image data classification method and device based on low-bit pulse neural network
CN115048979A (en) * 2022-04-29 2022-09-13 贵州大学 Robot touch pulse data classification method based on regularization
CN115984124A (en) * 2022-11-29 2023-04-18 北京大学 A neuromorphic pulse signal denoising and super-resolution method and device
CN116805391A (en) * 2023-02-05 2023-09-26 西北工业大学 Aircraft visual target sensing and positioning method based on impulse neural network
CN116453014B (en) * 2023-03-24 2025-05-30 大连理工大学 Multi-mode road scene target detection method based on images and events
CN116824326B (en) * 2023-04-17 2025-10-28 电子科技大学 Spiking neural network image recognition method based on the fusion of probability and attention mechanism
CN116394264B (en) * 2023-06-07 2023-08-25 安徽大学 Coordinated motion planning method for multi-manipulators based on population-encoded spiking neural network
CN117253039A (en) * 2023-09-25 2023-12-19 北京理工大学 Target segmentation method and system based on impulse neural network
CN117232638B (en) * 2023-11-15 2024-02-20 常州检验检测标准认证研究院 Robot vibration detection method and system
CN117934850B (en) * 2024-01-31 2024-11-22 重庆大学 A semantic segmentation method based on event and spiking neural network
CN118521822B (en) * 2024-05-15 2025-01-24 中国人民解放军军事航天部队航天工程大学 A neuromorphic data classification method based on multi-scale residual spiking neural network

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110210563A (en) * 2019-06-04 2019-09-06 北京大学 The study of pattern pulse data space time information and recognition methods based on Spike cube SNN
WO2020124902A1 (en) * 2018-12-19 2020-06-25 中国科学院自动化研究所 Supervised learning auditory attention-based voice extraction method and system, and apparatuses
CN111460906A (en) * 2020-03-05 2020-07-28 重庆大学 A spiking neural network pattern recognition method and system based on ensemble learning
CN111612136A (en) * 2020-05-25 2020-09-01 之江实验室 A neuromorphic visual target classification method and system
WO2020210673A1 (en) * 2019-04-10 2020-10-15 Cornell University Neuromorphic algorithm for rapid online learning and signal restoration
CN112130118A (en) * 2020-08-19 2020-12-25 复旦大学无锡研究院 SNN-based ultra-wideband radar signal processing system and processing method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020124902A1 (en) * 2018-12-19 2020-06-25 中国科学院自动化研究所 Supervised learning auditory attention-based voice extraction method and system, and apparatuses
WO2020210673A1 (en) * 2019-04-10 2020-10-15 Cornell University Neuromorphic algorithm for rapid online learning and signal restoration
CN110210563A (en) * 2019-06-04 2019-09-06 北京大学 The study of pattern pulse data space time information and recognition methods based on Spike cube SNN
CN111460906A (en) * 2020-03-05 2020-07-28 重庆大学 A spiking neural network pattern recognition method and system based on ensemble learning
CN111612136A (en) * 2020-05-25 2020-09-01 之江实验室 A neuromorphic visual target classification method and system
CN112130118A (en) * 2020-08-19 2020-12-25 复旦大学无锡研究院 SNN-based ultra-wideband radar signal processing system and processing method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
庄祖江 ; 房玉 ; 雷建超 ; 刘栋博 ; 王海滨 ; .基于STDP规则的脉冲神经网络研究.计算机工程.2020,(第09期),全文. *
程龙 ; 刘洋 ; .脉冲神经网络:模型、学习算法与应用.控制与决策.2018,(第05期),全文. *

Also Published As

Publication number Publication date
CN112699956A (en) 2021-04-23

Similar Documents

Publication Publication Date Title
CN112699956B (en) Neuromorphic visual target classification method based on improved impulse neural network
CN111858989B (en) An image classification method based on attention mechanism and spiking convolutional neural network
Li et al. Sign language recognition based on computer vision
CN111242063B (en) Construction method of small sample classification model based on transfer learning and application of iris classification
CN109829541A (en) Deep neural network incremental training method and system based on learning automaton
CN113205048B (en) Gesture recognition method and recognition system
CN114861838B (en) Intelligent classification method for pulsatile neural brains based on neuron complex dynamics
CN114998659B (en) Image data classification method using spiking neural network model trained online over time
CN111612136A (en) A neuromorphic visual target classification method and system
CN111723667A (en) Method and device for crowd behavior recognition of smart light poles based on human body joint point coordinates
CN116822592B (en) Target tracking method based on event data and impulse neural network
CN115908896A (en) Image recognition system based on spiking neural network with self-attention mechanism
CN114624715A (en) A Radar Echo Extrapolation Method Based on Self-Attention Spatiotemporal Neural Network Model
Niu et al. ALSTM: adaptive LSTM for durative sequential data
CN113988276B (en) Target identification method and system
CN117793553A (en) Video frame inserting method based on impulse neural network and event camera
CN110378469A (en) SCNN inference device based on asynchronous circuit, PE unit, processor and computer equipment thereof
CN117593790A (en) Segment-level multi-scale action segmentation method based on expansion attention mechanism
Liu et al. Real-time target tracking system with spiking neural networks implemented on neuromorphic chips
CN111209433A (en) Video classification algorithm based on feature enhancement
CN115841702A (en) Action classification method for dynamic vision sensor
CN117237658A (en) Feature extractor and method, point cloud processing network and chip
Cheng et al. An FPGA-Based Event-Driven SNN Accelerator for DVS Applications With Structured Sparsity and Early-Stop
CN116503787A (en) Handwritten Digit Recognition Method Based on Lightweight Spiking Neural Network Based on Tensor Decomposition
CN114330681A (en) A ST-GRU memristive neural network circuit and training method for gait prediction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant