CN103679267A - Method and device for constructing RBF neural network based on unmarked samples - Google Patents
Method and device for constructing RBF neural network based on unmarked samples Download PDFInfo
- Publication number
- CN103679267A CN103679267A CN201310654934.2A CN201310654934A CN103679267A CN 103679267 A CN103679267 A CN 103679267A CN 201310654934 A CN201310654934 A CN 201310654934A CN 103679267 A CN103679267 A CN 103679267A
- Authority
- CN
- China
- Prior art keywords
- neural network
- hidden layer
- neurons
- rbf neural
- module
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 84
- 238000000034 method Methods 0.000 title claims abstract description 36
- 210000002569 neuron Anatomy 0.000 claims abstract description 83
- 238000012549 training Methods 0.000 claims abstract description 27
- 230000035945 sensitivity Effects 0.000 claims abstract description 24
- 238000010276 construction Methods 0.000 claims abstract description 14
- 238000013507 mapping Methods 0.000 claims description 5
- 238000004364 calculation method Methods 0.000 abstract 1
- 230000006870 function Effects 0.000 description 16
- 238000013459 approach Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000004836 empirical method Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000011946 reduction process Methods 0.000 description 1
- 238000010187 selection method Methods 0.000 description 1
Images
Landscapes
- Feedback Control In General (AREA)
Abstract
本发明公开一种基于无标记样本的RBF神经网络构建方法及其装置,方法以RBF神经网络中隐层神经元的敏感性为基准,通过逐渐裁剪隐层神经元中敏感性低的神经元达到逐渐简化网络结构、提高分类器性能的目的。与传统方式相比,在计算RBF隐层神经元敏感性的过程中,本发明不但使用了有标记的样本,还大量使用了没有标记的样本,这样大大提高了敏感性计算精度。装置包括初始模块、训练模块、隐层神经元选择模块和输出模块,可以有效提高神经网络构建效率,并提高RBF神经网络的性能。
The invention discloses a method and device for constructing an RBF neural network based on unmarked samples. The method takes the sensitivity of the hidden layer neurons in the RBF neural network as a benchmark, and gradually cuts out neurons with low sensitivity in the hidden layer neurons to achieve The purpose of gradually simplifying the network structure and improving the performance of the classifier. Compared with the traditional method, in the process of calculating the sensitivity of RBF hidden layer neurons, the present invention not only uses marked samples, but also uses a large number of unmarked samples, which greatly improves the accuracy of sensitivity calculation. The device includes an initial module, a training module, a hidden layer neuron selection module and an output module, which can effectively improve the construction efficiency of the neural network and improve the performance of the RBF neural network.
Description
技术领域technical field
本发明涉及神经网络设计时的网络构建方法及其装置,尤其涉及一种可有效提高神经网络分类效率或回归效率的网络构建方法及其装置,属于智能科学与技术中的机器学习领域。The invention relates to a network construction method and a device for designing a neural network, in particular to a network construction method and a device that can effectively improve the classification efficiency or regression efficiency of a neural network, and belongs to the field of machine learning in intelligent science and technology.
背景技术Background technique
在设计RBF神经网络分类器时,如何确定神经网络的结构是一个重要而且关键的步骤。针对具体问题构建一个合适的网络对提高分类精度、泛化能力都有巨大的帮助。目前广泛使用的是三层神经网络。文献已经证明,三层神经网络当第二层(也称隐层、中间层)神经元数增多时可以逼近任何连续函数。在具体应用中,三层神经网络的第一层神经元数依赖于输入变量的维数,第三层神经元数依赖于输出变量的维数。因为输入变量和输出变量的维数一般都是确知的,所以第一层和第三层神经元的个数一般也是确定的。对于三层神经网络的网络构建实际上是确定第二层神经元数目的一个过程。When designing the RBF neural network classifier, how to determine the structure of the neural network is an important and critical step. Constructing a suitable network for specific problems is of great help to improve classification accuracy and generalization ability. Currently, three-layer neural networks are widely used. The literature has proved that the three-layer neural network can approach any continuous function when the number of neurons in the second layer (also known as the hidden layer and the middle layer) increases. In a specific application, the number of neurons in the first layer of the three-layer neural network depends on the dimension of the input variable, and the number of neurons in the third layer depends on the dimension of the output variable. Because the dimensions of the input variables and output variables are generally known, the number of neurons in the first layer and the third layer is generally determined. The network construction of the three-layer neural network is actually a process of determining the number of neurons in the second layer.
在训练神经网络的时候,通常使用有监督的学习方法。有监督的学习方法是指在训练神经网络的过程中通过告诉网络输入和相对应的输出来调节网络参数,达到训练神经网络的目的。在这个过程中需要使用有标记的训练样本。训练样本的标记一般是由专家来完成的,这往往要花费大量的金钱和时间代价。在实际应用中,无标记的样本比有标记的样本要容易获得的多。例如在某些互联网应用中,由专家标记的有标记样本和无标记样本相比只占很少的一部分。能否使用那些没有标记的样本帮助辅助确定网络的最优结构就变得很有必要。When training neural networks, supervised learning methods are usually used. The supervised learning method refers to adjusting the network parameters by telling the network input and corresponding output in the process of training the neural network, so as to achieve the purpose of training the neural network. This process requires the use of labeled training samples. The labeling of training samples is generally done by experts, which often costs a lot of money and time. In practical applications, unlabeled samples are much easier to obtain than labeled samples. For example, in some Internet applications, the labeled samples marked by experts account for only a small part compared to the unlabeled samples. It becomes necessary to be able to use those unlabeled samples to help determine the optimal structure of the network.
对于三层RBF神经网络,确定隐层神经网络的方法主要有:For the three-layer RBF neural network, the methods for determining the hidden layer neural network mainly include:
1)增枝法。这个方法确定隐层神经元数量的过程是首先选择一个很小的隐藏神经元数。由于隐层神经元数太少,所以网络的结构太简单,导致使用训练样本训练神经网络不成功。数学上的表征就是误差不收敛。在这个隐层神经元数量的基础上一个一个地增加隐层神经元数量,每增加一个隐层神经元重新训练一次网络,直到隐层神经元数量增加到某一数量时神经网络能够训练成功为止。能够使得神经网络训练成功的最小隐层神经元数量就是我们需要寻找的隐层神经元数。1) Branching method. The process of determining the number of hidden neurons in this method is to first choose a small number of hidden neurons. Because the number of neurons in the hidden layer is too small, the structure of the network is too simple, resulting in the failure of training the neural network with training samples. The mathematical characterization is that the error does not converge. On the basis of the number of neurons in the hidden layer, increase the number of neurons in the hidden layer one by one, and retrain the network every time a neuron in the hidden layer is added, until the number of neurons in the hidden layer increases to a certain number, and the neural network can be trained successfully. . The minimum number of hidden layer neurons that can make the neural network training successful is the number of hidden layer neurons we need to find.
2)减枝法。这个方法与增枝法相反,它的操作方法是首先确定一个足够大的隐层神经元数来构造一个三层神经网络,在这个结构下能够使用有标记的样本很容易地训练好神经网络。然后对隐层神经元去除一个,在去除之后的网络基础上使用有标记的样本进行训练,使得网络再次训练完成。重复上述去除过程,直到网络训练不能完成为止。这时候取最小能完成的隐层神经元数量作为最终确定的隐层神经元数。增枝法和减枝法背后的理论基础是统计学习理论要求对于一个具体的分类问题分类器要有一个合适的复杂度,保证既不过拟合也不欠拟合。只有这样的分类器才能具有最好的泛化能力。对于三层RBF神经网络这样的分类器而言,网络复杂度就体现在隐层神经元的数量上,神经元数量太少网络欠拟合,训练不能完成,神经元数量太多网络过拟合,泛化能力差。2) Branch reduction method. This method is opposite to the branching method. Its operation method is to first determine a sufficiently large number of neurons in the hidden layer to construct a three-layer neural network. Under this structure, the neural network can be easily trained using labeled samples. Then remove one of the neurons in the hidden layer, and use the marked samples for training on the basis of the removed network, so that the network is trained again. The above removal process is repeated until the network training cannot be completed. At this time, the minimum number of neurons in the hidden layer that can be completed is taken as the final number of neurons in the hidden layer. The theoretical basis behind the branching method and the branching method is that the statistical learning theory requires a classifier to have an appropriate complexity for a specific classification problem, ensuring that it is neither overfitting nor underfitting. Only such a classifier can have the best generalization ability. For a classifier such as a three-layer RBF neural network, the complexity of the network is reflected in the number of neurons in the hidden layer. If the number of neurons is too small, the network is underfitting, and the training cannot be completed. If the number of neurons is too large, the network is overfitting. , poor generalization ability.
3)经验法。这种方法确定隐层神经元数需要对具体问题所涉及的领域有深刻的理解,从而凭借经验确定隐层神经元的数量。即使这样也不能保证所取的隐层神经元数量是最优的。3) Empirical method. This method to determine the number of neurons in the hidden layer requires a deep understanding of the domain involved in the specific problem, so that the number of neurons in the hidden layer can be determined empirically. Even this cannot guarantee that the number of neurons in the hidden layer is optimal.
对于上述方法减枝法目前使用较多。在具体的减枝过程中,首先减去哪个隐层神经元、其次减去哪个隐层神经元对于确定最终的网络结构非常重要。一般认为每个隐层神经元在训练过程中起到的作用或者重要程度是不一样的。理论上首先去掉对分类没有作用的或者不重要的神经元可以使得最终训练完成的神经网络的泛化性能更好。如何利用无标记的样本来辅助确定隐层神经元以更好地确定网络就变得非常重要。For the above method, the branch reduction method is currently used more. In the specific branch reduction process, which hidden layer neuron is subtracted first, and which hidden layer neuron is subtracted second is very important for determining the final network structure. It is generally believed that each hidden layer neuron plays a different role or importance in the training process. In theory, removing neurons that have no effect on classification or are unimportant can make the generalization performance of the final trained neural network better. How to use unlabeled samples to help determine hidden layer neurons to better determine the network becomes very important.
发明内容Contents of the invention
发明目的:针对现有技术中存在的问题,本发明提供一种基于无标记样本的RBF神经网络构建方法及其装置。Purpose of the invention: Aiming at the problems existing in the prior art, the present invention provides a method and device for constructing a RBF neural network based on unlabeled samples.
技术方案:一种基于无标记样本的RBF神经网络构建方法,包括以下步骤:Technical solution: a method for constructing a RBF neural network based on unlabeled samples, comprising the following steps:
(S101)选取一个足够大的正整数m(m最大不超过训练样本的个数)作为隐层神经元数,构建一个三层RBF神经网络,并给定初始网络参数;(S101) Select a sufficiently large positive integer m (the maximum number of m does not exceed the number of training samples) as the number of neurons in the hidden layer, construct a three-layer RBF neural network, and give initial network parameters;
(S103)利用有标记的样本集训练神经网络直到代价函数收敛到某个给定的很小的阈值e(e取值小于10的-2次方),得到经过训练的分类器;(S103) Using the labeled sample set to train the neural network until the cost function converges to a given small threshold e (the value of e is less than 10 to the power of -2), and a trained classifier is obtained;
(S105)利用有标记和无标记样本计算隐层神经元的敏感性,并按照敏感性由小到大排序;(S105) Calculate the sensitivity of neurons in the hidden layer using labeled and unlabeled samples, and sort the sensitivity from small to large;
(S107)去掉敏感性最小的隐层神经元,得到新结构的RBF神经网络;(S107) removing the least sensitive hidden layer neurons to obtain a new structured RBF neural network;
(S109)对新的RBF神经网络在原有参数的基础上再次使用有标记的样本集进行训练,如果代价函数能够收敛到某个很小的阈值e,则得到经过参数更新的分类器并重复步骤(S107)、(S109);如果不能收敛则进入下一步;(S109) On the basis of the original parameters, the new RBF neural network is trained again using the marked sample set. If the cost function can converge to a small threshold e, a classifier with updated parameters is obtained and the steps are repeated (S107), (S109); if it cannot converge, go to the next step;
(S111)取隐层神经元数最小且能收敛的RBF神经网络网络结构为最终的网络结构,其网络为最终输出的分类器。(S111) Taking the RBF neural network structure with the smallest number of neurons in the hidden layer and capable of convergence as the final network structure, and its network is the final output classifier.
本发明所述方法为裁剪隐层神经元以确定网络结构提供依据,提高了神经网络分类器的分类性能。The method of the invention provides a basis for cutting hidden layer neurons to determine the network structure, and improves the classification performance of the neural network classifier.
本发明中还公开了一种基于无标记样本的RBF神经网络构建装置,包括:初始模块、训练模块、隐层神经元选择模块和输出模块;上述模块按照如下顺序依次构建:初始模块、训练模块、隐层神经元选择模块、输出模块。The present invention also discloses a RBF neural network construction device based on unlabeled samples, including: an initial module, a training module, a hidden layer neuron selection module and an output module; the above-mentioned modules are sequentially constructed in the following order: initial module, training module , a hidden layer neuron selection module, and an output module.
(1)构建初始模块:其选取一个足够大的正整数m作为隐层神经元数,构建一个三层RBF神经网络,并给定初始网络参数;(1) Construct the initial module: it selects a sufficiently large positive integer m as the number of neurons in the hidden layer, constructs a three-layer RBF neural network, and gives initial network parameters;
(2)构建训练模块:其利用有标记的样本集训练神经网络直到代价函数收敛到某个给定的很小的阈值e,得到经过训练的分类器;(2) Build a training module: it uses a labeled sample set to train the neural network until the cost function converges to a given small threshold e, and obtains a trained classifier;
(3)隐层神经元选择模块:其利用有标记的样本和无标记样本计算隐层神经元的敏感性,按照敏感性由小到大排序,并去掉敏感性最小的隐层神经元,形成新结构的RBF神经网络;重新利用有标记样本集训练新结构的RBF神经网络,如果代价函数能够收敛到某个很小的阈值e,则得到经过参数更新的分类器并重复本步骤;如果代价函数不能够收敛到某个很小的阈值e,则进入下一个步骤;(3) Hidden layer neuron selection module: it uses labeled samples and unlabeled samples to calculate the sensitivity of hidden layer neurons, sorts the sensitivity from small to large, and removes the least sensitive hidden layer neurons to form RBF neural network with a new structure; re-use the labeled sample set to train the RBF neural network with a new structure, if the cost function can converge to a small threshold e, get a classifier with updated parameters and repeat this step; if the cost If the function cannot converge to a small threshold e, enter the next step;
(4)构建输出模块:取隐层神经元数最小且能收敛的神经网络网络结构为最终的网络结构,并输出这个网络为最终的分类器。(4) Build the output module: take the neural network structure with the smallest number of neurons in the hidden layer that can converge as the final network structure, and output this network as the final classifier.
本发明所属基于无标记样本的RBF神经网络构建装置有效利用了无标记样本所蕴含的信息,从而在判断隐层神经元个体重要性方面比以往的方法更加准确,更加方便。The RBF neural network construction device based on unlabeled samples of the present invention effectively utilizes the information contained in unlabeled samples, so that it is more accurate and convenient than previous methods in judging the importance of individual neurons in the hidden layer.
与现有技术相比,本发明具有以下有益效果:Compared with the prior art, the present invention has the following beneficial effects:
本发明所述基于无标记样本的RBF神经网络构建方法为裁剪隐层神经元以确定网络结构提供依据,提高了神经网络分类器的分类性能。The RBF neural network construction method based on unmarked samples in the invention provides a basis for clipping hidden layer neurons to determine the network structure, and improves the classification performance of the neural network classifier.
本发明所述基于无标记样本的RBF神经网络构建装置有效利用无标记样本所蕴含的信息,从而在判断隐层神经元个体重要性方面比以往的方法更加准确,更加方便,同时采用基于无标记样本的RBF神经网络构建装置的工作方法裁剪过的神经网络具有更好的泛化性。The RBF neural network construction device based on unlabeled samples of the present invention effectively utilizes the information contained in unlabeled samples, so that it is more accurate and convenient than previous methods in judging the importance of individual neurons in the hidden layer. The working method of the sample RBF neural network construction device The pruned neural network has better generalization.
附图说明Description of drawings
图1为RBF神经网络结构图;Fig. 1 is the structural diagram of RBF neural network;
图2为本发明实施例的基于无标记样本的RBF神经网络构建方法流程图。FIG. 2 is a flow chart of a method for constructing an RBF neural network based on unlabeled samples according to an embodiment of the present invention.
具体实施方式Detailed ways
下面结合具体实施例,进一步阐明本发明,应理解这些实施例仅用于说明本发明而不用于限制本发明的范围,在阅读了本发明之后,本领域技术人员对本发明的各种等价形式的修改均落于本申请所附权利要求所限定的范围。Below in conjunction with specific embodiment, further illustrate the present invention, should be understood that these embodiments are only used to illustrate the present invention and are not intended to limit the scope of the present invention, after having read the present invention, those skilled in the art will understand various equivalent forms of the present invention All modifications fall within the scope defined by the appended claims of the present application.
现以RBF神经网络为例,说明根据本发明的前向神经网络样本选择方法。Taking the RBF neural network as an example, the sample selection method of the feedforward neural network according to the present invention is described.
RBF神经网络是一种全连接的前向神经网络,适用于目标的分类。RBF神经网络的结构如图1所示,它是一种三层前向网络:输入层MA由输入模式节点组成,xi表示输入模式向量的第i个分量(i=1,2,...,n);第二层是隐含层MB,它由m个节点bj(j=1,2,...,m)组成。第三层是输出层MC,它由p个节点ck(k=1,2,...,p)组成。The RBF neural network is a fully connected feedforward neural network, which is suitable for object classification. The structure of the RBF neural network is shown in Figure 1. It is a three-layer forward network: the input layer MA is composed of input pattern nodes, and x i represents the i-th component of the input pattern vector (i=1,2,.. .,n); the second layer is the hidden layer MB, which consists of m nodes b j (j=1,2,...,m). The third layer is the output layer MC, which consists of p nodes c k (k=1,2,...,p).
在训练之前需要对输入向量的每个元素规范化,这里将每个元素规范化到[-1,1]。Each element of the input vector needs to be normalized before training, where each element is normalized to [-1,1].
对于上述RBF神经网络的训练在这里采用标准BP算法。The standard BP algorithm is used here for the training of the above-mentioned RBF neural network.
下面我们定义上述RBF神经网络隐层神经元的敏感性,这样的定义可以很容易地推广到其它前向神经网络。Below we define the sensitivity of the neurons in the hidden layer of the RBF neural network above, and such a definition can be easily extended to other forward neural networks.
当神经网络训练完成后,它的映射关系也就确定了。设映射关系函数为F(X)(其中X为输入向量)。假设裁掉的是第j个隐层神经元,那么去掉第j个隐层神经元后神经网络的映射关系函数变成Fj(X),定义第j个隐层神经元的敏感性为:When the training of the neural network is completed, its mapping relationship is determined. Let the mapping relationship function be F(X) (where X is the input vector). Assuming that the jth hidden layer neuron is cut off, the mapping relationship function of the neural network becomes F j (X) after removing the jth hidden layer neuron, and the sensitivity of the jth hidden layer neuron is defined as:
Sj(X)=E(||F(X)-Fj(X)||2) (1)S j (X)=E(||F(X)-F j (X)|| 2 ) (1)
||·||2是求取·的欧几里德范数的算符。E为求取期望的算符。||·|| 2 is an operator for obtaining the Euclidean norm of ·. E is the operator for obtaining expectations.
由敏感性的定义可以看出,敏感性实际上代表了有第j个隐层神经元和没有第j个隐层神经元函数输出的差别。这个差别越小表示第j个隐层神经元越不重要,反之则越重要。From the definition of sensitivity, it can be seen that the sensitivity actually represents the difference between the function output of the jth hidden layer neuron and the function output without the jth hidden layer neuron. The smaller the difference, the less important the jth hidden layer neuron is, and vice versa.
在计算Sj(X)的过程中,Sj(X)可以变形为:In the process of calculating S j (X), S j (X) can be transformed into:
Sj(X)=∫Ω(F(X)-Fj(X))2p(X)dX (2)S j (X)= ∫Ω (F(X)-F j (X)) 2 p(X)dX (2)
其中Ω是定义域,p(X)为X在定义域Ω中的密度函数。一般来说p(X)是未知的,由于有标记的样本数量很少,只依靠有标记的样本估计p(X)非常困难。无标记的样本数量非常多,并且里面也蕴含了样本的分布信息。这里我们采用有标记的样本和无标记的样本一起估计p(X),这能使估计精度大大提高。Where Ω is the domain, and p(X) is the density function of X in the domain Ω. Generally speaking, p(X) is unknown. Due to the small number of marked samples, it is very difficult to estimate p(X) only relying on marked samples. The number of unlabeled samples is very large, and it also contains the distribution information of the samples. Here we use labeled samples and unlabeled samples to estimate p(X), which can greatly improve the estimation accuracy.
假设(X1,y1),(X2,y2),…(Xb,yb)是有标记的样本集,Xb+1,Xb+2…XN是无标记的样本集,那么Sj(X)的一个好的估计是:Suppose (X 1 ,y 1 ),(X 2 ,y 2 ),…(X b ,y b ) are labeled sample sets, X b+1 ,X b+2 …X N are unlabeled sample sets , then a good estimate of S j (X) is:
本文采用(3)式计算隐层神经元的敏感性。In this paper, formula (3) is used to calculate the sensitivity of neurons in the hidden layer.
如图2所示为本发明基于无标记样本的神经网络构建方法流程图。FIG. 2 is a flow chart of the neural network construction method based on unlabeled samples in the present invention.
在步骤S101中,选取一个足够大的正整数m作为RBF隐层神经元数,构建一个三层神经网络,并给定初始网络参数。In step S101, a sufficiently large positive integer m is selected as the number of neurons in the RBF hidden layer, a three-layer neural network is constructed, and initial network parameters are given.
在步骤S103中,利用有标记的样本集训练神经网络直到代价函数收敛到某个给定的很小的阈值e,得到经过训练的分类器。In step S103, the neural network is trained using the labeled sample set until the cost function converges to a given small threshold e, and a trained classifier is obtained.
在步骤S105中,利用有标记和无标记样本计算隐层神经元的敏感性,并按照敏感性由小到大排序。In step S105, the sensitivities of neurons in the hidden layer are calculated by using the labeled and unlabeled samples, and sorted according to the sensitivity from small to large.
在步骤S107中,去掉敏感性最小的隐层神经元,得到新结构的RBF神经网络。In step S107, the neuron in the hidden layer with the least sensitivity is removed to obtain an RBF neural network with a new structure.
在步骤S109中,对新的RBF神经网络在原有参数的基础上再次使用有标记的样本集进行训练,如果代价函数能够收敛到某个很小的阈值e,则得到经过参数更新的分类器并重复步骤S107、S109。如果不能收敛则进入下一步。In step S109, on the basis of the original parameters, the new RBF neural network is trained again using the marked sample set. If the cost function can converge to a small threshold e, the classifier with updated parameters is obtained and Repeat steps S107 and S109. If it cannot converge, go to the next step.
在步骤S111中,取隐层神经元数最小且能收敛的RBF神经网络网络结构为最终的网络结构,其网络为最终输出的分类器。In step S111, the RBF neural network structure with the smallest number of neurons in the hidden layer and capable of convergence is taken as the final network structure, and its network is the final output classifier.
本实施例所属方法基于所述基于无标记样本的神经网络构建装置,该装置包括:初始模块、训练模块、隐层神经元选择模块和输出模块;上述模块按照如下顺序依次构建:初始模块、训练模块、隐层神经元选择模块、输出模块。The method of this embodiment is based on the neural network construction device based on unlabeled samples. The device includes: an initial module, a training module, a hidden layer neuron selection module, and an output module; the above modules are sequentially constructed in the following order: initial module, training module, hidden layer neuron selection module, and output module.
(1)构建初始模块:其选取一个足够大的正整数m作为RBF隐层神经元数,构建一个三层神经网络,并给定初始网络参数;(1) Construct the initial module: it selects a sufficiently large positive integer m as the number of neurons in the RBF hidden layer, constructs a three-layer neural network, and gives initial network parameters;
(2)构建训练模块:其利用有标记的样本集训练RBF神经网络直到代价函数收敛到某个给定的很小的阈值e,得到经过训练的分类器;(2) Build a training module: it uses a labeled sample set to train the RBF neural network until the cost function converges to a given small threshold e to obtain a trained classifier;
(3)隐层神经元选择模块:其利用有标记的样本和无标记样本计算隐层神经元的敏感性,按照敏感性由小到大排序,并去掉敏感性最小的隐层神经元,形成新结构的RBF神经网络;重新利用有标记样本集训练新结构的神经网络,如果代价函数能够收敛到某个很小的阈值e,则得到经过参数更新的分类器并重复本步骤;如果代价函数不能够收敛到某个很小的阈值e,则进入下一个步骤;(3) Hidden layer neuron selection module: it uses labeled samples and unlabeled samples to calculate the sensitivity of hidden layer neurons, sorts the sensitivity from small to large, and removes the least sensitive hidden layer neurons to form RBF neural network with a new structure; re-use the labeled sample set to train the neural network with a new structure, if the cost function can converge to a small threshold e, get a classifier with updated parameters and repeat this step; if the cost function If it cannot converge to a small threshold e, enter the next step;
(4)构建输出模块:取隐层神经元数最小且能收敛的RBF神经网络网络结构为最终的网络结构,并输出这个网络为最终的分类器。(4) Build the output module: take the RBF neural network structure with the smallest number of neurons in the hidden layer and can converge as the final network structure, and output this network as the final classifier.
Claims (3)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201310654934.2A CN103679267A (en) | 2013-12-05 | 2013-12-05 | Method and device for constructing RBF neural network based on unmarked samples |
Applications Claiming Priority (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN201310654934.2A CN103679267A (en) | 2013-12-05 | 2013-12-05 | Method and device for constructing RBF neural network based on unmarked samples |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| CN103679267A true CN103679267A (en) | 2014-03-26 |
Family
ID=50316751
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| CN201310654934.2A Pending CN103679267A (en) | 2013-12-05 | 2013-12-05 | Method and device for constructing RBF neural network based on unmarked samples |
Country Status (1)
| Country | Link |
|---|---|
| CN (1) | CN103679267A (en) |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104504441A (en) * | 2014-12-09 | 2015-04-08 | 河海大学 | Method and device for constructing MADALINE neural network based on sensitivity |
| CN107223260A (en) * | 2015-02-17 | 2017-09-29 | 高通股份有限公司 | Method for dynamicalling update grader complexity |
| CN110472736A (en) * | 2019-08-26 | 2019-11-19 | 联想(北京)有限公司 | A kind of method and electronic equipment cutting neural network model |
| CN112762957A (en) * | 2020-12-29 | 2021-05-07 | 西北工业大学 | Multi-sensor fusion-based environment modeling and path planning method |
-
2013
- 2013-12-05 CN CN201310654934.2A patent/CN103679267A/en active Pending
Cited By (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN104504441A (en) * | 2014-12-09 | 2015-04-08 | 河海大学 | Method and device for constructing MADALINE neural network based on sensitivity |
| CN107223260A (en) * | 2015-02-17 | 2017-09-29 | 高通股份有限公司 | Method for dynamicalling update grader complexity |
| CN110472736A (en) * | 2019-08-26 | 2019-11-19 | 联想(北京)有限公司 | A kind of method and electronic equipment cutting neural network model |
| CN110472736B (en) * | 2019-08-26 | 2022-04-22 | 联想(北京)有限公司 | Method for cutting neural network model and electronic equipment |
| CN112762957A (en) * | 2020-12-29 | 2021-05-07 | 西北工业大学 | Multi-sensor fusion-based environment modeling and path planning method |
| CN112762957B (en) * | 2020-12-29 | 2022-12-30 | 西北工业大学 | Multi-sensor fusion-based environment modeling and path planning method |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Duan et al. | Short-term wind power forecasting using the hybrid model of improved variational mode decomposition and Correntropy Long Short-term memory neural network | |
| CN110659693B (en) | Method, system and medium for rapid topology identification of distribution network based on K nearest neighbor classification | |
| CN102056183B (en) | Network flow prediction method and device based on cognitive network | |
| CN106022954B (en) | Multiple BP neural network load prediction method based on grey correlation degree | |
| CN108346293B (en) | A method for short-term forecasting of real-time traffic flow | |
| CN107633301A (en) | The training method of testing and its application system of a kind of BP neural network regression model | |
| CN105760952A (en) | Load prediction method based on Kalman filtering and self-adaptive fuzzy neural network | |
| CN103077267A (en) | Parameter sound source modeling method based on improved BP (Back Propagation) neural network | |
| CN104569747A (en) | System and method for checking insulativity of power-off cable | |
| CN106053988A (en) | Inverter fault diagnosis system and method based on intelligent analysis | |
| CN104299031A (en) | Ultra-short-term load prediction method of BP neural network | |
| CN103679267A (en) | Method and device for constructing RBF neural network based on unmarked samples | |
| CN107274028A (en) | A kind of many wind fields based on mixing Copula functions are exerted oneself Forecasting Methodology | |
| CN116843083A (en) | Carbon emission prediction system and method based on hybrid neural network model | |
| CN107528649A (en) | A kind of Forecasting Methodology of the cognitive radio networks frequency spectrum based on Bagging | |
| CN104504441A (en) | Method and device for constructing MADALINE neural network based on sensitivity | |
| Duong-Ngoc et al. | Short term load forcast using deep learning | |
| CN113485986B (en) | Electric power data restoration method | |
| CN103020485A (en) | Short-term wind speed forecast method based on beta noise kernel ridge regression technology | |
| CN117076851A (en) | Wind speed prediction method based on ICEEMD-ICA-BiLSTM | |
| CN104537391A (en) | Meta learning method of extreme learning machine | |
| CN115908054A (en) | Building energy consumption prediction method, device, equipment and medium | |
| CN106503793A (en) | A kind of neural network short-term wind speed forecasting method based on improvement difference algorithm | |
| CN104573864A (en) | A data analysis and alarm method based on autoregressive prediction | |
| Wang et al. | A wind speed prediction method based on improved empirical mode decomposition and support vector machine |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| PB01 | Publication | ||
| PB01 | Publication | ||
| C10 | Entry into substantive examination | ||
| SE01 | Entry into force of request for substantive examination | ||
| C02 | Deemed withdrawal of patent application after publication (patent law 2001) | ||
| WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20140326 |