[go: up one dir, main page]

CN118871929A - A data processing system, method and computer program product including a network - Google Patents

A data processing system, method and computer program product including a network Download PDF

Info

Publication number
CN118871929A
CN118871929A CN202380027102.XA CN202380027102A CN118871929A CN 118871929 A CN118871929 A CN 118871929A CN 202380027102 A CN202380027102 A CN 202380027102A CN 118871929 A CN118871929 A CN 118871929A
Authority
CN
China
Prior art keywords
nodes
node
weights
inputs
data processing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202380027102.XA
Other languages
Chinese (zh)
Inventor
H·乔恩泰欧
L·玛特森
U·隆加拉
J·埃南德
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intuit Inc
Original Assignee
Intuit Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intuit Inc filed Critical Intuit Inc
Priority claimed from PCT/SE2023/050153 external-priority patent/WO2023163637A1/en
Publication of CN118871929A publication Critical patent/CN118871929A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/18Complex mathematical operations for evaluating statistical data, e.g. average values, frequency distributions, probability functions, regression analysis
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B13/00Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
    • G05B13/02Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
    • G05B13/0265Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
    • G05B13/027Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biophysics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Pure & Applied Mathematics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Mathematical Optimization (AREA)
  • Mathematical Analysis (AREA)
  • Computational Mathematics (AREA)
  • Algebra (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Operations Research (AREA)
  • Databases & Information Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Devices For Executing Special Programs (AREA)
  • Multi Processors (AREA)

Abstract

本公开涉及一种方法、计算机程序产品和数据处理系统(100),数据处理系统(100)具有包括待处理数据的一个或多个系统输入(110a、110b、…、110z)和系统输出(120),数据处理系统(100)包括:网络NW(130);网络NW(130)包括多个节点(130a、130b、…、130x),每个节点被配置为具有多个输入(132a、132b、…、132y),每个节点(130a、130b、…、130x)包括针对每个输入(132a、132b、…、132y)的权重(Wa、…、Wy),并且每个节点被配置为产生输出(134a、134b、…、134x);以及一个或多个更新单元(150),所述一个或多个更新单元(150)在学习模式期间基于所述节点(130a)的每个相应输入(132a、…、132c)与所述对应输出(134a)的相关性来更新每个节点的所述权重(Wa、…、Wy);一个或多个处理单元(140x),其接收处理单元输入并产生处理单元输出,每个相应的处理单元(140x)将所述处理单元输出作为输入(132b、132e、…)提供给所述一个或多个其他节点(130a、130b、…)。

The present disclosure relates to a method, a computer program product and a data processing system (100), the data processing system (100) having one or more system inputs (110a, 110b, ..., 110z) and a system output (120) including data to be processed, the data processing system (100) comprising: a network NW (130); the network NW (130) comprising a plurality of nodes (130a, 130b, ..., 130x), each node being configured to have a plurality of inputs (132a, 132b, ..., 132y), each node (130a, 130b, ..., 130x) comprising a weight (Wa, ..., Wy) for each input (132a, 132b, ..., 132y), and each nodes are configured to generate outputs (134a, 134b, ..., 134x); and one or more update units (150), which update the weights (Wa, ..., Wy) of each node based on the correlation of each corresponding input (132a, ..., 132c) of the node (130a) with the corresponding output (134a) during the learning mode; one or more processing units (140x) that receive processing unit inputs and generate processing unit outputs, each corresponding processing unit (140x) providing the processing unit output as input (132b, 132e, ...) to the one or more other nodes (130a, 130b, ...).

Description

Data processing system, method and computer program product comprising a network
Technical Field
The present disclosure relates to a data processing system, method and computer program product comprising a network. More particularly, the present disclosure relates to a data processing system, method and computer program product comprising a network as defined in the preambles of the independent claims.
Background
Artificial Intelligence (AI) is known. One example of AI is artificial neural networks (ans). The ANNs may be affected by the rigid representation, which appears to result in the network focusing on limited features for identification. Such a rigid representation may lead to inaccurate predictions. Thus, it may be advantageous to create a network/data processing system that does not rely on a rigid representation, e.g., one in which reasoning is based on a broad representation across all nodes/elements and/or one in which individual features are not allowed to become too dominant, thereby providing a more accurate prediction and/or a more accurate data processing system. The network where all nodes contribute to all representations is called a dense code network. Up to now, the implementation of densely coded networks has been hampered by the lack of rules for autonomous network formation, making it difficult to generate functional networks with high capacity/variation.
Accordingly, an AI system having an increased capacity may be desired. And (5) processing. Preferably, such AI systems provide or implement one or more of the following: better performance, higher reliability, higher efficiency, faster training speed, less computer power used, less training data used, less memory used, less complexity, and/or less energy used.
SE 2051375 A1 alleviates some of the above problems. However, there may still be a need for a more efficient AI/data processing system and/or alternative methods.
Disclosure of Invention
It is an object of the present disclosure to mitigate, alleviate or eliminate one or more of the above-identified deficiencies in the art and disadvantages singly or in any combination and at least solve the above mentioned problems.
According to a first aspect, a data processing system is provided. The data processing system is configured to have one or more system inputs and system outputs that include data to be processed. The data processing system includes: a network NW comprising a plurality of nodes, each node configured to have a plurality of inputs, each node comprising a weight for each input, and each node configured to produce an output; and one or more updating units configured to update the weight of each node during the learning mode based on the correlation of each respective input and corresponding output of the node; one or more processing units configured to receive the processing unit input and configured to generate a processing unit output by changing a sign of the received processing unit input. The system output includes the output of each node. Further, a node of the first set of nodes of the plurality of nodes is configured to stimulate one or more other nodes of the plurality of nodes by providing an output of each node of the first set of nodes as an input to the one or more other nodes of the plurality of nodes. Further, the nodes in the second group of the plurality of nodes are configured to suppress one or more other nodes in the plurality of nodes by providing the output of each node in the second group as a processing unit input to a respective processing unit, each respective processing unit being configured to provide the processing unit output as an input to one or more other nodes. Each of the plurality of nodes belongs to one of the first and second groups of nodes.
According to some embodiments, the system inputs sensor data comprising a plurality of contexts/tasks.
According to some embodiments, the update unit comprises a probability value for each weight for increasing the weight, and during the learning mode, the data processing system is configured to limit the ability of the node to suppress or excite one or more other nodes by: providing a first set point of a sum of ownership weights associated with inputs to one or more other nodes, comparing the first set point to the sum of ownership weights associated with inputs to one or more other nodes, decreasing the probability value associated with the weight associated with the input to the one or more other nodes if the first set point is less than the sum of ownership weights associated with the inputs to the one or more other nodes, and increasing the probability value associated with the weight associated with the input to the one or more other nodes if the first set point is greater than the sum of ownership weights associated with the input to the one or more other nodes. Thereby, the uniqueness of each node is improved, learning is improved/accelerated, and/or precision/accuracy is improved/increased.
According to some embodiments, during the learn mode, the data processing system is configured to limit the ability of the system input to suppress or stimulate one or more nodes by: providing a first set point of a sum of ownership weights associated with inputs to one or more nodes, comparing the first set point to a sum of ownership weights associated with inputs to one or more nodes, decreasing the probability value associated with the weights associated with the inputs to the one or more nodes if the first set point is less than the sum of ownership weights associated with the inputs to the one or more nodes, and increasing the probability value associated with the weights associated with the inputs to the one or more nodes if the first set point is greater than the sum of ownership weights associated with the inputs to the one or more nodes. Thereby, learning is improved/accelerated, and/or precision/accuracy is improved/increased.
According to some embodiments, each input to one or more other nodes has coordinates in the network space, and the amount of decrease/increase in the weight of the input to the one or more other nodes is based on the distance between the coordinates of the inputs associated with the weight in the network space.
According to some embodiments, the system is further configured to set the weight to zero if the weight does not increase within a preset period of time.
According to some embodiments, the system is further configured to increase the probability value of the weight having a zero value if the sum of all weights associated with the inputs to the one or more other nodes does not exceed the first set point within a preset period of time.
According to some embodiments, during the learning mode, the data processing system is configured to increase the relevance of the output of a node to one or more other nodes by: providing a first set point of a sum of ownership weights associated with inputs to one or more other nodes, comparing the first set point to the sum of ownership weights associated with inputs to one or more other nodes over a first period of time, increasing the probability of changing the weight of the input to the node if the first set point is less than the sum of ownership weights associated with inputs to the one or more other nodes over the entire length of the first period of time, and decreasing the probability of changing the weight of the input to the node if the first set point is greater than the sum of ownership weights associated with inputs to the one or more other nodes over the entire length of the first period of time. Thereby, learning is improved/accelerated, and/or precision/accuracy is improved/increased.
According to some embodiments, the updating unit comprises a probability value for each weight for increasing the weight, and wherein, during the learning mode, the data processing system is configured to provide a second set point of the sum of all weights associated with the input to the node, the data processing system is configured to calculate the sum of all weights associated with the input to the node, to compare the calculated sum with the second set point, and if the calculated sum is greater than the second set point, to decrease the probability value associated with the weight associated with the input to the node, and if the calculated sum is less than the second set point, to increase the probability value associated with the weight associated with the input to the node. Thereby, learning is improved/accelerated, and/or precision/accuracy is improved/increased.
According to some embodiments, each node comprises a plurality of compartments, and each compartment is configured with a plurality of compartment inputs, each compartment comprises a compartment weight for each compartment input, and each compartment is configured to generate a compartment output, and each compartment comprises an updating unit configured to update the compartment weights based on the correlations during the learning mode, and the compartment output of each compartment is used to adjust the output of the node comprising the compartment based on the transfer function. Thus, each individual node is made more useful/powerful (e.g., increased capacity), learning is improved/accelerated, and/or precision/accuracy is improved/increased.
According to some embodiments, during the learning mode, the data processing system is configured to: detecting whether the network is sparsely connected by comparing the cumulative weight change of the system input over the second time period with a threshold; and if the data processing system detects that the network is sparsely connected, increasing the output of one or more of the plurality of nodes by adding a predetermined waveform to the output of one or more of the plurality of nodes for a duration of the third period of time. Thus, a more efficient data processing system is achieved that can handle a wider range of contexts/tasks given network resources, thereby reducing power consumption.
According to some embodiments, each node comprises an updating unit, each updating unit being configured to update the weight of each respective input of a node based on the correlation of the output of that node, and each updating unit being configured to apply a first function to the correlation if the associated node belongs to a first group of a plurality of nodes, and to apply a second function different from the first function to the correlation if the associated node belongs to the second group of the plurality of nodes, so as to update the weight during the learning mode. By updating the weights of the respective nodes based on the correlation of each respective input of the node with the output of the (same) node, and if the associated node belongs to a first group of the plurality of nodes, a first function is applied to the correlation, and if the associated node belongs to a second group of the plurality of nodes, a second function different from the first function is applied to the correlation, so that the weights are updated (during learning mode), so that each node is more independent of the other nodes, and a higher accuracy is obtained (compared to prior art e.g. back propagation). Thus, a technical effect is that higher precision/accuracy is achieved/obtained.
According to some embodiments, the data processing system is configured to calculate a total variance of the outputs of the nodes of the network after the updating of the weights has been performed, compare the calculated total variance with a power law; and minimizing the error or mean square error between the ensemble and the power law by adjusting parameters of the network. Thus, each node is made more independent of the other nodes (and a measure of how independent the nodes are from each other can be obtained). Thus, a more efficient data processing system is achieved that can handle a wider range of contexts/tasks for a given network resource, thereby reducing power consumption.
According to some embodiments, the data processing system is configured to learn from the sensor data to identify one or more entities when in a learning mode, and thereafter the data processing system is configured to identify the one or more entities when in an execution mode, and the identified entities are one or more of: a speaker, spoken letter, syllable, phoneme, word or phrase present in the sensor data, or an object or feature of an object present in the sensor data, or a new contact event present in the sensor data, an end of a contact event, a gesture, or an applied pressure. In some embodiments, higher precision/accuracy in identifying one or more entities or measurable features thereof is achieved/obtained.
According to some embodiments, the network is a recurrent neural network.
According to some embodiments, the network is a recurrent neural network.
According to a second aspect, a computer-implemented or hardware-implemented method for processing data is provided. The method includes a) receiving one or more system inputs including data to be processed; b) Providing a plurality of inputs to a network NW comprising a plurality of first nodes, at least one of the plurality of inputs being a system input; c) Receiving an output from each first node; d) Providing a system output comprising an output of each first node; e) Energizing, by a node in a first group of the plurality of nodes, one or more other nodes in the plurality of nodes by providing an output of each node in the first group of nodes as an input to the one or more other nodes; f) Suppressing, by a node in a second group of the plurality of nodes, one or more other nodes in the plurality of nodes by providing an output of each node of the second group as a processing unit input to a respective processing unit, each respective processing unit configured to provide the processing unit output as an input to the one or more other nodes; and g) updating the weights, optionally by one or more updating units, based on the correlations, h) optionally repeating a) -g) until a learning criterion is met, and i) repeating a) -f) until a stopping criterion is met, and each of the plurality of nodes belongs to one of the first and second sets of points of nodes.
According to some embodiments, the method further comprises initializing the weights by setting the weights to zero and adding a predetermined waveform to the output of one or more of the plurality of nodes for a duration of a third time period, the third time period beginning at a time of receiving one or more system inputs comprising data to be processed.
According to some embodiments, the method further comprises: the weights are initialized by randomly assigning values between 0 and 1 to the weights, and a predetermined waveform is added to the output of one or more of the plurality of nodes for the duration of the third time period.
According to a third aspect, there is provided a computer program product comprising a non-transitory computer readable medium having stored thereon a computer program comprising program instructions, the computer program being loadable into a data-processing unit and configured to perform the method of the third aspect or any of the embodiments described above when the computer program is run by the data-processing unit.
The effects and features of the second and third aspects are largely analogous to those described in connection with the first aspect above and vice versa. The embodiments mentioned in relation to the first aspect are to a large extent compatible with the second and third aspects and vice versa.
An advantage of some embodiments is that data/information is more efficiently processed, for example during a learning/training mode.
Another advantage of some embodiments is that an efficient network is provided, e.g., maximizing the utilization of available network capacity, thereby providing a more efficient data processing system.
Another advantage of some embodiments is that the system/network complexity is lower, e.g., with a reduced number of nodes (with the same accuracy and/or for the same context/input scope).
Yet another advantage of some embodiments is that data usage is more efficient.
Another advantage of some embodiments is that the utilization of available network capacity is improved (e.g., maximized), thereby providing an efficient data processing system.
Yet another advantage of some embodiments is that the system/network is more efficient and/or training/learning time is shorter/faster.
Another advantage of some embodiments is that a network of lower complexity is provided.
Another advantage of some embodiments is improved/increased versatility (e.g., across different tasks/contexts).
Yet another advantage of some embodiments is that the sensitivity of the system/network to noise is reduced.
Other advantages of some embodiments are improved performance, improved reliability/enhancement, improved accuracy, improved efficiency (for training and/or performance), faster training/learning speed/shorter time, lower power consumption of the required computer, less training data required, less storage space required, lower complexity and/or lower energy consumption.
In some embodiments, each node is made more independent of the other nodes. This results in an increase in the total capability of representing information in the data processing system (and thus may for example represent more information in the data processing system or be used to identify one or more entities/objects and/or one or more features of one or more objects) and thus achieve a higher accuracy (compared to prior art such as back propagation).
The present disclosure will become apparent from the detailed description given below. The detailed description and specific examples disclose the preferred embodiments of the present disclosure by way of illustration only. Those skilled in the art will appreciate from the teachings of the detailed description that changes and modifications can be made within the scope of the disclosure.
Accordingly, it is to be understood that the disclosure disclosed herein is not limited to the particular components of the apparatus described or to the steps of the methods described, as such devices and methods may vary. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to be limiting. It should be noted that as used in the specification and the appended claims, "a," "an," "the," and "the" are intended to mean that there are one or more elements unless the context clearly dictates otherwise. Thus, for example, reference to "a unit" or "the unit" may include several devices, etc. Moreover, the words "comprise", "include", "contain" and similar words do not exclude other elements or steps.
Drawings
The above objects, as well as additional objects, features and advantages of the present disclosure will be more fully understood by reference to the following illustrative and non-limiting detailed description of exemplary embodiments of the present disclosure when taken in conjunction with the accompanying drawings.
FIG. 1 is a schematic block diagram of a data processing system shown in accordance with some embodiments;
FIG. 2 is a schematic block diagram of a data processing system shown in accordance with some embodiments;
FIG. 3 is a flowchart illustrating exemplary method steps according to some embodiments;
FIG. 4 is a schematic diagram of an exemplary computer-readable medium shown according to some embodiments;
FIG. 5 is a schematic diagram of an update unit shown according to some embodiments; and
Fig. 6 is a schematic diagram of one compartment shown according to some embodiments.
Detailed Description
The present disclosure will now be described with reference to the accompanying drawings, in which preferred example embodiments of the disclosure are shown. However, the present disclosure may be embodied in other forms and should not be construed as limited to the embodiments set forth herein. The disclosed embodiments are provided to fully convey the scope of the disclosure to the skilled artisan.
Terminology
Hereinafter referred to as "nodes". The term "node" may refer to a neuron, such as a neuron of an artificial neural network, another processing element of a network of processing elements, such as a processor, or a combination thereof. Thus, the term "network" (NW) may refer to an artificial neural network, a network of processing elements, or a combination thereof.
Hereinafter referred to as "processing unit". The processing unit may also be referred to as a synapse, such as an input unit (with processing unit) for a node. However, in some embodiments, the processing unit is a (general purpose) processing unit (other than a synapse) associated with (connected to, connectable to, or included in) a node of the NW, or a (general purpose) processing unit located between two different nodes of the NW.
Hereinafter referred to as "context". Context is the environment or situation involved. The context is related to the type of (input) data that is intended, e.g. different types of tasks, where each different task has its own context. As an example, if the system input is a pixel from an image sensor, and the image sensor is exposed to different lighting conditions, each different lighting condition may be a different context of an object imaged by the image sensor, such as a ball, car, or tree. As another example, if the system input is an audio frequency band from one or more microphones, each different speaker may be a different context of phonemes present in the one or more audio frequency bands.
Hereinafter referred to as "measurable". The term "measurable" should be interpreted as something that can be measured or detected, i.e. something that is detectable. The terms "measurement" and "sensing" should be interpreted as synonyms.
Hereinafter referred to as "entities". The term entity is to be interpreted as an entity, such as a physical entity or a more abstract entity, such as a financial entity, e.g. one or more financial data sets. The term "physical entity" will be interpreted as an entity having physical presence, such as an object, a feature (of an object), a gesture, an applied pressure, a speaker, a spoken letter, a syllable, a phoneme, a word or a phrase.
Hereinafter referred to as "update unit". The update unit may be an update module or an update object.
Embodiments will be described below, wherein FIG. 1 is a schematic block diagram of a data processing system 100 shown in accordance with some embodiments, and FIG. 2 is a schematic block diagram of a data processing system 100 shown in accordance with some embodiments. In some embodiments, data processing system 100 is a Network (NW) or data processing system 100 includes an NW. In some embodiments, data processing system 100 is or includes a deep neural network, a deep belief network, a deep reinforcement learning system, a recurrent neural network, or a convolutional neural network.
The data processing system 100 has or is configured with one or more system inputs 110a, 110b, …, 110z. One or more of the system inputs 110a, 110b, …, 110z include data to be processed. The data may be multidimensional. For example, a plurality of signals are provided in parallel. In some embodiments, the system inputs 110a, 110b, …, 110z include or consist of time-continuous data. In some embodiments, the data to be processed includes data from sensors, such as image sensors, touch sensors, and/or sound sensors (e.g., microphones). Further, in some embodiments, one or more of the system inputs 110a, 110b, …, 110z includes sensor data for a plurality of contexts/tasks, e.g., when the data processing system 100 is in a learn mode and/or when the data processing system 100 is in an execute mode. That is, in some embodiments, data processing system 100 is in both an execution mode and a learning mode.
Further, data processing system 100 has or is configured with a system output 120. Data processing system 100 includes Network (NW) 130.NW 130 includes a plurality of nodes 130a, 130b, …,130 x. Each node 130a, 130b, …,130 x has or is configured with a plurality of inputs 132a, 132b, …, 132y. In some embodiments, at least one of the plurality of inputs 132a, 132b, …, 132y is a system input 110a, 110b, …, 110z. Further, in some embodiments, all of the system inputs 110a, 110b, …, 110z are used as inputs 132a, 132b, …, 132y to one or more of the nodes 130a, 130b, …,130 x. Further, in some embodiments, each of the nodes 130a, 130b, …,130 x has one or more system inputs 110a, 110b, …, 110z as inputs 132a, 132b, …, 132y. Each node 130a, 130b, …,130 x has or includes a weight Wa, wb, …, wy for each input 132a, 132b, …, 132y, i.e., each input 132a, 132b, …, 132y is associated with a respective weight Wa, wb, …, wy. In some embodiments, each weight Wa, wb, …, wy has a value in the range from 0 to 1. Further, NW 130, or each node thereof, generates or is configured to generate an output 134a, 134b, …, 134x. In some embodiments, each node 130a, 130b, …,130 x calculates (to that node) a combination of inputs 132a, 132b, …, 132y multiplied by respective weights Wa, wb, …, wy, such as a (linear) sum, square sum, or average, to produce outputs 134a, 134b, …, 134x.
The data processing system 100 comprises one or more updating units 150 configured to update the weights Wa, …, wy of each node during the learning mode based on (in accordance with) the correlation of each respective input 132a, …,132 c of the node (e.g., 130 a) with the corresponding output (e.g., 134 a), i.e., with the output of the same node (e.g., 130 a). In some embodiments, the weights are not updated during the execution mode. In one example, the updating of the weights Wa, wb, wc is based on (according to) the correlation of each respective input 132a, …,132 c to a node 130a with the combined activation of all inputs 132a, …,132 c to that node 130a, i.e., the correlation of each respective input 132a, …,132 c to a node 130a with the output 134a of that node 130a (node 130a is exemplary and applicable to all other nodes 130b, …,130 x). Thus, the correlation (value) between the first input 132a and the respective output 134a is calculated, the correlation (value) between the second input 132b and the respective output 134a is calculated, and the correlation (value) between the third input 132c and the respective output 134a is calculated. In some embodiments, different calculated correlation (series) values are compared to each other and the weights are updated based on (according to) the comparison. In some embodiments, updating the weights Wa, …, wy of each node based on (according to) the correlation of each respective input (e.g., 132a, …,132 c) of the node (e.g., 130 a) with the corresponding output (e.g., 134 a) includes evaluating each input (e.g., 132a, …,132 c) of the node (e.g., 130 a) based on (according to) a scoring function. The scoring function gives an indication of how useful each input (e.g., 132a, …,132 c) of a node (e.g., 130 a) is in space, e.g., for a corresponding output (e.g., 134 a) compared to other inputs (e.g., 132a, …,132 c) to the node, and/or in time, e.g., as the data processing system (100) processes the input (e.g., 132 a). As described above, the updating of the weights Wa, …, wy of each node is based on or in accordance with the correlation of each respective input 132a, …,132 c of the node (e.g., 130 a) with the corresponding output (e.g., 134 a), i.e., with the output of (only) the same node. Thus, the updating of the weight of each node is independent of the updating/learning in the other nodes, i.e. each node has independent learning.
Further, the data processing system 100 includes one or more processing units 140x configured to receive the processing unit input 142x and configured to generate a processing unit output 144x by changing the sign of the received processing unit input 142 x. In some implementations, the sign of the received processing unit input 142x is changed by multiplying the processing unit input 142x by-1. However, in other embodiments, the sign of the received processing unit input 142x is changed by shifting the received processing unit input 142x by 180 degrees. Or by reversing the sign of the received processing unit input 142x, for example from positive to negative or from negative to positive. The system output 120 includes an output 134a, 134b, …, 134x for each node 130a, 130b, …, 130 x. In some embodiments, the system output 120 is an array of outputs 134a, 134b, …, 134x. Further, in some embodiments, the system output 120 is used to identify one or more entities or measurable characteristics thereof (or characteristics) while in the execution mode, e.g., from sensor data.
In some embodiments, NW 130 includes only a first set 160 of the plurality of nodes 130a, 130b, …,130x (as shown in fig. 1). However, in some embodiments, NW 130 includes a first group 160 of the plurality of nodes 130a, 130b, …,130x and a second group 162 of the plurality of nodes 130a, 130b, …,130x (as shown in fig. 2). Each node (e.g., 130a, 130 b) in the first set 160 of the plurality of nodes (i.e., excitatory nodes) is configured to stimulate one or more other nodes (e.g., 130 x) in the plurality of nodes 130a, 130b, …,130x by providing an output (e.g., 134a, 134 b) of each node (e.g., 130a, 130 b) in the first set 160 of nodes (directly) as an input (132 d, …, 132 y) to one or more other nodes (e.g., 130 x) in the plurality of nodes 130a, 130b, …,130x, such as all other nodes 130b, …,130 x.
Further, the nodes (e.g., 130 x) in the second set 162 of the plurality of nodes are configured to suppress one or more other nodes 130a, 130b, …, such as one or more other nodes 130a, 130b, … being all other nodes 130a, 130b, … in the plurality of nodes 130a, 130b, …,130x, by providing an output (e.g., 134 x) of each node (e.g., 130 x) in the second set 162 as a processing unit input 142x to a respective processing unit (e.g., 140 x), each respective processing unit (e.g., 140 x) being configured to provide a processing unit output 144x as an input (e.g., 132b, 132 e) to one or more other nodes, e.g., 130a, 130 b). Each node of the plurality of nodes 130a, 130b, …,130x belongs to one of the first and second groups (160, 162) of nodes. Further, as described above, in some embodiments, all nodes 130a, 130b, …,130x belong to a first group 160 of nodes. In some embodiments, each node 130a, 130b, …,130x is configured to suppress or energize some/all of the other nodes 130b, …,130x in the plurality of nodes 130a, 130b, …,130x by multiplying the output 134a, 134b, …,134x (of each node 130a, 130b, …,130 x) by-1 or providing it directly as an input 132d, …,132 y to one or more other nodes 130b, …,130 x. By configuring one of the nodes to suppress other nodes and another of the nodes to excite other nodes and perform updates based on (based on) the correlation during the learning mode, a more efficient network may be provided, e.g., the utilization of available network capacity may be maximized, thereby providing a more efficient data processing system.
In some embodiments, the updating unit 150 comprises a probability value Pa, …, py for each weight Wa, …, wy for increasing the weight (and possibly a probability value Pad, …, pyd for decreasing the weight, in some embodiments the probability value is 1-Pa, …, 1-Py, i.e. Pad = 1-Pa, pbd = 1-Pb, etc.). In some embodiments, the updating unit 150 comprises a look-up table (LUT) for storing probability values Pa, …, py. During the learn mode, data processing system 100 is configured to limit the ability of a node (e.g., 130 a) to inhibit or stimulate one or more other nodes (e.g., 130b, …,130 x) by: providing a first set point of a sum of ownership weights (e.g., wd, wy) associated with inputs (e.g., 132d, …,132 y) to one or more other nodes (e.g., 130b, …,130 x), comparing the first set point to a sum of ownership weights (e.g., wd, wy) associated with inputs (e.g., 132d, …,132 y) to one or more other nodes (e.g., 130b, …,130 x), decreasing a probability value (e.g., wd, wy) associated with the input (e.g., 132d,132 y) to one or more other nodes (e.g., 130b, …,130 x), increasing a probability value (e.g., pd, 132 y) associated with the input (e.g., 132d,132 y) to one or more other nodes (e.g., 132d,132 y) if the first set point is less than a sum of ownership weights (e.g., wd, wy) associated with inputs (e.g., 132d,132 y) to one or more other nodes (e.g., 130b, …,130 x), increasing a probability value (e.g., pd, 132 y) associated with the input (e.g., 132d,132 y) to one or more nodes (e.g., 132d,132 y).
Further, in some embodiments, data processing system 100 is configured during a learn mode to limit the ability of a system input (e.g., 110 z) to inhibit or stimulate one or more nodes (e.g., 130b, 130 x) by: providing a first set point of a sum of ownership weights (e.g., wg, wx) associated with inputs (e.g., 132g, 132 x) to one or more nodes (e.g., 130b, 130 x) to a first set point, the first set point being compared to the sum of ownership weights (e.g., wg, 132 x) associated with inputs (e.g., 132g, 132 x) to one or more nodes (e.g., 130b, 130 x), decreasing a probability value (e.g., pg, px) associated with the weights (e.g., wg, wx) if the first set point is less than the sum of ownership weights (e.g., wg, 132 x) associated with inputs (e.g., 132g, 132 x) to one or more nodes (e.g., 130b, 130 x), and increasing a probability value (e.g., pg, px) associated with the first set point (e.g., wg, wx) if the first set point is greater than the sum of ownership weights (e.g., 132g, 132 x) associated with the first set point (e.g., 132 g., 132g, 132 x) to the one or more nodes (e.g., 132 g., 132g, 132 x).
Further, in some embodiments, each input (e.g., 132d, 132 y) to one or more other nodes (e.g., 130b, 130 x) has coordinates in the network space, and the amount of decrease/increase in the weight (e.g., wd, wy) of the input (e.g., 132d, 132 y) to the one or more other nodes (e.g., 130b, 130 x) is based on (in terms of) the distance between the coordinates in the network space of the input (e.g., 132d, 132 y) associated with the weight (e.g., wd, wy). In these embodiments, the decrease/increase in weight is based on (according to) the probability of decreasing/increasing the weight (indicated by the probability value) and is based on (according to) the amount of decreasing/increasing the weight (which is calculated based on the distance between the coordinates entered in the network space).
In some embodiments, data processing system 100 is (further) configured to: if the weights Wa, …, wy (discussed) do not increase for the (first) preset period of time, the weights Wa, …, wy (e.g., any one of one or more of the weights) are set to zero. Moreover, in some embodiments, data processing system 100 is (further) configured to: if the sum of all weights (e.g., wd, wy) associated with inputs (e.g., 132d, 132 y) to one or more other nodes (e.g., 130b, 130 x) does not exceed the first set point within a (second) preset time period, then the probability value Pa, …, py of the weight Wa, …, wy having a zero value is increased.
In some embodiments, data processing system 100 is configured during a learn mode to increase the relevance of the output (e.g., 134 a) of a node (e.g., 130 a) to one or more other nodes (e.g., 130b, 130 x) by: providing a first set point of a sum of ownership weights (e.g., wd, wy) associated with inputs (e.g., 132d, 132 y) to a plurality of other nodes (e.g., 130b, 130 x), comparing the first set point to the sum of ownership weights (e.g., 132d, 132 y) associated with inputs (e.g., 132d, 132 y) to one or more other nodes (e.g., 130b, 130 x), increasing a probability of changing the input (e.g., 132a, 132b, 132 c) to the sum of ownership weights (e.g., wd, wy) to the node (e.g., 132 a), and if the first set point is less than the sum of ownership weights (e.g., wd, wy) associated with inputs (e.g., 132b, 130 x) to one or more other nodes (e.g., 130b, 130 x) over the entire length of the first time period, decreasing the probability of changing the sum of ownership weights (e.g., 132d, wy) to the node (e.g., 132a, 132b, 132 c) over the entire length of the first time period, and if the first set point is less than the sum of ownership weights (e.g., 132d, 132 y) associated with inputs (e.g., 132d, 132 y) to the one or more other nodes (e.g., 130b, 132 y) and if the probability of changing the sum of ownership weights (e.g., 132b, 132 c) is not changing over the entire length of the total length of the first set point (e.g., 132b, 132 c) to be less than the sum of ownership).
Furthermore, in some embodiments, the updating unit 150 comprises a probability value Pa, …, py for each weight Wa, …, wy for increasing the weight (and possibly a probability value Pad, …, pyd for decreasing the weight, in some embodiments the probability value is 1-Pa, …, 1-Py, i.e. Pad = 1-Pa, pbd = 1-Pb, etc.). In these embodiments, during the learn mode, the data processing system 100 is configured to provide a second set point of the sum of the ownership weights Wa, wb, wc associated with the inputs 132a, 132b, 132c to the node 130a, the data processing system 100 is configured to calculate the sum of the ownership weights Wa, wb, wc associated with the inputs 132a, 132b, 132c to the node 130a, the data processing system 100 is configured to compare the calculated sum with the second set point, and if the calculated sum is greater than the second set point, to decrease the probability values Pa, pb, pc associated with the ownership weights Wa, wb, wc associated with the inputs 132a, 132b, 132c to the node 130a, and if the calculated sum is less than the second set point, to increase the probability values Pa, pb, pc associated with the weights Wa, wc associated with the inputs 132a, 132b, 132c (to the other nodes 130b, …, 130 x) as an example.
Further, in some embodiments, during the learning mode, the data processing system 100 is configured to detect whether the network 130 is sparsely connected by comparing the cumulative weight change of the one or more system inputs 110a, 110b, …, 110z over the second period of time to a threshold. The cumulative weight change refers to a change in weight Wa, wf, wg, wx associated with one or more system inputs 110a, 110b, …, 110z over a second period of time. The second time period may be a predetermined time period. If the cumulative weight change is greater than the threshold, then the network 130 is determined to be sparsely connected. Further, data processing system 100 is configured to: if the data processing system 100 detects a sparse connection of the network 130, the output 134a, 134b, …, 134x of one or more of the plurality of nodes 130a, 130b, …,130 x is increased by adding a predetermined waveform to the output 134a, 134b, …, 134x of one or more of the plurality of nodes 130a, 130b, …,130 x for the duration of the third period of time. The third time period may be a predetermined time period. The nodes may be better grouped together by adding a predetermined waveform to the outputs 134a, 134b, …, 134x of one or more of the plurality of nodes 130a, 130b, …,130 x for the duration of the third time period.
Furthermore, in some embodiments, each node includes an update unit 150. Each updating unit 150 is configured to update the weights Wa, wb, wc of a respective node 130a based on (according to) the correlation of each respective input 132a, …, 132c of that node 130a with the output 134a of that node 130 a. Furthermore, each updating unit 150 is configured to apply a first function to the correlation if the associated node belongs to a first group 160 of the plurality of nodes and to apply a second function different from the first function to the correlation if the associated node belongs to a second group 162 of the plurality of nodes, in order to update the weights Wa, wb, wc during the learning mode (node 130a is an example and is also applicable to all other nodes 130b, …, 130 x). In some embodiments, the first (learning) function is a function in which the output, i.e. the weight change (value), increases exponentially if the input, i.e. the correlation (value), increases and vice versa (decreasing the input results in exponentially decreasing the output). In some embodiments, the second (learning) function is a function in which the output, i.e. the weight change (value), decreases exponentially if the input, i.e. the correlation (value), increases, and vice versa (decreasing the input results in exponentially increasing the output).
In some embodiments, the data processing system 100 is configured to calculate the overall variance of the outputs 134a, 134b, …, 134x of the nodes 130a, 130b, …, 130x of the network after the updating of the weights Wa, …, wy has been performed, compare the calculated overall variance to a power law; and minimizing errors between the population and the power law, such as mean absolute error or mean square error, by adjusting parameters of the network. Thus, the overall variance of the outputs 134a, 134b, …, 134x of the nodes 130a, 130b, …, 130x of the network may approximate a power law distribution. Hereby, an optimal resource utilization is achieved and/or each node is allowed to contribute in an optimal way, thereby providing a more efficient data utilization. The power law may be, for example, a logarithm of the amount of change based on (in terms of) a logarithmic interpretation of the number of components resulting from the principal component analysis. In another example, the power law is based on principal component analysis of finite time vectors of activity/output across all neurons, with each principal component score in the abscissa being replaced with a node number. It is assumed that the input data to which the system is exposed has a greater number of principal components than nodes. In this case, each node added to the system potentially expands the maximum capacity of the system when following the power law. Examples of (adjustable) parameters of the network include: the type of scaling learned (how the weights are composed, the range of weights, etc.), the induced change in synaptic weights at update (e.g., exponentially, linearly), the amount of gain in learning, one or more time constants of the state memory of the or each node, the specific learning function (e.g., first function and/or second function), the transfer function of each node, the total capacity of the connection between the node and the sensor, the total capacity of the node across all nodes.
Further, in some embodiments, the data processing system 100 is configured to learn from the sensor data to identify one or more (unidentified) entities or (unidentified) measurable characteristics (or properties) thereof while in the learning mode, and thereafter to identify the one or more entities or measurable characteristics (or properties) thereof while in the execution mode, e.g., from the sensor data. In some embodiments, the identified entities are one or more of speakers, spoken letters, syllables, phonemes, words or phrases present in the (audio) sensor data. Alternatively or additionally, the identified entities are one or more objects or one or more features (e.g., pixels) of objects present in the sensor data. As a further alternative or in addition, the identified entity is a new contact event present in the (touch) sensor data, the end of the contact event, a gesture or an applied pressure. Although in some embodiments all sensor data is a particular type of sensor data, such as audio sensor data, image sensor data, or touch sensor data, in other embodiments the sensor data is a mix of different types of sensor data, such as audio sensor data, image sensor data, and touch sensor data, i.e., the sensor data includes different modalities. In some embodiments, the data processing system 100 is configured to learn from the sensor data to identify the measurable characteristic (or characteristics) of the entity. The measurable characteristic may be a feature of the object, a portion of a feature, a time-evolving trajectory of the location, a trajectory of the applied pressure, or a frequency feature or time-evolving frequency feature of a certain speaker when speaking a certain letter, syllable, phoneme, word or phrase. Such measurable characteristics may then be mapped to an entity. For example, features of an object may be mapped to the object, a portion of the features may be mapped to features (of the object), a locus of locations may be mapped to gestures, a locus of applied pressure may be mapped to (maximum) applied pressure, a frequency feature of a certain speaker may be mapped to the speaker, and spoken letters, syllables, phonemes, words or phrases may be mapped to actual letters, syllables, phonemes, words or phrases. Such a mapping may simply be a lookup in a memory, a lookup table or a database. The searching may be based on (from) finding the entity of the plurality of physical entities having the closest characteristic to the identified measurable characteristic. From this lookup, the actual entity can be identified. Furthermore, the data processing system 100 may be used in a warehouse, e.g. as part of a fully automated warehouse (machine), in a robot, e.g. connected to a robot actuator (or robot control circuit) via middleware (for connecting the data processing system 100 to the actuator), or in a system with a low complexity event based camera, whereby trigger data from the event based camera may be fed/sent directly to the data processing system 100.
FIG. 3 is a flowchart illustrating exemplary method steps according to some embodiments. Fig. 3 illustrates a computer-implemented or hardware-implemented method 300 for processing data. The method may be implemented in analog hardware/electronic circuitry, in digital circuitry such as gates and flip-flops, in mixed signal circuitry, in software, and in any combination thereof. In some embodiments, the method 300 includes entering a learning mode. Alternatively, method 300 includes providing data processing system 100 that has been trained. In this case, steps 370 and 380 (steps g and h) are not performed. The method 300 includes receiving 310 one or more system inputs 110a, 110b, …, 110z including data to be processed. Further, the method 300 includes providing 320 a plurality of inputs 132a, 132b, …, 132y to the network NW 130 including the plurality of first nodes 130a, 130b, …, 130x, at least one of the plurality of inputs being a system input. Further, the method 300 includes receiving 330 an output 134a, 134b, …, 134x from each first node 130a, 130b, …, 130 x. The method 300 includes providing 340 a system output 120 including an output 34a, 134b, …, 134x of each first node 130a, 130b, …, 130 x. Further, the method 300 includes exciting 350, by a node 130a, 130b in the first set 160 of nodes, one or more other nodes …, 130x in the plurality of nodes 130a, 130b, …, 130x by providing an output 134a, 134b of each node 130a, 130b in the first set 160 of nodes as an input 132d, …, 132y to one or more other nodes …, 130x in the plurality of nodes 130a, 130b, …, 130x. Further, the method 300 includes suppressing 360, by a node 130x in the second set 162 of the plurality of nodes, one or more other nodes 130a, 130b, … in the plurality of nodes 130a, 130b, …, 130x by providing an output 134x of each node 130x in the second set 162 as a processing unit input 142x to a respective processing unit 140x, each respective processing unit 140x configured to provide a processing unit output 144x as an input 132b, 132e, … to one or more other nodes 130a, 130b, …. The method 300 includes updating 370 the weights Wa, …, wy by the one or more updating units 150 based on (according to) the correlations (during the learn mode and as described above in connection with fig. 1 and 2). Further, the method 300 includes repeating 380 (during the learning mode) steps 310, 320, 330, 340, 350, 360, and 370 (as described above) until the learning criterion is met (thus exiting the learning mode when the learning criterion is met). In some embodiments, the learning criteria is that data processing system 100 is fully trained. In some embodiments, the learning criteria is that the weights Wa, wb, …, wy converge and/or the total error is below an error threshold. In some embodiments, method 300 includes entering an execution/recognition mode. Further, the method 300 includes repeating 390 (during the execution/recognition mode) steps 310, 320, 330, 340, 350, and 360 (as described above) until the stopping criteria are met (thus exiting the execution/recognition mode when the stopping criteria are met). The stop criteria/conditions may be that all pending data has been processed or that a certain number of data/a certain number of cycles have been processed/executed. Alternatively, the stop criteria is that the entire data processing system 100 is shut down. As another alternative, the stopping criteria is that data processing system 100 (or a user of system 100) has found that further training of data processing system 100 is required. in this case, data processing system 100 enters/reenters the learn mode (and performs steps 310, 320, 330, 340, 350, 360, 370, 380, and 390). Each node of the plurality of nodes 130a, 130b, …, 130x belongs to one of the first set 160 and the second set 162 of nodes.
In some embodiments, the method 300 includes initializing 304 the weights Wa, …, wy by setting the weights Wa, …, wy to zero. Alternatively, the method 300 includes initializing 306 the weights Wa, …, wy by randomly assigning values between 0 and 1 to the weights Wa, …, wy. Further, in some embodiments, the method 300 includes adding 308 a predetermined waveform to the output 134a, 134b, …, 134x of one or more of the plurality of nodes 130a, 130b, …,130 x for the duration of the third period of time. In some embodiments, the third time period begins simultaneously with receiving 310 one or more system inputs 110a, 110b, …, 110z that include data to be processed.
According to some embodiments, a computer program product includes a non-transitory computer readable medium 400, such as Universal Serial Bus (USB) memory, a plug-in card, an embedded drive, a Digital Versatile Disk (DVD), or Read Only Memory (ROM). Fig. 4 illustrates an example computer-readable medium in the form of a Compact Disk (CD) ROM 400. A computer readable medium has stored thereon a computer program comprising program instructions. The computer program may be loaded into a data Processor (PROC) 420, which data Processor (PROC) 420 may for example be comprised in a computer or computing device 410. When loaded into a data processing unit, the computer program may be stored in a memory (MEM) 430 associated with or included in the data processing unit. According to some embodiments, the computer program may perform the method steps of the method shown in fig. 3, for example, when the computer program is loaded into and executed by a data processing unit. Furthermore, in some embodiments, a computer program product is provided comprising instructions that, when executed on at least one processor of a processing device, cause the processing device to perform the method shown in fig. 3. Further, in some embodiments, a non-transitory computer readable storage medium storing one or more programs configured for execution by one or more processors of a processing device is provided, the one or more programs comprising instructions, which when executed by the processing device, cause the processing device to perform the method shown in fig. 3.
Fig. 5 illustrates an update unit, according to some embodiments. The update unit 150a is for the node 130 a. However, all update units 150, 150a (for all nodes) are the same or similar. The update unit 150a receives each respective input 132a, …, 132c of the node 130a (or all nodes if the central update unit 150). In addition, the update unit 150a receives the outputs 134a of the nodes 130a (or, if the central update unit 150, the outputs 134a of all nodes). Further, the updating unit 150a includes a correlator 152a. The correlator 152a calculates the correlation of each respective input 132a, …, 132c of the node 130a with the corresponding output (134 a) during the learn mode, producing a correlation value(s) for each input 132a, …, 132 c. In some embodiments, the different calculated (series of) correlation values are compared to each other (to produce a correlation ratio), and the updating of the weights is based on (based on) the comparison. Furthermore, in some embodiments, the updating unit 150a is configured to apply a first function 154 to the correlations (values, ratios) if the node (130 a) belongs to a first group 160 of the plurality of nodes and to apply a second function 156, different from the first function, to the correlations (values, ratios) if the node (130 a) belongs to a second group 162 of the plurality of nodes, in order to update the weights Wa, wb, wc during the learning mode. In some embodiments, the update unit 150a tracks whether a node belongs to the first group 160 or the second group 162 by utilizing a lock table (LUTs). Further, in some embodiments, the updating unit 150a includes a probability value Pa, pb, pc for each weight Wa, wb, wc for increasing the weight. In some embodiments, the updating unit 150a comprises a probability value Pad, pbd, pcd for each weight Wa, wb, wc for reducing the weight, the probability value being 1-Pa, 1-Pb, 1-Pc in some embodiments, i.e. pad=1-Pa, pbd=1-Pb and pcd=1-Pc). In some embodiments, the probability values Pa, pb, pc and optionally the probability value Pad, pbd, pcd are included in the memory unit 158a of the update unit 150 a. In some embodiments, memory unit 158a is a lock table (LUT). In some embodiments, the updating unit 150a applies one of the first and second functions and/or the probability values Pa, pb, pc and optionally the probability value Pad, pbd, pcd to the calculated (series of) correlation values (or resulting correlation ratios) to obtain the updating signal 159, and then applies the updating signal 159 to the weights Wa, wb, wc to update the weights Wa, wb, wc. the function/structure of the update unit for the other nodes 150b, …, 150x is the same as that for the node 150 a. Further, in some embodiments, the central update unit 150 includes each of the update units for each of the nodes 130a, 130b, …,130 x.
Fig. 6 illustrates a compartment according to some embodiments. In some embodiments, each node 130a, 130b, …,130 x includes a plurality of compartments 900. Each compartment is configured with a plurality of compartment inputs 910a, 910b, …, 910x. Further, each compartment 900 includes compartment weights 920a, 920b, …, 920x for each compartment input 910a, 910b, …, 910x. Further, each compartment 900 is configured to produce a compartment output 940. In some embodiments, the compartment output 940 is calculated by the compartment as a weighted combination, such as a sum, of all compartment inputs 930a, 930b, …, 930x to the compartment. To calculate the sum/combination, the compartment may be equipped with a summer (or adder/summing unit) 935. Each compartment 900 includes an updating unit 995, the updating unit 995 being configured to update the compartment weights 920a, 920b, …, 920x based on (according to) the correlations during the learning mode (in the same manner as described above in connection with fig. 5 and elsewhere for the updating unit 150a, and for one or more compartments may include evaluating each input of the node based on a scoring function). Further, the compartment output 940 of each compartment is used to adjust the output 134a, 134b, …,134 x (e.g., 134 a) including the nodes 130a, 130b, …,130 x (e.g., 130 a) of the compartment 900 based on (according to) the transfer function. Examples of transfer functions that may be utilized are time constants, such as one or more of RC filters, resistors, spike generators, and active elements, such as transistors or operational amplifiers. The compartments 900a, 900b, …, 900x may include sub-compartments 900aa, 900ab, …, 900ba, 900bb, …, 900xx. Thus, each compartment 900a, 900b, …, 900x may have sub-compartments 900aa, 900ab, …, 900ba, 900bb, sub-compartments, etc. functionally connected to the compartment, i.e. the compartments are cascaded. The number of compartments (and sub-compartments) of a particular node is based on (based on) the type of input to the particular node, such as a suppression input, a sensor input, and an excitatory input. Further, the compartments 900 of the nodes are arranged such that each compartment 900 has a majority of one type of input (e.g., inhibitory input, sensor input, or excitatory input). Thus, any type of input (e.g., inhibitory input, sensor input, or excitatory input) is not allowed to become too dominant.
In some embodiments, still referring to fig. 6, the update unit 995 of each compartment 900 includes probability values PCa, …, PCy for increasing weights (and possibly probability values PCad, …, PCyd for decreasing weights, in some embodiments probability values 1-PCa, …, 1-PCy, i.e., PCad = 1-PCa, PCbd = 1-PCb, etc.) for each compartment weight 920a, 920b, …, 920 x. In these embodiments, during the learn mode, the data processing system 100 is configured to calculate a third set point of the sum of all compartment weights 920a, 920b, …, 920x associated with the compartment inputs 910a, 910b, …, 910x to the compartment 900, the data processing system 100 is configured to calculate the sum of all compartment weights 920a, 920b, …, 920x associated with the compartment inputs 910a, 910b, …, 910x to the compartment 900, the data processing system 100 is configured to compare the calculated sum with the third set point, and if the calculated sum is greater than the third set point, to decrease the probability value PCa, …, PCy associated with the weight 920a, 920b, …, 920x (associated therewith) of the compartment input 910a, 910b, …, 910x to the compartment 900, and if the calculated sum is less than the third set point, to increase the probability value PCa, 920a, 920b, 920x associated with the weight 920a, 920b, 920x (associated with the weight 920b, 920 x) of the compartment input 910a, …, 910x to the compartment 900. The third set point is based on (according to) the type of input, such as a system input (sensor input), an input from a node in the first set 160 of the plurality of nodes (excitatory input), or an input from a node in the second set 162 of the plurality of nodes (inhibitory input).
In some embodiments, data processing system 100 is a time-continuous data processing system, i.e., all signals within data processing system 100, including signals between different nodes and including one or more system inputs 110a, 110b, …, 110z and system output 120 are time-continuous (e.g., without spikes).
List of embodiments:
Example 1, a data processing system (100) configured with one or more system inputs (110 a, 110b, …, 110 z) including data to be processed and a system output (120), comprising:
A network NW (130) comprising a plurality of nodes (130 a, 130b, …, 130 x), each node being configured to have a plurality of inputs (132 a, 132b, …, 132 y), each node (130 a, 130b, …, 130 x) comprising a weight (Wa, …, wy) for each input (132 a, 132b, …, 132), and each node being configured to produce an output (134 a, 134b, …, 134 x); and
One or more updating units (150) configured to update the weight (Wa, …, wy) of each node (130 a) during a learning mode based on a correlation of each respective input (132 a, …, 132 c) of the node with the corresponding output (134 a);
One or more processing units (140 x) configured to receive the processing unit input and configured to generate a processing unit output by changing a sign of the received processing unit input; and
Wherein the system output (120) comprises the output (134 a, 134b, …, 134 x) of each node (130 a, 130b, …, 130 x),
Wherein the nodes (130 a, 130 b) in the first group (160) of the plurality of nodes are configured to stimulate one or more other nodes (…, 130 x) in the plurality of nodes (130 a, 130b, …, 130 x) by providing an output (134 a, 134 b) of each node (130 a, 130 b) in the first group (160) of nodes as an input (132 d, …,132 y) to the one or more other nodes (…, 130 x),
Wherein a node (130 x) in a second group (162) of the plurality of nodes is configured to suppress one or more other nodes (130 a, 130b, …) in the plurality of nodes (130 a, 130b, …, 130 x) by providing the output (134 x) of each of the nodes (130 x) in the second group (162) as a processing unit input to a respective processing unit (140 x), each respective processing unit (140 x) being configured to provide the processing unit output as an input (132 b, 132e, …) to the one or more other nodes (130 a, 130b, …), and
Wherein each node of the plurality of nodes (130 a, 130b, …, 130 x) belongs to one of the first and second groups (160, 162) of nodes.
Example 2, the data processing system of example 1, wherein the system input includes sensor data for a plurality of contexts/tasks.
Example 3, the data processing system of any of examples 1-2, wherein the update unit 150 includes a probability value (Pa, …, py) for each weight (Wa, …, wy) for increasing the weight, and wherein during the learning mode, the data processing system is configured to limit a node's (130 a) ability to suppress or energize the one or more other nodes (130 b, …, 130 x) by: providing a first set point of a sum of all weights (Wd, wy) associated with the inputs (132 d, …, 132 y) to the one or more other nodes (130 b, …, 130 x), decreasing a probability value (Pd, py) associated with the weights (Wd, wy) and a sum of all weights (Wd, wy) associated with the inputs (132 d, …, 132 y) to the one or more other nodes (130 b, …, 130 x), increasing the probability value (Pd, py) associated with the weights (Wd, wy) to one or more other nodes (130 b, …, 130 x) and the sum of all weights (Wd, wy) associated with the inputs (132 d, 3434, 132 y) to the one or more other nodes (130 b, …, 130 x), if the first set point is less than the sum of all weights (Pd, wy) associated with the inputs (132 d, 3434, 132 y) to the one or more other nodes (130 b, …, 130 x), increasing the probability value (Pd, wy) associated with the one or more inputs (Pd, 34, 132 y) to the one or more other nodes (132 b, 132 x), and increasing the probability value (Pd, wy) associated with the one or more inputs (Pd, 132 y) to the sum of the inputs (132 d, 34 y).
Example 4, the data processing system of any of examples 1 to 3, wherein during the learn mode, the data processing system is configured to limit an ability of a system input (110 z) to inhibit or stimulate one or more nodes (130 a, …, 130 x) by: -providing the first set point of a sum of all weights (Wg, wx) associated with the inputs (132 g, 132 x) to the one or more nodes (130 a, …, 130 x), -comparing the first set point with a sum of all weights (Wg, wx) associated with the inputs (132 g, 132 x) to the one or more nodes (130 a, …, 130 x), -reducing the probability value (Pg, px) associated with the weights (Wg, wx), if the first set point is smaller than a sum of all weights (Wg, wx) associated with the inputs (132 g, 132 x) to the one or more nodes (130 a, …, 130 x), if the weights (Wg, wx) associated with the inputs (132 g, 132 x) to the one or more nodes (130 a, …, 130 x) are greater than the probability value (Pg, px) associated with the one or more inputs (132 g, 35 x) associated with the one or more nodes (132 g, 132 x), if the weights (Wg, 35 x) associated with the one or more inputs (132 g, 132 x) are greater than one or more than one of the sum of the weights (Pg, px) associated with the inputs (132 g, 132 x).
Example 5, the data processing system of any of examples 3 to 4, wherein each of the inputs (132 d, …, 132 y) to the one or more other nodes (130 b, …, 130 x) has coordinates in a network space, wherein the decrease/increase in the weight (Wd, wy) of the input (132 d, 132 y) to the one or more other nodes (130 b, …, 130 x) is based on a distance between coordinates of the input (132 d, 132 y) associated with the weight (Wd, wy) in the network space.
Example 6, the data processing system of any of examples 3-5, wherein the system is further configured to set the weight (Wa, …, wy) to zero if the weight (Wa, …, wy) does not increase within a preset period of time; and/or
Wherein the system is further configured to increase a probability value (Pa, …, py) of a weight (Wa, …, wy) having a zero value if a sum of all weights (Wd, wy) associated with the inputs (132 d, …, 132 y) to the one or more other nodes (130 b, …, 130 x) does not exceed the first set point within a preset period of time.
Example 7, the data processing system of any of examples 1-2, wherein during the learning mode, the data processing system is configured to increase a correlation of the output (134 a) of a node (130 a) with the one or more other nodes (130 b, …, 130 x) by: -providing a first set point of a sum of all weights (Wd, wy) associated with the inputs (132 d, …, 132 y) to the one or more other nodes (130 b, …, 130 x), -comparing the first set point with a sum of all weights (Wd, wy) associated with the inputs (132 d, …, 132 y) to the one or more other nodes (130 b, …, 130 x) over a first time period, -increasing a probability of changing the weights (Wa, wb, 132 c) to the node (130 a) and-decreasing a probability of changing the weights (Wd, wc) to the one or more other nodes (132 b, …, 130 x) over the first time period by less than a sum of all weights (Wd, wy) associated with the inputs (132 d, …, 132 y) to the one or more other nodes (130 b, …, 130 x), -decreasing a probability of changing the weights (Wd, wc) to the one or more nodes (132 b, 132 c) over the first time period by more than the sum of all weights (Wd, wc) associated with the inputs (132 b, 132 c).
Example 8, the data processing system of any of examples 1-2, wherein the update unit (150) comprises a probability value (Pa, …, py) for increasing the weight for each weight (Wa, …, gy), and wherein during the learning mode the data processing system is configured to provide a second set point of a sum of all weights (Wa, wb, wc) associated with the inputs (132 a, 132b, 132 c) to a node (130 a), the data processing system is configured to calculate a sum of all weights (Wa, wb, wc) associated with the inputs (132 a, 132b, 132 c) to the node (130 a), the data processing system is configured to compare the calculated sum with the second set point, and if the calculated sum is greater than the second set point, to reduce the probability value (Pa, wb, wc) associated with the weights (Wa, pb, wb, wc) and if the calculated sum is less than the second set point, the probability value (Pb, wb, wc) associated with the inputs (132 a, 132b, wc) to the node (132 a).
Example 9, the data processing system of any of examples 1-2, wherein each node (130 a, 130b, …, 130 x) comprises a plurality of compartments (900), and each compartment is configured to have a plurality of compartment inputs (910 a, 910b, …, 910 x), each compartment (900) comprises a compartment weight (920 a, 920b, …, 920 x) for each compartment input (910 a, 910b, …, 910 x), and each compartment (900) is configured to generate a compartment output (940), and wherein each compartment (900) comprises an updating unit (995) configured to update a compartment weight (920 a, 920b, …, 920 x) based on a correlation during the learning mode and wherein the compartment output (940) of each compartment is used to adjust the output (134 a, 134b, 134 x) of the node (130 a, 130b, …, 130 x) comprising the compartment based on a transfer function.
Example 10, the data processing system of example 9, wherein the update unit (995) of each compartment (900) includes a third set point for each compartment weight (920 a, 920b, …, 920 x) for increasing the probability value (PCa, …, PCy) of the weight, and wherein during the learning mode the data processing system is configured to provide a sum of all compartment weights (920 a, 920b, …, 920 x) associated with the compartment input (910 a, 910b, …, 910 x) to compartment (900), the data processing system is configured to calculate a sum of all compartment weights (920 a, 920b, …, 910 x) associated with compartment weights (920 a, 920b, …, 920 x) to compartment (900), the data processing system is configured to compare the calculated sum with the third set point, and if the calculated sum is greater than the third set point, the probability value is configured to decrease the data processing system as a sum of all compartment weights (920 a, 920b, …, 920 x) associated with compartment weights (920 a, 920b, 920 x) and the sum of the compartment weights (920 a, 920b, …, 910 x) associated with compartment weights (920 a, 920b, 920 x) to compartment (900), the data processing system is configured to decrease the sum of the calculated sum of all compartment weights (920 a, 920b, …, 920 x) associated with the calculated sum of the compartment weights (920 a, 920b, 920 x) associated with the calculated sum of the compartment weights (920 a, 920 x) and the calculated by the sum of the calculated by the data processing system is greater than is calculated by the sum of the calculated by the sum, 920 x) is associated with the compartment input (910 a, 910b, …, 910 x) to the compartment (900), and wherein the third setpoint is based on a type of input, such as a system input, an input from a node in the first set (160) of the plurality of nodes, or an input from a node in the second set (162) of the plurality of nodes.
Example 11, the data processing system of any of examples 1-2, wherein during the learning mode, the data processing system is configured to:
Detecting whether the network (130) is sparsely connected by comparing a cumulative weight change of the system inputs (110 a, 110b, …, 110 z) over a second period of time with a threshold; and
If the data processing system detects a sparse connection of the network (130), the output (134 a, 134b, …, 134 x) of one or more of the plurality of nodes (130 a, 130b, …, 130 x) is increased by adding a predetermined waveform to the output (134 a, 134b, …, 134 x) of one or more of the plurality of nodes (130 a, 130b, …, 130 x) for a duration of a third period of time.
Example 12, the data processing system of any of examples 1-11, wherein each node comprises an updating unit (150), wherein each updating unit (150) is configured to update a weight (Wa, wb, wc) of each respective input (132 a, …, 132 c) of the node (130 a) based on a correlation of the respective input to the output (134 a) of the node (130 a), and wherein each updating unit (150) is configured to apply a first function to the correlation if the associated node belongs to a first group (160) of the plurality of nodes, and to apply a second function, different from the first function, to the correlation if the associated node belongs to the second group (162) of the plurality of nodes, to update the weight (Wa, wb, wc) during the learning mode.
Example 13, the data processing system of any of examples 1-12, wherein the data processing system is configured to: after the updating of the weights (Wa, …, wy) has been performed, calculating a global variance of the outputs (134 a, 134b, …, 134 x) of the nodes (130 a, 130b, …, 130 x) of the network, comparing the calculated global variance with a power law; and minimizing the error or mean square error between the ensemble and the power law by adjusting parameters of the network.
Example 14, the data processing system of any of examples 2-13, wherein the data processing system is configured to learn from the sensor data to identify one or more entities when in a learning mode, and thereafter configured to identify the one or more entities when in an execution mode, and wherein the identified entities are one or more of: a speaker, spoken letter, syllable, phoneme, word or phrase present in the sensor data, or an object or feature of an object present in the sensor data, or a new contact event present in the sensor data, an end of a contact event, a gesture or an applied pressure.
Example 15, a computer-implemented or hardware-implemented method (300) for processing data, comprising:
a) Receiving (310) one or more system inputs (110 a, 110b, …, 110 z) comprising data to be processed;
b) Providing (320) a plurality of inputs (132 a, 132b, …, 132 y) to a network NW (130) comprising a plurality of first nodes (130 a, 130b, …, 130 x), at least one of the plurality of inputs being a system input;
c) -receiving (330) an output (134 a, 134b, …, 134 x) from each first node (130 a, 130b, …, 130 x);
d) -providing (340) a system output (120), the system output (120) comprising the output (134 a, 134b, …, 134 x) of each first node (130 a, 130b, …, 130 x);
e) Exciting (350) one or more other nodes (…, 130 x) in a first set (160) of the plurality of nodes by a node (130 a, 130 b) in the first set (160) of the plurality of nodes by providing an output (134 a, 134 b) of each node (130 a, 130 b) as an input (132 d, …, 132 y) to the one or more other nodes (…, 130 x) in the plurality of nodes (130 a, 130b, …, 130 x);
f) Suppressing (360) one or more other nodes (130 a, 130b, …) of the plurality of nodes (130 a, 130b, …, 130 x) by a node (130 x) of a second group (162) of the plurality of nodes by providing an output (134 x) of each node (130 x) of the second group (162) as a processing unit input to a respective processing unit (140 x), each respective processing unit (140 x) configured to provide the processing unit output as an input (132 b, 132e, …) to the one or more other nodes (130 a, 130b, …);
g) Optionally, updating (370) the weights (Wa, …, wy) by one or more updating units (150) based on the correlations; and
H) -optionally repeating (380) a) -g) until a learning criterion is met;
i) Repeating (390) a) -f) until the stopping criteria are met, and
Wherein each node of the plurality of nodes (130 a, 130b, …, 130 x) belongs to one of the first and second groups (160, 162) of nodes.
Example 16, the method of example 15, further comprising:
Initializing (304) the weights (Wa, …, wy) by setting the weights (Wa, …, wy) to zero; and
A predetermined waveform is added (308) to the output (134 a, 134b, …, 134 x) of one or more of the plurality of nodes (130 a, 130b, …, 130 x) for a duration of a third period of time, the third period of time beginning at a time of receiving (310) one or more system inputs (110 a, 110b, …, 110 z) comprising data to be processed.
Example 17, the method of example 15, further comprising:
Initializing (306) weights (Wa, …, wy) by randomly assigning values between 0 and 1 to the weights (Wa, …, wy); and
A predetermined waveform is added (308) to the output (134 a, 134b, …, 134 x) of one or more of the plurality of nodes (130 a, 130b, …, 130 x) for a duration of a third period of time.
Example 18, a computer program product comprising a non-transitory computer readable medium (400), having stored thereon a computer program comprising program instructions, the computer program being loadable into a data-processing unit (420) and configured to perform the method according to any of examples 15-17 when the computer program is run by the data-processing unit (420).
Those skilled in the art will recognize that the present disclosure is not limited to the preferred embodiments described above. Those skilled in the art will further recognize that modifications and variations may be made within the scope of the appended claims. For example, signals from other sensors, such as fragrance sensors or flavor sensors, may be processed by the data processing system. Furthermore, the described data processing system may equally well be used for unsegmented, connected handwriting recognition, speech recognition, speaker recognition and anomaly detection in a network traffic or Intrusion Detection System (IDSs). Further, variations to the disclosed embodiments can be understood and effected by those skilled in the art in practicing the claimed disclosure, from a study of the drawings, the disclosure, and the appended claims.

Claims (26)

1.一种数据处理系统(100),被配置为具有包括待处理数据的一个或多个系统输入(110a、110b、…、110z)和系统输出(120),包括:1. A data processing system (100) configured to have one or more system inputs (110a, 110b, ..., 110z) and a system output (120) including data to be processed, comprising: 包括多个节点(130a、130b、…、130x)的网络NW(130),每个节点被配置为具有多个输入(132a、132b、…、132y),每个节点(130a、130b、…、130x)包括针对每个输入(132a、132b、…、132y)的权重(Wa、…、Wy),并且每个节点被配置为产生输出(134a、134b、…、134x);以及A network NW (130) comprising a plurality of nodes (130a, 130b, ..., 130x), each node being configured to have a plurality of inputs (132a, 132b, ..., 132y), each node (130a, 130b, ..., 130x) comprising a weight (Wa, ..., Wy) for each input (132a, 132b, ..., 132y), and each node being configured to generate an output (134a, 134b, ..., 134x); and 一个或多个处理单元(140x),被配置为接收处理单元输入,并且被配置为通过改变所接收的处理单元输入的符号来产生处理单元输出;并且one or more processing units (140x) configured to receive a processing unit input and configured to generate a processing unit output by changing the sign of the received processing unit input; and 其中所述系统输出(120)包括每个节点(130a、130b、…、130x)的所述输出(134a、134b、…、134x),wherein the system output (120) includes the output (134a, 134b, ..., 134x) of each node (130a, 130b, ..., 130x), 其中所述多个节点中的第一组(160)中的节点(130a、130b)被配置为通过将节点中的所述第一组(160)中的所述节点(130a、130b、…、130x)中的每个所述节点(130a、130b)的所述输出(134a、134b)作为输入提供给所述多个节点(130a、130b、…、130x)中的一个或多个其他节点来激励所述多个节点(130a、130b、…、130x)中的一个或多个其他节点,wherein the nodes (130a, 130b) in the first group (160) of the plurality of nodes are configured to stimulate one or more other nodes of the plurality of nodes (130a, 130b, ..., 130x) by providing the output (134a, 134b) of each of the nodes (130a, 130b) in the first group (160) of nodes (130a, 130b, ..., 130x) as an input to one or more other nodes of the plurality of nodes (130a, 130b, ..., 130x), 其中,所述多个节点中的第二组(162)中的节点(130x)被配置为通过将所述第二组(162)中的每个所述节点(130x)的所述输出(134x)作为处理单元输入提供给相应的处理单元(140x)来抑制所述多个节点(130a、130b、…、130x)中的一个或多个其他节点,每个相应的处理单元(140x)被配置为将所述处理单元输出作为输入(132b、132e、…)提供给所述一个或多个其他节点,wherein a node (130x) in a second group (162) of the plurality of nodes is configured to suppress one or more other nodes of the plurality of nodes (130a, 130b, ..., 130x) by providing the output (134x) of each of the nodes (130x) in the second group (162) as a processing unit input to a corresponding processing unit (140x), each corresponding processing unit (140x) being configured to provide the processing unit output as an input (132b, 132e, ...) to the one or more other nodes, 其中,所述多个节点(130a、130b、…、130x)中的每个节点属于节点中的所述第一组和第二组(160、162)中的一个,wherein each of the plurality of nodes (130a, 130b, ..., 130x) belongs to one of the first group and the second group (160, 162) of nodes, 其中,每个节点包括更新单元(150),其中,每个更新单元(150)被配置为基于节点(130a)的每个相应输入(132a、…、132c)与所述节点(130a)的所述输出(134a)的相关性来更新相应节点(130a)的权重(Wa、Wb、Wc),以及wherein each node comprises an updating unit (150), wherein each updating unit (150) is configured to update a weight (Wa, Wb, Wc) of a corresponding node (130a) based on a correlation of each corresponding input (132a, ..., 132c) of the node (130a) with the output (134a) of the node (130a), and 其中,每个更新单元(150)被配置为若所述关联节点属于所述多个节点中的所述第一组(160),则将第一函数应用于所述相关性,并且若所述关联节点属于所述多个节点中的所述第二组(162),则将不同于所述第一函数的第二函数应用于所述相关性,以便在所述学习模式期间更新所述权重(Wa、Wb、Wc)。Wherein, each updating unit (150) is configured to apply a first function to the correlation if the associated node belongs to the first group (160) of the plurality of nodes, and to apply a second function different from the first function to the correlation if the associated node belongs to the second group (162) of the plurality of nodes, so as to update the weights (Wa, Wb, Wc) during the learning mode. 2.根据权利要求1所述的数据处理系统,其中所述一个或多个系统输入包括多个语境/任务的传感器数据。2. The data processing system of claim 1, wherein the one or more system inputs include sensor data of multiple contexts/tasks. 3.根据权利要求1-2中任一项所述的数据处理系统,其中,所述更新单元150包括针对每个权重(Wa、…、Wy)的用于增加所述权重的概率值(Pa、…、Py),并且其中,在所述学习模式期间,所述数据处理系统被配置为通过以下方式来限制节点(130a)抑制或激励所述一个或多个其他节点(130b、…、130x)的能力:提供与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和的第一设定点,将所述第一设定点和与到所述一个或多个其他节点的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和进行比较,若所述第一设定点小于与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和,则减小与所述权重(Wd、Wy)相关联的概率值(Pd、Py),所述权重(Wd、Wy)与到所述一个或多个其他节点(130b、…、130x)的所述输入(13b、…、130x)相关联,以及若所述第一设定点大于与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和,则增加与所述权重(Wd,Wy)相关联的所述概率值(Pd、Py),所述权重(Wd、Wy)与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联。3. The data processing system according to any one of claims 1-2, wherein the updating unit 150 includes, for each weight (Wa, ..., Wy), a probability value (Pa, ..., Py) for increasing the weight, and wherein, during the learning mode, the data processing system is configured to limit the ability of the node (130a) to inhibit or stimulate the one or more other nodes (130b, ..., 130x) by providing a first set point of the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x), comparing the first set point with the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes, and if the first set point is less than the sum of all weights (Wd, Wy) associated with the inputs to the one or more other nodes, If the first set point is greater than the sum of all weights (Wd, Wy) associated with the input (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x), the probability value (Pd, Py) associated with the weight (Wd, Wy) associated with the input (13b, ..., 130x) to the one or more other nodes (130b, ..., 130x) is increased. If the first set point is greater than the sum of all weights (Wd, Wy) associated with the input (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x), the probability value (Pd, Py) associated with the weight (Wd, Wy) associated with the input (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x) is increased. 4.根据权利要求1-3中任一项所述的数据处理系统,其中,在所述学习模式期间,所述数据处理系统被配置为通过以下方式来限制系统输入(110z)抑制或激励一个或多个节点(130a、…、130x)的能力:提供与到所述一个或多个节点(130a、…、130x)的所述输入(132g、132x)相关联的所有权重(Wg、Wx)的总和的所述第一设定点,将所述第一设定点和与到所述一个或多个节点(130a、…、130x)的所述输入(132g、132x)相关联的所有权重(Wg、Wx)的总和进行比较,若所述第一设定点小于与到所述一个或多个节点(130a、…、130x)的所述输入(132g、132x)相关联的所有权重(Wg、Wx)的总和,则减小与所述权重(Wg、Wx)相关联的所述概率值(Pg、Px),所述权重(Wg、Wx)与到所述一个或多个节点(130a、…、130x)的所述输入(132g、132x)相关联,以及若所述第一设定点大于与到所述一个或多个节点(130a、…、130x)的所述输入(132g、132x)相关联的所有权重(Wg、Wx)的总和,则增加与所述权重(Wg、Wx)相关联的所述概率值(Pg、Px),所述权重(Wg、Wx)与到所述一个或多个节点(130a、…、130x)的输入(132g、132x)相关联。4. The data processing system of any one of claims 1 to 3, wherein during the learning mode, the data processing system is configured to limit the ability of a system input (110z) to inhibit or stimulate one or more nodes (130a, ..., 130x) by providing a first set point that is a sum of all weights (Wg, Wx) associated with the inputs (132g, 132x) to the one or more nodes (130a, ..., 130x), comparing the first set point to the sum of all weights (Wg, Wx) associated with the inputs (132g, 132x) to the one or more nodes (130a, ..., 130x), and if the first set point is less than the sum of all weights (Wg, Wx) associated with the inputs (132g, 132x) to the one or more nodes (130a, ..., 130x), the first set point is compared to the sum of all weights (Wg, Wx) associated with the inputs (132g, 132x) to the one or more nodes (130a, ..., 130x). If the first set point is greater than the sum of all weights (Wg, Wx) associated with the input (132g, 132x) to the one or more nodes (130a, ..., 130x), then the probability value (Pg, Px) associated with the weight (Wg, Wx) is reduced, and the weight (Wg, Wx) is associated with the input (132g, 132x) to the one or more nodes (130a, ..., 130x), and if the first set point is greater than the sum of all weights (Wg, Wx) associated with the input (132g, 132x) to the one or more nodes (130a, ..., 130x), then the probability value (Pg, Px) associated with the weight (Wg, Wx) is increased, and the weight (Wg, Wx) is associated with the input (132g, 132x) to the one or more nodes (130a, ..., 130x). 5.根据权利要求3-4中任一项所述的数据处理系统,其中,到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)中的每个输入在网络空间中具有坐标,其中,到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、132y)的所述权重(Wd、Wy)的减少/增加量是基于与所述权重(Wd、Wy)相关联的所述输入(132d、132y)在所述网络空间中的坐标之间的距离。5. A data processing system according to any one of claims 3-4, wherein each of the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x) has coordinates in a network space, and wherein the amount of decrease/increase in the weight (Wd, Wy) of the input (132d, 132y) to the one or more other nodes (130b, ..., 130x) is based on the distance between the coordinates of the input (132d, 132y) associated with the weight (Wd, Wy) in the network space. 6.根据权利要求3-5中任一项所述的数据处理系统,其中,所述系统还被配置为若权重(Wa、…、Wy)在预设时间段内不增加,则将所述权重(Wa、…、Wy)设置为零;和/或6. A data processing system according to any one of claims 3 to 5, wherein the system is further configured to set the weight (Wa, ..., Wy) to zero if the weight (Wa, ..., Wy) does not increase within a preset time period; and/or 其中,所述系统还被配置为若与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和在预设时间段内不超过所述第一设定点,则增加具有零值的权重(Wa、…、Wy)的概率值(Pa、…、Py)。Wherein, the system is also configured to increase the probability value (Pa, ..., Py) of the weight (Wa, ..., Wy) having a zero value if the sum of all weights (Wd, Wy) associated with the input (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x) does not exceed the first set point within a preset time period. 7.根据权利要求1-2中任一项所述的数据处理系统,其中,在所述学习模式期间,所述数据处理系统被配置为通过以下方式来增加节点(130a)的所述输出(134a)与所述一个或多个其他节点(130b、…、130x)的相关性:提供与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和的第一设定点,在第一时间段内,将所述第一设定点和与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和进行比较,若在所述第一时间段的整个长度上所述第一设定点小于与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和,则增加改变到所述节点(130a)的所述输入(132a、132b、132c)的所述权重(Wa、Wb、Wc)的概率,以及若在所述第一时间段的整个长度上所述第一设定点大于与到所述一个或多个其他节点(130b、…、130x)的所述输入(132d、…、132y)相关联的所有权重(Wd、Wy)的总和,则减小改变到所述节点(130a)的所述输入(132a、132b、132c)的所述权重(Wa、Wb、Wc)的概率。7. A data processing system according to any one of claims 1-2, wherein, during the learning mode, the data processing system is configured to increase the relevance of the output (134a) of the node (130a) to the one or more other nodes (130b, ..., 130x) by providing a first set point of the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x), and within a first time period, comparing the first set point with the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x), and if the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x) is greater than the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x), the first set point is compared with the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x) during the entire first time period. If the first set point is less than the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x) over the length, the probability of changing the weights (Wa, Wb, Wc) of the inputs (132a, 132b, 132c) to the node (130a) is increased, and if the first set point is greater than the sum of all weights (Wd, Wy) associated with the inputs (132d, ..., 132y) to the one or more other nodes (130b, ..., 130x) over the entire length of the first time period, the probability of changing the weights (Wa, Wb, Wc) of the inputs (132a, 132b, 132c) to the node (130a) is reduced. 8.根据权利要求1-2中任一项所述的数据处理系统,其中,所述更新单元(150)包括针对每个权重(Wa、…、Wy)的用于增加所述权重的概率值(Pa、…、Py),并且其中,在所述学习模式期间,所述数据处理系统被配置为提供与到节点(130a)的所述输入(132a、132b、132c)相关联的所有权重(Wa、Wb、Wc)的总和的第二设定点,所述数据处理系统被配置为计算与到所述节点(130a)的所述输入(132a、132b、132c)相关联的所有权重(Wa、Wb、Wc)的总和,所述数据处理系统被配置为将所计算的总和与所述第二设定点进行比较,并且若所计算的总和大于所述第二设定点,则被配置为减小与所述权重(Wa、Wb、Wc)相关联的所述概率值(Pa、Pb、Pc),所述权重(Wa、Wb、Wc)与到所述节点(130a)的所述输入(132a、132b、132c)相关联,以及若所计算的总和小于所述第二设定点,则被配置为增加与所述权重(Wa、Wb、Wc)相关联的所述概率值(Pa、Pb、Pc),所述权重(Wa、Wb、Wc)与到所述节点(130a)的所述输入(132a、132b、132c)相关联。8. A data processing system according to any one of claims 1-2, wherein the updating unit (150) comprises, for each weight (Wa, ..., Wy), a probability value (Pa, ..., Py) for increasing the weight, and wherein, during the learning mode, the data processing system is configured to provide a second set point for the sum of all weights (Wa, Wb, Wc) associated with the inputs (132a, 132b, 132c) to the node (130a), the data processing system being configured to calculate the sum of all weights (Wa, Wb, Wc) associated with the inputs (132a, 132b, 132c) to the node (130a), the data processing system being configured to The device is configured to compare the calculated sum with the second set point, and if the calculated sum is greater than the second set point, it is configured to reduce the probability value (Pa, Pb, Pc) associated with the weight (Wa, Wb, Wc) associated with the input (132a, 132b, 132c) to the node (130a), and if the calculated sum is less than the second set point, it is configured to increase the probability value (Pa, Pb, Pc) associated with the weight (Wa, Wb, Wc) associated with the input (132a, 132b, 132c) to the node (130a). 9.根据权利要求1-2中任一项所述的数据处理系统,其中每个节点(130a、130b、…、130x)包括多个隔室(900),并且每个隔室被配置为具有多个隔室输入(910a、910b、…、910x),每个隔室(900)包括针对每个隔室输入(910a、910b、…、910x)的隔室权重(920a、920b、…、920x),并且每个隔室(900)被配置为产生隔室输出(940),并且其中每个隔室(900)包括更新单元(995),所述更新单元(995)被配置为在所述学习模式期间基于相关性更新所述隔室权重(920a、920b、…、920x)并且其中每个隔室的所述隔室输出(940)被用于基于传递函数来调整包括所述隔室的所述节点(130a、130b、…、130x)的所述输出(134a、134b、…、134x)。9. The data processing system according to any one of claims 1 to 2, wherein each node (130a, 130b, ..., 130x) comprises a plurality of compartments (900), and each compartment is configured to have a plurality of compartment inputs (910a, 910b, ..., 910x), each compartment (900) comprises a compartment weight (920a, 920b, ..., 920x) for each compartment input (910a, 910b, ..., 910x), and each compartment (900) is configured to produce The invention relates to a method for generating a compartment output (940) and wherein each compartment (900) comprises an update unit (995) configured to update the compartment weights (920a, 920b, ..., 920x) based on the correlation during the learning mode and wherein the compartment output (940) of each compartment is used to adjust the output (134a, 134b, ..., 134x) of the node (130a, 130b, ..., 130x) comprising the compartment based on a transfer function. 10.根据权利要求9所述的数据处理系统,其中,每个隔室(900)的所述更新单元(995)包括针对每个隔室权重(920a、920b、…、920x)的用于增加所述权重的概率值(PCa、…、PCy),并且其中,在所述学习模式期间,所述数据处理系统被配置为提供与到隔室(900)的所述隔室输入(910a、910b、…、910x)相关联的所有隔室权重(920a、920b、…、920x)的总和的第三设定点,所述数据处理系统被配置为计算与到隔室(900)的所述隔室输入(910a、910b、…、910x)相关联的所有隔室权重(920a、920b、…、920x)的总和,所述数据处理系统被配置为将所计算的总和与所述第三设定点进行比较,并且若所述计算总和大于所述第三设定点,则所述数据处理系统被配置为减小与所述隔室权重(920a、920b、…、920x)相关联的概率值(PCa、…、PCy),所述隔室权重(920a、920b、…、920x)与到所述隔室(900)的所述隔室输入(910a、910b、…、910x)相关联,以及若所述计算总和小于所述第三设定点,则所述数据处理系统被配置为增大与所述隔室权重(920a、920b、…、920x)相关联的概率值(PCa、…、PCy),其中所述隔室权重(920a、920b、…、920x)与到所述隔室(900)的所述隔室输入(910a、910b、…、910x)相关联,并且其中所述第三设定点是基于输入的类型的,诸如系统输入、来自所述多个节点中的所述第一组(160)中的节点的输入或来自所述多个节点中的所述第二组(162)中的节点的输入。10. A data processing system according to claim 9, wherein the updating unit (995) of each compartment (900) comprises, for each compartment weight (920a, 920b, ..., 920x), a probability value (PCa, ..., PCy) for increasing the weight, and wherein, during the learning mode, the data processing system is configured to provide all compartment weights associated with the compartment input (910a, 910b, ..., 910x) to the compartment (900) The data processing system is configured to calculate the sum of all compartment weights (920a, 920b, ..., 920x) associated with the compartment inputs (910a, 910b, ..., 910x) to the compartment (900), the data processing system is configured to compare the calculated sum with the third set point, and if the calculated sum is greater than the third set point, the data processing system is configured to The data processing system is configured to decrease the probability values (PCa, ..., PCy) associated with the compartment weights (920a, 920b, ..., 920x) associated with the compartment inputs (910a, 910b, ..., 910x) to the compartment (900), and if the calculated sum is less than the third set point, the data processing system is configured to increase the probability values (PCa, ..., PCy) associated with the compartment weights (920a, 920b, ..., 920x) associated with the compartment inputs (910a, 910b, ..., 910x) to the compartment (900). ), wherein the compartment weights (920a, 920b, ..., 920x) are associated with the compartment inputs (910a, 910b, ..., 910x) to the compartment (900), and wherein the third set point is based on the type of input, such as system input, input from a node in the first group (160) of the plurality of nodes, or input from a node in the second group (162) of the plurality of nodes. 11.根据权利要求1-2中任一项所述的数据处理系统,其中,在所述学习模式期间,所述数据处理系统被配置为:11. The data processing system according to any one of claims 1-2, wherein during the learning mode, the data processing system is configured to: 通过将所述一个或多个系统输入(110a、110b、…、110z)在第二时间段内的累积权重变化与阈值进行比较来检测所述网络(130)是否稀疏连接;以及detecting whether the network (130) is sparsely connected by comparing a cumulative weight change of the one or more system inputs (110a, 110b, ..., 110z) over a second time period to a threshold; and 若所述数据处理系统检测到所述网络(130)稀疏连接,则通过在第三时间段的持续时间内将预定波形添加到所述多个节点(130a、130b、…、130x)中的一个或多个的所述输出(134a、134b、…、134x)来增加所述多个节点(130a、130b、…、130x)中的一个或多个的所述输出(134a、134b、…、134x)。If the data processing system detects that the network (130) is sparsely connected, the output (134a, 134b, ..., 134x) of one or more of the multiple nodes (130a, 130b, ..., 130x) is increased by adding a predetermined waveform to the output (134a, 134b, ..., 134x) of one or more of the multiple nodes (130a, 130b, ..., 130x) for the duration of a third time period. 12.根据权利要求1至11中任一项所述的数据处理系统,其中,所述数据处理系统被配置为:在已经执行所述权重(Wa、…、Wy)的更新之后,计算所述网络(130)的所述节点(130a、130b、…、130x)的所述输出(134a、134b、…、134x)的总体方差,将所计算的总体方差与幂律进行比较;以及通过调整所述网络(130)的参数来最小化所述总体与所述幂律之间的误差或均方误差。12. A data processing system according to any one of claims 1 to 11, wherein the data processing system is configured to: calculate the population variance of the outputs (134a, 134b, ..., 134x) of the nodes (130a, 130b, ..., 130x) of the network (130) after the update of the weights (Wa, ..., Wy) has been performed, compare the calculated population variance with a power law; and minimize the error or mean square error between the population and the power law by adjusting the parameters of the network (130). 13.根据权利要求12所述的数据处理系统,其中调整所述网络(130)的参数包括调整以下中的一个或多个:13. The data processing system of claim 12, wherein adjusting the parameters of the network (130) comprises adjusting one or more of: 所述学习的缩放类型,诸如权重的范围;The type of scaling of the learning, such as the range of weights; 当更新时,突触权重的诱导变化,诸如指数地或线性地;When updating, the induced changes in synaptic weights, such as exponentially or linearly; 所述学习中的增益量;an amount of gain in said learning; 每个节点的状态存储器的一个或多个时间常数;One or more time constants for each node's state memory; 一个或多个学习函数,诸如所述第一函数和第二函数;one or more learning functions, such as the first and second functions; 每个节点的传递函数;The transfer function of each node; 节点和传感器之间的连接的总容量;以及The total capacity of the connections between nodes and sensors; and 跨所有节点的节点的总容量。The total capacity of the node across all nodes. 14.根据权利要求2-13中任一项所述的数据处理系统,其中,所述数据处理系统被配置为在处于学习模式时从所述传感器数据中学习以识别一个或多个实体,并且此后所述数据处理系统被配置为在处于执行模式时识别所述一个或多个实体。14. A data processing system according to any one of claims 2-13, wherein the data processing system is configured to learn from the sensor data to identify one or more entities when in a learning mode, and thereafter the data processing system is configured to identify the one or more entities when in an execution mode. 15.根据权利要求14所述的数据处理系统,其中,所识别的实体是所述传感器数据中存在的说话者、口语字母、音节、音素、单词或短语中的一个或多个。15. The data processing system of claim 14, wherein the identified entities are one or more of speakers, spoken letters, syllables, phonemes, words, or phrases present in the sensor data. 16.根据权利要求14所述的数据处理系统,其中,所识别的实体是传感器数据中存在的对象或对象的特征。16. The data processing system of claim 14, wherein the identified entity is an object or a feature of an object present in the sensor data. 17.根据权利要求14所述的数据处理系统,其中,所识别的实体是所述传感器数据中存在的新接触事件、接触事件的结束、手势或施加的压力。17. The data processing system of claim 14, wherein the identified entity is a new contact event, an end of a contact event, a gesture, or an applied pressure present in the sensor data. 18.根据权利要求1-17中任一项所述的数据处理系统,其中,所述网络(130)是循环神经网络。18. A data processing system according to any one of claims 1-17, wherein the network (130) is a recurrent neural network. 19.根据权利要求1-17中任一项所述的数据处理系统,其中,所述网络(130)是递归神经网络。19. The data processing system according to any one of claims 1-17, wherein the network (130) is a recurrent neural network. 20.一种用于处理数据的计算机实现或硬件实现的方法(300),包括:20. A computer-implemented or hardware-implemented method (300) for processing data, comprising: 接收(310)包括待处理数据的一个或多个系统输入(110a、110b、…、110z);Receiving (310) one or more system inputs (110a, 110b, ..., 110z) including data to be processed; 向包括多个第一节点(130a、130b、…、130x)的网络NW(130)提供(320)多个输入(132a、132b、…、132y),所述多个输入中的至少一个是系统输入;Providing (320) a plurality of inputs (132a, 132b, ..., 132y) to a network NW (130) comprising a plurality of first nodes (130a, 130b, ..., 130x), at least one of the plurality of inputs being a system input; 接收(330)来自每个第一节点(130a、130b、…、130x)的输出(134a、134b、…、134x);receiving (330) an output (134a, 134b, ..., 134x) from each first node (130a, 130b, ..., 130x); 提供(340)系统输出(120),所述系统输出(120)包括每个第一节点(130a、130b、…、130x)的所述输出(134a、134b、…、134x);providing (340) a system output (120), the system output (120) comprising the output (134a, 134b, ..., 134x) of each first node (130a, 130b, ..., 130x); 由所述多个节点中的第一组(160)中的节点(130a、130b)通过将所述节点中的所述第一组(160)中的每个节点(130a、130b)的输出(134a、134b)作为输入(132d、…、132y)提供给所述多个节点(130a、130b、…、130x)中的一个或多个其他节点(…、130x),来激励(350)所述一个或多个其他节点(…、130x);stimulating (350) one or more other nodes (..., 130x) of the plurality of nodes (130a, 130b) by providing an output (134a, 134b) of each node (130a, 130b) of the first group (160) of the nodes as an input (132d, ..., 132y) to the other nodes (..., 130x) of the plurality of nodes (130a, 130b, ..., 130x); 由所述多个节点中的第二组(162)中的节点(130x)通过将所述第二组(162)中的每个节点(130x)的所述输出(134x)作为处理单元输入提供给相应的处理单元(140x)来抑制(360)所述多个节点(130a、130b、…、130x)中的一个或多个其他节点(130a、130b、…),每个相应的处理单元(140x)被配置为将所述处理单元输出作为输入(132b、132e、…)提供给所述一个或多个其他节点(130a、130b、…);以及suppressing (360) by a node (130x) in a second group (162) of the plurality of nodes one or more other nodes (130a, 130b, ..., 130x) of the plurality of nodes by providing the output (134x) of each node (130x) in the second group (162) as a processing unit input to a corresponding processing unit (140x), each corresponding processing unit (140x) being configured to provide the processing unit output as an input (132b, 132e, ...) to the one or more other nodes (130a, 130b, ...); and 针对每个节点,基于所述节点(130a)的每个相应输入(132a、…、132c)与该节点(130a)的所述输出(134a)的相关性来更新(370)所述权重(Wa、Wb、Wc),以及若所述关联节点属于所述多个节点中的所述第一组(160),则将第一函数应用于所述相关性,以及若所述关联节点属于所述多个节点中的所述第二组(162),则将不同于所述第一函数的第二函数应用于所述相关性,以便在所述学习模式期间更新所述权重(Wa、Wb、Wc),updating (370) the weights (Wa, Wb, Wc) for each node based on a correlation of each respective input (132a, ..., 132c) of the node (130a) with the output (134a) of the node (130a), and applying a first function to the correlation if the associated node belongs to the first group (160) of the plurality of nodes, and applying a second function different from the first function to the correlation if the associated node belongs to the second group (162) of the plurality of nodes, so as to update the weights (Wa, Wb, Wc) during the learning mode, 其中,所述多个节点(130a、130b、…、130x)中的每个节点属于节点中的所述第一组和第二组(160,162)中的一个。Each of the plurality of nodes (130a, 130b, ..., 130x) belongs to one of the first and second groups (160, 162) of nodes. 21.根据权利要求19所述的计算机实现或硬件实现的方法,还包括:21. The computer-implemented or hardware-implemented method of claim 19, further comprising: 重复(380)接收(310)一个或多个系统输入、提供(320)多个输入、接收(330)输出、提供(340)系统输出、激励(350)、抑制(360)和更新(370)的步骤,直到满足学习标准。The steps of receiving (310) one or more system inputs, providing (320) multiple inputs, receiving (330) outputs, providing (340) system outputs, stimulating (350), inhibiting (360), and updating (370) are repeated (380) until a learning criterion is met. 22.根据权利要求20或21所述的计算机实现或硬件实现的方法,还包括:22. The computer-implemented or hardware-implemented method of claim 20 or 21, further comprising: 重复(390)接收(310)一个或多个系统输入、提供(320)多个输入、接收(330)输出、提供(340)系统输出、激励(350)和抑制(360)的步骤,直到满足停止标准。The steps of receiving (310) one or more system inputs, providing (320) a plurality of inputs, receiving (330) an output, providing (340) a system output, stimulating (350), and inhibiting (360) are repeated (390) until a stopping criterion is met. 23.根据权利要求20至22中任一项所述的方法,还包括:23. The method according to any one of claims 20 to 22, further comprising: 通过将权重(Wa、…、Wy)设置为零来初始化(304)权重(Wa、…、Wy);以及Initializing (304) weights (Wa, ..., Wy) by setting them to zero; and 在第三时间段的持续时间内将预定波形添加(308)到所述多个节点(130a、130b、…、130x)中的一个或多个节点的所述输出(134a、134b、…、134x),所述第三时间段开始于接收(310)包括待处理数据的一个或多个系统输入(110a、110b、…、110z)的时间。A predetermined waveform is added (308) to the output (134a, 134b, ..., 134x) of one or more of the plurality of nodes (130a, 130b, ..., 130x) for the duration of a third time period, the third time period beginning at the time when one or more system inputs (110a, 110b, ..., 110z) including data to be processed are received (310). 24.根据权利要求20至22中任一项所述的方法,还包括:24. The method according to any one of claims 20 to 22, further comprising: 通过向所述权重(Wa、…、Wy)随机地分配位于0和1之间的数值来初始化(306)所述权重(Wa、…、Wy);以及Initializing (306) the weights (Wa, ..., Wy) by randomly assigning them values between 0 and 1; and 在第三时间段的持续时间内将预定波形添加(308)到所述多个节点(130a、130b、…、130x)中的一个或多个节点的所述输出(134a、134b、…、134x)。A predetermined waveform is added (308) to the output (134a, 134b, ..., 134x) of one or more of the plurality of nodes (130a, 130b, ..., 130x) for a duration of a third time period. 25.一种包括指令的计算机程序产品,当所述指令在处理设备的至少一个处理器上执行时使所述处理设备执行根据权利要求20至24中任一项所述的方法。25. A computer program product comprising instructions which, when executed on at least one processor of a processing device, cause the processing device to perform the method according to any one of claims 20 to 24. 26.一种存储一个或多个程序的非临时性计算机可读存储介质,所述一个或多个程序被配置为由处理设备的一个或多个处理器执行,所述一个或多个程序包括指令,当所述指令被所述处理设备执行时使所述处理设备执行根据权利要求20-24中任一项所述的方法。26. A non-transitory computer-readable storage medium storing one or more programs, wherein the one or more programs are configured to be executed by one or more processors of a processing device, and the one or more programs include instructions that, when executed by the processing device, cause the processing device to perform the method according to any one of claims 20-24.
CN202380027102.XA 2022-02-23 2023-02-21 A data processing system, method and computer program product including a network Pending CN118871929A (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US202263313076P 2022-02-23 2022-02-23
US63/313,076 2022-02-23
SE2250397-3 2022-03-30
PCT/SE2023/050153 WO2023163637A1 (en) 2022-02-23 2023-02-21 A data processing system comprising a network, a method, and a computer program product

Publications (1)

Publication Number Publication Date
CN118871929A true CN118871929A (en) 2024-10-29

Family

ID=88018604

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202380027102.XA Pending CN118871929A (en) 2022-02-23 2023-02-21 A data processing system, method and computer program product including a network

Country Status (2)

Country Link
CN (1) CN118871929A (en)
SE (1) SE547197C2 (en)

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7904398B1 (en) * 2005-10-26 2011-03-08 Dominic John Repici Artificial synapse component using multiple distinct learning means with distinct predetermined learning acquisition times
US7958071B2 (en) * 2007-04-19 2011-06-07 Hewlett-Packard Development Company, L.P. Computational nodes and computational-node networks that include dynamical-nanodevice connections
US7814038B1 (en) * 2007-12-06 2010-10-12 Dominic John Repici Feedback-tolerant method and device producing weight-adjustment factors for pre-synaptic neurons in artificial neural networks
JP5393589B2 (en) * 2010-05-17 2014-01-22 本田技研工業株式会社 Electronic circuit
US20150278680A1 (en) * 2014-03-26 2015-10-01 Qualcomm Incorporated Training, recognition, and generation in a spiking deep belief network (dbn)
US10713558B2 (en) * 2016-12-30 2020-07-14 Intel Corporation Neural network with reconfigurable sparse connectivity and online learning

Also Published As

Publication number Publication date
SE547197C2 (en) 2025-05-27
SE2250397A1 (en) 2023-08-24

Similar Documents

Publication Publication Date Title
CN109754066B (en) Method and apparatus for generating a fixed-point neural network
CN110799994B (en) Adaptive bit width reduction for neural networks
CN112508186B (en) Method for training neural network for image recognition and neural network device
CN107622303B (en) Methods for neural networks and apparatus for performing the methods
CN110956256B (en) Method and device for realizing Bayes neural network by using memristor intrinsic noise
KR102410820B1 (en) Method and apparatus for recognizing based on neural network and for training the neural network
WO2020061489A1 (en) Training neural networks for vehicle re-identification
TWI806987B (en) Method for training of neural networks and neural network training system
CN115062769B (en) Model training method, device, equipment and storage medium based on knowledge distillation
Kuo et al. Automatic clustering using an improved particle swarm optimization
US20250165779A1 (en) A data processing system comprising a network, a method, and a computer program product
CN116108353A (en) A small-sample deep learning underwater acoustic target recognition method based on data grouping
KR20220166176A (en) Method and apparatus for quantizing deep neural network
CN118871929A (en) A data processing system, method and computer program product including a network
US20250148263A1 (en) Computer-implemented or hardware-implemented method of entity identification, a computer program product and an apparatus for entity identification
CN115699018A (en) Computer-implemented or hardware-implemented entity recognition method, computer program product and device for entity recognition
CN118884382A (en) A method for detecting small floating targets on the sea surface with controllable false alarm based on pulse neural network
CN114970732B (en) Posterior calibration method, device, computer equipment and medium for classification model
CN114822562B (en) Voiceprint recognition model training method, voiceprint recognition method and related equipment
US20240385987A1 (en) A computer-implemented or hardware-implemented method, a computer program product, an apparatus, a transfer function unit and a system for identification or separation of entities
CN116542139A (en) Method and device for predicting surface roughness of liquid jet polishing
CN115761343A (en) Image classification method and image classification device based on continuous learning
Purbanugraha et al. Improvement Accuracy Identification and Learning Speed of Offline Signatures Based on SqueezeNet with ADAM Backpropagation
EP4369258A1 (en) Systems and methods for finding nearest neighbors
Li et al. A 512-nW 0.003-mm2 Forward-Forward Black Box Trainer for an Analog Voice Activity Detector in 28-nm CMOS

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination