WO2010035690A2 - Dispositif, système et procédé d'entraînement de structure, programme et support d'enregistrement - Google Patents
Dispositif, système et procédé d'entraînement de structure, programme et support d'enregistrement Download PDFInfo
- Publication number
- WO2010035690A2 WO2010035690A2 PCT/JP2009/066259 JP2009066259W WO2010035690A2 WO 2010035690 A2 WO2010035690 A2 WO 2010035690A2 JP 2009066259 W JP2009066259 W JP 2009066259W WO 2010035690 A2 WO2010035690 A2 WO 2010035690A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- network structure
- learning
- network
- agent
- candidate
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
Definitions
- the present invention relates to a structure learning device, a structure learning system, a structure learning method, a program, and a recording medium that improve the network structure learning result of the agent by sharing and using the network structure information of other agents.
- Non-Patent Document 1 The learning method of the network structure learning system is described in Non-Patent Document 1, for example.
- the network structure learning system described in Non-Patent Document 1 includes network structure learning means and network structure mixed parameter estimation means.
- the network structure learning system having such a configuration operates as follows.
- the network structure learning and the network structure mixed parameter estimation are performed by a method called an EM algorithm until the result converges.
- Non-Patent Document 1 since the technique described in Non-Patent Document 1 does not consider sharing of the network structure by a plurality of agents, the plurality of agents share the network structure and cannot improve each learning result.
- the present invention has been made in view of the above problems, and an object thereof is to provide a structure learning apparatus in which a plurality of agents can share a network structure and improve each learning result.
- a structure learning device learns a network structure from a shared network and determines a candidate network structure, and a network structure made a candidate by the network structure learning means
- Network structure mixed parameter estimating means for estimating the mixed parameters of the network structure at the network structure level
- agent mixed parameter estimating means for estimating the network structure mixed parameters that are candidates by the network structure learning means at the agent level. It is characterized by.
- the network structure generating means for generating a new network structure when the network structure learning means determines that there is no suitable network structure as a candidate.
- the network structure learning means is characterized by repeatedly learning until the estimation results by the network structure mixed parameter estimating means and the agent mixed parameter estimating means converge.
- the structure learning system according to the present invention includes a plurality of the structure learning devices described above.
- the structure learning method is a network structure learning step for learning a network structure from a shared network and determining a network structure as a candidate, and a network structure mixed parameter determined by the network structure learning step as a network.
- a network structure mixed parameter estimating step for estimating at a structure level, and an agent mixed parameter estimating step for estimating a network structure mixed parameter made a candidate by the network structure learning step at an agent level.
- a network structure generation step is provided for generating a new network structure when it is determined in the network structure learning step that there is no network structure suitable as a candidate.
- the network structure learning step is characterized in that learning is repeated until the estimation results of the network structure mixed parameter estimation step and the agent mixed parameter estimation step converge.
- the program according to the present invention is characterized by causing a computer to execute the structure learning method described above.
- the recording medium in the present invention is a computer-readable recording medium storing the above-described program.
- a plurality of agents can share the network structure and improve each learning result.
- FIG. 1 is a hardware configuration diagram of a network structure learning apparatus according to an embodiment of the present invention.
- the apparatus includes an operation unit 1, a CPU 2, a memory 3, a ROM 4, and a network interface 5.
- the CPU 2 controls the entire apparatus, and the ROM 4 stores programs executed by the CPU 2 and other fixed data.
- the memory 3 is used for temporary storage of data when the CPU 2 performs calculations. Further, a network is realized by the network interface 5, and information can be input / output by the operation unit 1.
- FIG. 2 is a system block diagram of the network structure learning apparatus according to the embodiment of the present invention.
- the network structure learning apparatus includes an input unit 11, a network structure learning unit 12, a network structure generation unit 13, a network structure mixed parameter estimation unit 14, an agent mixed parameter estimation unit 15, and an output unit 16. Configured.
- the input unit 11 takes in the data to be observed, and the network structure learning unit 12 learns the network structure that generates the observation data.
- the network structure generation unit 13 generates a new network structure when the existing network structure does not match the observation value well by the network structure learning unit 12.
- the network structure mixed parameter estimation means 14 estimates the mixing probability at the network structure level
- the agent mixing parameter estimation means 15 estimates the mixing probability at the agent level.
- the output means 16 outputs the learned network structure and the estimation result in the mixing probability.
- step S1 data to be observed is captured by the input means 11 (step S1).
- step S2 the network structure learning unit 12 learns the most suitable structure as a network structure for generating observation data (step S2). It is determined whether or not there is (step S3).
- step S4 If there is no network structure that has been successfully matched by learning so far, a new most suitable network structure is generated from the distribution serving as the base of the network structure (step S4).
- the network structure level is calculated to calculate what mixing probability the network structure has, and the mixing probability is estimated (step S5). It is determined whether or not it has converged (step S6). Here, when it does not converge, it returns to step S2 and learns again.
- step S7 calculates what mixing probability the agent has at the agent level, estimates the mixing probability (step S7), and determines whether the estimation result has converged (step S8). .
- step S8 determines whether the estimation result has converged.
- step S9 output the network structure learned in step S2 and the mixing probability estimated in steps S5 and S7 (step S9).
- the probability distribution T of the network is
- Nj represents the number of observations of data at agent j.
- the observation value is assumed to be exchangeable in the agent, and is generated according to the probability distribution xji to T ( ⁇
- the network structure and its parameter ⁇ T: ji are generated from Gj,
- nDP nested Dirichlet process
- the index cj represents the mixed network structure that generates the network structure of agent j.
- the index sji represents the network structure for generating the data point xji.
- nDP and shared network structure can be performed based on Gibbs sampler, and the following steps are repeated until convergence.
- (ii) and (iv) correspond to the network structure mixed parameter estimating means 14 of FIG. 2 and step S5 of FIG. 3, and (i) and (iii) are the agent mixed parameters of FIG. 3 corresponds to the estimation means 15 and step S7 in FIG. 3, and (v) corresponds to the network structure learning means 12, the network structure generation means 13 in FIG. 2, and steps S2 to S4 in FIG.
- a plurality of agents can share a network structure and improve each learning result.
- a mixed model is used for each agent model, a plurality of agents can be clustered into similar ones.
- modeling, learning, and estimation are further performed using a nonparametric model, so that it is not necessary to set a finite value as the number of models mixed in advance.
- the present invention can be used as follows.
- the basic part correlation structure such as engine control shows similar behavior, but for example in the case of a hybrid vehicle, electronic motor control, etc. It has a parts correlation structure unique to the model.
- the learning accuracy is improved by sharing information on other vehicle types, and the part correlation structure that can be found only in a specific vehicle type is learned independently for each vehicle type.
- the present invention can be applied to such a case that vehicle types that share a part correlation structure with high uniqueness in the vehicle type are clustered as similar vehicle types.
- the learning accuracy can be improved by sharing information of other banks.
- each bank the number of specific operations, regional characteristics such as the occurrence of earthquakes and typhoons
- the present invention can also be applied to applications such as sharing information between banks having the specific loss event correlation structure and clustering them as banks having a similar loss event correlation structure.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Medical Informatics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Artificial Intelligence (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
L'invention concerne un dispositif d'entraînement de structure dans lequel une pluralité d'agents partagent une structure de réseau, et le résultat d'entraînement de chaque agent peut être amélioré. Le dispositif comprend : un moyen d'entraînement de structure de réseau destiné à enseigner une structure de réseau d'un réseau partagé et à déterminer une structure de réseau candidate, un moyen d'évaluation de paramètres mélangés au niveau structure de réseau destiné à évaluer les paramètres mélangés de la structure de réseau candidate du moyen d'entraînement de structure de réseau au niveau de la structure de réseau, et un moyen d'évaluation de paramètres mélangés au niveau agent destiné à évaluer des paramètres mélangés de la structure de réseau candidate du moyen d'entraînement de structures de réseau au niveau agent.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2010530824A JPWO2010035690A1 (ja) | 2008-09-24 | 2009-09-17 | 構造学習装置、構造学習システム、構造学習方法、プログラム及び記録媒体 |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2008-244168 | 2008-09-24 | ||
| JP2008244168 | 2008-09-24 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| WO2010035690A2 true WO2010035690A2 (fr) | 2010-04-01 |
| WO2010035690A9 WO2010035690A9 (fr) | 2011-02-03 |
Family
ID=42060202
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/JP2009/066259 Ceased WO2010035690A2 (fr) | 2008-09-24 | 2009-09-17 | Dispositif, système et procédé d'entraînement de structure, programme et support d'enregistrement |
Country Status (2)
| Country | Link |
|---|---|
| JP (1) | JPWO2010035690A1 (fr) |
| WO (1) | WO2010035690A2 (fr) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111602148A (zh) * | 2018-02-02 | 2020-08-28 | 谷歌有限责任公司 | 正则化神经网络架构搜索 |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN106156856A (zh) * | 2015-03-31 | 2016-11-23 | 日本电气株式会社 | 用于混合模型选择的方法和装置 |
-
2009
- 2009-09-17 JP JP2010530824A patent/JPWO2010035690A1/ja not_active Withdrawn
- 2009-09-17 WO PCT/JP2009/066259 patent/WO2010035690A2/fr not_active Ceased
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN111602148A (zh) * | 2018-02-02 | 2020-08-28 | 谷歌有限责任公司 | 正则化神经网络架构搜索 |
| US11669744B2 (en) | 2018-02-02 | 2023-06-06 | Google Llc | Regularized neural network architecture search |
| CN111602148B (zh) * | 2018-02-02 | 2024-04-02 | 谷歌有限责任公司 | 正则化神经网络架构搜索 |
| US12400121B2 (en) | 2018-02-02 | 2025-08-26 | Google Llc | Regularized neural network architecture search |
Also Published As
| Publication number | Publication date |
|---|---|
| JPWO2010035690A1 (ja) | 2012-02-23 |
| WO2010035690A9 (fr) | 2011-02-03 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN115943394B (zh) | 用于安全纵向联邦学习的方法、装置和系统 | |
| Berg et al. | Structure and evolution of protein interaction networks: a statistical model for link dynamics and gene duplications | |
| Krull et al. | Poseagent: Budget-constrained 6d object pose estimation via reinforcement learning | |
| CN113344016B (zh) | 深度迁移学习方法、装置、电子设备及存储介质 | |
| Dulac-Arnold et al. | Fast reinforcement learning with large action sets using error-correcting output codes for mdp factorization | |
| KR101700030B1 (ko) | 사전 정보를 이용한 영상 물체 탐색 방법 및 이를 수행하는 장치 | |
| Liu et al. | WTFM layer: An effective map extractor for unsupervised shape correspondence | |
| Zhu et al. | Trivialized momentum facilitates diffusion generative modeling on lie groups | |
| Langford et al. | Learning nonlinear dynamic models | |
| Zeng et al. | A nearsighted force-training approach to systematically generate training data for the machine learning of large atomic structures | |
| Fan et al. | Image steganalysis via random subspace fisher linear discriminant vector functional link network and feature mapping | |
| Ruan et al. | A robust and scalable solution for interpolative multidimensional scaling with weighting | |
| Meier et al. | Efficient Bayesian local model learning for control | |
| WO2010035690A2 (fr) | Dispositif, système et procédé d'entraînement de structure, programme et support d'enregistrement | |
| CN116070437A (zh) | 锂电池表面温度建模方法、装置、计算机设备和存储介质 | |
| Wong et al. | State-space decomposition for reinforcement learning | |
| Chen et al. | Handwritten CAPTCHA recognizer: a text CAPTCHA breaking method based on style transfer network | |
| Clementi et al. | Folding Lennard‐Jones proteins by a contact potential | |
| Chiu et al. | Automatic complexity reduction in reinforcement learning | |
| Varshney et al. | Reduced order modeling and dynamic optimization of multiscale PDE/kMC process systems | |
| JP2010134926A (ja) | マルチ・モーダル・グリーン・コンピューティング統合システム、コンピュータ・プログラム、方法、およびシステムの展開方法(問題解析を用いたマルチ・モーダル・グリーン・コンピューティング統合) | |
| CN119790405A (zh) | 基于模型的任务处理 | |
| WO2020040007A1 (fr) | Dispositif d'apprentissage, procédé d'apprentissage et programme d'apprentissage | |
| CN114255374B (zh) | 一种针对软间隔支持向量机的量子方法 | |
| Wang et al. | Generalized flow matching for transition dynamics modeling |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09816105 Country of ref document: EP Kind code of ref document: A2 |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2010530824 Country of ref document: JP |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| 122 | Ep: pct application non-entry in european phase |
Ref document number: 09816105 Country of ref document: EP Kind code of ref document: A2 |