Palagin et al., 2017 - Google Patents
Resource and energy optimization oriented development of FPGA-based adaptive logical networks for classification problemPalagin et al., 2017
- Document ID
- 6586467143704334271
- Author
- Palagin A
- Opanasenko V
- Kryvyi S
- Publication year
- Publication venue
- Green IT Engineering: Components, Networks and Systems Implementation
External Links
Snippet
New FPGA-based devices with a high degree of reconfigurability that allow dynamic to change architecture and structure are created. These devices improve the efficiency of reconfigurable devices and systems according to the optimal criteria of hardware resources …
- 230000003044 adaptive 0 title abstract description 13
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
- G06N3/0635—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means using analogue means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
- G06F17/5009—Computer-aided design using simulation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/11—Complex mathematical operations for solving equations, e.g. nonlinear equations, general mathematical optimization problems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F7/00—Methods or arrangements for processing data by operating upon the order or content of the data handled
- G06F7/38—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation
- G06F7/48—Methods or arrangements for performing computations using exclusively denominational number representation, e.g. using binary, ternary, decimal representation using non-contact-making devices, e.g. tube, solid state device; using unspecified devices
- G06F7/52—Multiplying; Dividing
- G06F7/523—Multiplying only
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
- G06F17/5045—Circuit design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/02—Knowledge representation
- G06N5/022—Knowledge engineering, knowledge acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/04—Inference methods or devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F15/00—Digital computers in general; Data processing equipment in general
- G06F15/76—Architectures of general purpose stored programme computers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computer systems based on specific mathematical models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F2217/00—Indexing scheme relating to computer aided design [CAD]
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Palagin et al. | Resource and energy optimization oriented development of FPGA-based adaptive logical networks for classification problem | |
| Taud et al. | Multilayer perceptron (MLP) | |
| Klidbary et al. | Digital hardware realization of a novel adaptive ink drop spread operator and its application in modeling and classification and on-chip training | |
| Kheyrinataj et al. | Fractional Chebyshev functional link neural network‐optimization method for solving delay fractional optimal control problems with Atangana‐Baleanu derivative | |
| Deresse et al. | A deep learning approach: physics-informed neural networks for solving a nonlinear telegraph equation with different boundary conditions | |
| Okano et al. | Permutation elementary cellular automata: analysis and application of simple examples | |
| Skibinsky-Gitlin et al. | Cyclic reservoir computing with FPGA devices for efficient channel equalization | |
| Bodiwala et al. | An efficient stochastic computing based deep neural network accelerator with optimized activation functions | |
| Basterretxea | Recursive sigmoidal neurons for adaptive accuracy neural network implementations | |
| Chen et al. | Uniform projection nested Latin hypercube designs | |
| Rothmann et al. | STANN–Synthesis Templates for Artificial Neural Network Inference and Training | |
| Joseph et al. | Performance-driven LSTM accelerator hardware using split-matrix-based MVM | |
| Liu et al. | Hardware Spiking Neural Networks with Pair-Based STDP Using Stochastic Computing | |
| Bilski et al. | A new proposition of the activation function for significant improvement of neural networks performance | |
| Dudek | Are direct links necessary in random vector functional link networks for regression? | |
| Nepomnyashchiy et al. | Method of recurrent neural network hardware implementation | |
| Nejadmorad Moghanloo et al. | A new modified elman neural network with stable learning algorithms for identification of nonlinear systems | |
| Gorbounov et al. | Context-switching neural node for constrained-space hardware | |
| Tanaka et al. | Basics of neural networks | |
| Lozada et al. | Performance/resources comparison of hardware implementations on fully connected network inference | |
| Lee et al. | Reservoir Computing for Scalable Hardware with Block‐Based Neural Network | |
| Aizaz et al. | Efficient approximate multipliers for neural network applications | |
| Li et al. | Expensive Optimization | |
| Mayannavar et al. | Fast implementation of tunable ARN nodes | |
| Thasnimol et al. | A Hardware Accelerator Implementation of Multilayer Perceptron |