EP3769270A1 - A method, an apparatus and a computer program product for an interpretable neural network representation - Google Patents
A method, an apparatus and a computer program product for an interpretable neural network representationInfo
- Publication number
- EP3769270A1 EP3769270A1 EP19771931.3A EP19771931A EP3769270A1 EP 3769270 A1 EP3769270 A1 EP 3769270A1 EP 19771931 A EP19771931 A EP 19771931A EP 3769270 A1 EP3769270 A1 EP 3769270A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- interpretability
- neural network
- measure
- inference model
- model
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H50/00—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics
- G16H50/20—ICT specially adapted for medical diagnosis, medical simulation or medical data mining; ICT specially adapted for detecting, monitoring or modelling epidemics or pandemics for computer-aided diagnosis, e.g. based on medical expert systems
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
- G06N20/20—Ensemble learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/042—Knowledge-based neural networks; Logical representations of neural networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
- G06N3/0442—Recurrent networks, e.g. Hopfield networks characterised by memory or gating, e.g. long short-term memory [LSTM] or gated recurrent units [GRU]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0464—Convolutional networks [CNN, ConvNet]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/0895—Weakly supervised learning, e.g. semi-supervised or self-supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/09—Supervised learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/092—Reinforcement learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/02—Knowledge representation; Symbolic representation
- G06N5/022—Knowledge engineering; Knowledge acquisition
- G06N5/025—Extracting rules from data
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/04—Inference or reasoning models
- G06N5/045—Explanation of inference; Explainable artificial intelligence [XAI]; Interpretable artificial intelligence
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/04—Inference or reasoning models
- G06N5/048—Fuzzy inferencing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Definitions
- the present solution generally relates to machine learning and data analytics.
- the solution relates to determining and signaling a level of interpretability of neural network representation.
- Semantic information is represented by metadata which may express the type of scene, the occurrence of a specific action/activity, the presence of a specific object, etc. Such semantic information can be obtained by analyzing the media.
- neural networks have been criticized for their black-box nature. Such kind of black-box decision making is unacceptable in many use cases, such as in medical diagnosis or autonomous driving in which even rare mistakes can be costly or fatal.
- a method comprising obtaining a neural network; distilling the neural network into at least one inference model; based on the inference model, determining an interpretability measure, said interpretability measure indicating a level of interpretability of the inference model; and outputting the interpretability measure.
- an apparatus comprising means for obtaining a neural network; means for distilling the neural network into at least one inference model; based on the inference model, means for determining an interpretability measure, said interpretability measure indicating a level of interpretability of the inference model; and means for outputting the interpretability measure.
- apparatus comprising at least one processor, memory including computer program code, the memory and the computer program code configured to, with the at least one processor, cause the apparatus to perform at least the following obtain a neural network; distil the neural network into at least one inference model; based on the inference model, to determine an interpretability measure, said interpretability measure indicating a level of interpretability of the inference model; and output the interpretability measure.
- a computer program product embodied on a non-transitory computer readable medium, comprising computer program code configured to, when executed on at least one processor, cause an apparatus or a system to obtain a neural network; distil the neural network into at least one inference model; based on the inference model, to determine an interpretability measure, said interpretability measure indicating a level of interpretability of the inference model; and output the interpretability measure.
- the input data is an image.
- the apparatus is used for medical diagnosis.
- said at least one distilled model comprises one or more decision trees extracted from the neural network. According to an embodiment, said distilling is performed by using fuzzy logic.
- an interpretability measure is determined for a distilling method used for distilling the neural network.
- level of interpretability is determined based on at least one of: a number of tests associated with the one or more decision trees; at least one Hamming distance between decisions of the decision trees; an average fuzziness of decisions of the one or more decision trees.
- Fig. 1 shows a computer system according to an embodiment suitable to be used in data processing
- Fig. 2 shows an example of a Convolutional Neural Network (CNN);
- Fig. 3 shows an example of a Recurrent Neural Network (RNN);
- Figs. 4a-d show examples of 64 channels of merged neuron weights of a
- Fig. 5 shows an example of an encapsulation of interpretable neural network representation
- Fig. 6 illustrates a performance-interpretability trade-off curve with respect to that of an ideal model
- Fig. 7 is a flowchart illustrating a method according to an embodiment.
- Figure 1 shows a computer system suitable to be used in data processing, for example in machine learning according to an embodiment.
- the generalized structure of the computer system will be explained in accordance with the functional blocks of the system.
- Several functionalities can be carried out with a single physical device, e.g. all calculation procedures can be performed in a single processor if desired.
- a data processing system of an apparatus according to an example of Fig. 1 comprises a main processing unit 100, a memory 102, a storage device 104, an input device 106, an output device 108, and a graphics subsystem 1 10, which are all connected to each other via a data bus 1 12.
- the main processing unit 100 is a conventional processing unit arranged to process data within the data processing system.
- the main processing unit 100 may comprise or be implemented as one or more processors or processor circuitry.
- the memory 102, the storage device 104, the input device 106, and the output device 108 may include conventional components as recognized by those skilled in the art.
- the memory 102 and storage device 104 store data in the data processing system 100.
- Computer program code resides in the memory 102 for implementing, for example, machine learning process.
- the input device 106 inputs data into the system while the output device 108 receives data from the data processing system and forwards the data, for example to a display.
- the data bus 1 12 is a conventional data bus and while shown as a single line it may be any combination of the following: a processor bus, a PCI bus, a graphical bus, an ISA bus. Accordingly, a skilled person readily recognizes that the apparatus may be any data processing device, such as a computer device, a personal computer, a server computer, a mobile phone, a smart phone or an Internet access device, for example Internet tablet computer.
- various processes of the computer system may be carried out in one or more processing devices; for example, entirely in one computer device, or in one server device or across multiple user devices.
- the elements of machine learning process may be implemented as a software component residing on one device or distributed across several devices, as mentioned above, for example so that the devices form a so-called cloud.
- Deep learning is a sub-field of machine learning which has emerged in the recent years. Deep learning may involve learning of multiple layers of nonlinear processing units, either in supervised or in unsupervised manner, or in semi-supervised manner. These layers form a hierarchy of layers. Each learned layer extracts feature representations from the input data. Features from lower layers represent low-level semantics (i.e. less abstract concepts, such as edges and texture), whereas higher layers represent higher-level semantics (i.e., more abstract concepts, like scene class). Unsupervised learning applications typically include pattern analysis and representation (i.e., feature) learning, whereas supervised learning applications may include classification of image objects (in the case of visual data).
- Deep learning techniques may be used e.g. for recognizing and detecting objects in images or videos with great accuracy, outperforming previous methods.
- the fundamental difference of deep learning image recognition technique compared to previous methods is learning to recognize image objects directly from the raw data, whereas previous techniques are based on recognizing the image objects from hand- engineered features (e.g. SIFT features).
- CNN Convolutional Neural Network
- the input to a CNN is an image, but any other data could be used as well.
- Each layer of a CNN represents a certain abstraction (or semantic) level, and the CNN extracts multiple feature maps.
- a feature map may for example comprise a dense matrix of Real numbers representing values of the extracted features.
- the CNN in Fig. 2 has only three feature (or abstraction, or semantic) layers C1 , C2, C3 for the sake of simplicity, but CNNs may have more than three, and even over convolution layers.
- the first convolution layer C1 of the CNN consists of extracting 4 feature-maps from the first layer (i.e. from the input image). These maps may represent low-level features found in the input image, such as edges and corners.
- the second convolution layer C2 of the CNN consisting of extracting 6 feature-maps from the previous layer, increases the semantic level of extracted features.
- the third convolution layer C3 may represent more abstract concepts found in images, such as combinations of edges and corners, shapes, etc.
- the last layer of the CNN referred to as fully connected Multi-Layer Perceptron (MLP) may include one or more fully-connected (i.e., dense) layers and a final classification layer.
- the MLP uses the feature-maps from the last convolution layer in order to predict (recognize) for example the object class. For example, it may predict that the object in the image is a house.
- Deep learning is a field, which studies artificial neural networks (ANN), also referred to as neural network (NN).
- ANN artificial neural networks
- N neural network
- a neural network is a computation graph representation, usually made of several layers of successive computation. Each layer is made of units or neurons computing an elemental/basic computation.
- the goal of a neural network is to transform the input data into a more useful output.
- classification where input data is classified into one of N possible classes (e.g., classifying if an image contains a cat or a dog).
- regression where input data is transformed into a Real number (e.g. determining the music beat of a song).
- the power of neural networks comes from the internal representation which is built inside the layers. This representation is distributed among many units and is hierarchical, where complex concepts build on top of simple concepts.
- a neural network has two main modes of operation: training phase and testing phase.
- the training phase is the development phase, where the network learns to perform the final task. Learning consists in iteratively updating the weights or connections between units.
- the testing phase is the phase in which the network actually performs the task. Learning can be performed in several ways. The main ones are supervised, unsupervised, and reinforcement learning.
- supervised training the model is provided with input-output pairs, where the output is usually a label.
- the network is provided only with input data (and also with output raw data in case of self-supervised training).
- reinforcement learning the supervision is more sparse and less precise; instead of input-output pairs, the network gets input data and, sometimes, delayed rewards in the form of scores (E.g., -1 , 0, or +1 ).
- FIG. 1 Another example of neural network is a Recurrent Neural Network (RNN), where the hidden representation (hidden state h) is updated based not only on the current input but also on the hidden representations obtained from past inputs.
- RNNs work by recurrently (iteratively) looking at the input at each time step t and building an internal representation of the whole sequence so far.
- This internal representation is a “summary” and can be thought of as a“memory”.
- One of the RNN types is Long Short- Term Memory (LSTM) network, which uses special gating mechanisms that help training RNNs more effectively.
- Figure 3 illustrates an encoder-decoder model with additional RNN in the form of LSTM 310 for modeling the temporal aspect in data. As shown in Figure 3, an input at time t is given to the system 300, and getting the data at time t+1 as an output.
- NNR neural network representation
- the present embodiments are targeted to an interpretable neural network encapsulation.
- the embodiments utilize a technique of distillation, where a simple student decision making model is learnt through a distillation from a complex teacher model.
- three examples of known distillation methods are given.
- a trained neural network is first used to produce the class probabilities, which are subsequently used as“soft targets” for training the soft decision tree (SDT) model.
- SDT models trained as such are able to achieve higher performances and generalize better, than decision tree models directly trained without exploiting the“soft targets” distilled from the original neural network model.
- the hierarchical decisions employed in a SDT model are much more easier to explain, as compared with the hierarchical features learned in the original neural network.
- a SDT model may comprise for example 15 inner nodes at 4 different depths.
- p,(x) o(x w, +bi) where x is the input pattern to the model and s is the sigmoid logistic function.
- deep models are optimized for human-simulatability via a new model complexity penalty function (“tree regularization”).
- the tree regularization favors models whose decision boundaries can be well-approximated by small decision- trees, thus penalizing models that requires many calculations to simulate predictions.
- the true-average-path-length cost function Q(W) is replaced by an estimate of the average-path-length H(W), which is differentiable with respect to neural network parameters on the one hand, and makes reliable estimation on the other hand. This implementation step turns out to be a key technical contribution of the method.
- a decision tree is learned to explain the logic of each prediction of a pre-trained convolutional neural networks (CNNs).
- CNNs convolutional neural networks
- the decision tree learns to identify, qualitatively, which object parts in high convolution-layers of the CNN contributes to the predication in question. By doing do, a parse tree is recursively inferenced in a top-down manner for a given test image. Moreover, the (relative) contributions of each decision tree node can be exactly computed using optimized weights parameters of selected filters).
- evaluation metrics for CNN model interpretability include the following methods: 1 ) the compatibility between a convolution filter f and a specific semantic concept (or parts) may be quantified by the intersection-over-union score. The higher the score, the more relevant the filter is to the semantic concept (thus more interpretable), and vice versa; 2) location stability may account for the interpretability of underlying CNN models. The higher the stability, the more interpretable the model is. Implementation-wise, the location stability can be measured by the average deviation of relative location of detected object parts w.r.t. some reference landmarks in test images.
- GNN Generalized Hamming Network
- each neuron outputs quantifies the degree of fuzzy equivalence between inputs X and neuron weights W.
- each neuron evaluates the fuzzy truth value of the statement“x ⁇ ® w”, where“ ⁇ ®” denotes a fuzzy equivalence relation.
- neighboring neuron outputs from the previous layer are integrated to form composite statements, e.g.“(x 1 i ⁇ ® w 1 i, . . .
- Figures 4a-4d illustrate 64 channels of merged neuron weights, up to layers 1 ,2, 3 and 4 respectively, of a Generalized Hamming Network, which is trained for CIFAR10 classification problem.
- a merged neural network filters up to 4 different layers.
- the merged weights are independent of input test images; and the merging of weights is purely analytic and no optimization of learning steps are needed.
- Each merged neuron weight represents different types of salient image features from small to large scale, including e.g. dots or corner points (Fig. 4a), oriented edgelets (Fig. 4b), textons at different scales (Fig. 4c-4d). Note that for color images, textons may have associated color information not shown in Figs. 4a-d.
- the present embodiments are targeted to an encapsulation of interpretable neural network representation, as illustrated in Figure 5.
- the encapsulation method comprises
- a top-level neural network model 510 allows the highest performance decision- making (e.g. the highest accuracy) among a set of neural networks.
- Such a NN model may be complex, and thus, its inference logic may not be easy to explain, i.e. its interpretability measure is low.
- sub-optimal model 520 which is distilled (or learned) from the model 510.
- Such a sub-optimal (i.e. distilled) model may be simpler and easier to explain, but nevertheless, has a sub- optimal performance as compared with the original model.
- the inference logic of the distilled model can be explained in terms of e.g.“IF-THEN” propositional logic, or fuzzy logic or combination of simple inference rules.
- a numeric interpretability measure can be concretely defined for each distilled model. Such interpretability measures can be used to provide quantitative comparison between different distilled models.
- a neural network 510 is trained for a given learning task, e.g. image-based medical diagnosis, or anomaly detection in autonomous driving scenario.
- a neural network 510 is often complex, e.g. consists of thousands of convolutional layers, and hard to explain.
- the encapsulation of neural network representations according to embodiments provide more explainable justification for decision making in following aspects:
- the original neural network 510 For the original neural network 510, a number of parameters can be used to quantify the interpretability. However, the more parameters are used in the model, the less interpretable the model is. Therefore, the original neural network 510 is distilled into at least one distilled model 520, 530 according any of the distillation methods discussed above. The distillation method is used to extract simple inference models 520, 530 from the optimal neural network models 510.
- the above mentioned distillation methods can be adopted to extract decision-trees from original neural network model 510.
- the decision made at a leaf-node of the tree can be re- casted at sequence of IF-THEN propositional logic, e.g.“if body-temperature >39°C AND if white_blood_cell level > 9.0 then ...”
- the interpretability of the decision-tree can be concretely defined as the inversely proportional to the number of tests involved, i.e. where /() can be an arbitrary monotonic function [0,1] ® [0,1] and ttests 3 1 is assumed.
- decision forest i.e. ensemble of decision trees
- the above-mentioned propositional logic inference and interpretability measure are also applicable.
- the degree of consensus between different decision tree leaves can be taken into account by: where 7 is the interpretability for tree I, and the consensus ⁇ s defined as the (average) consistency of decisions made by different trees for each data sample, e.g. by the averaged hamming distance between decision made by different trees.
- the distillation method being used may adopt fuzzy logic, the interpretability of such distilled model can be explained as a sequence of Fuzzy Logic predicates e.g.“if body-temperature > 39°C with degree 0.85 fuzzy_AND if white_blood_cell level >9.0 with degree 0.9 then ...”.
- the interpretability of fuzzy logic can then be defined in by the number of tests and the average fuzziness over all test predicates.
- the performance measure may be classification/recognition accuracy, regression accuracy, or even camera pose estimation accuracy.
- the proposed measure is generic and aims to evaluate each model in terms of its performance- interpretability trade-off curve. Often a distilled model sacrifices performance for better interpretability and vice-versa (see below the curve marked with“Model A” ).
- the area under the curve (denoted as SA) is thus a good summarization of the model’s performance-interpretability trade off.
- SA an ideal model whose performance and interpretability are both maximized (marked as the green model below) and the area in the green rectangle is denoted as S b .
- the ratio of SA / S b is thus between [0,1 ] and can be used to quantify the gap of performance- interpretability curve from an ideal model.
- Figure 6 illustrates a performance-interpretability trade-off curve with respect to that of an ideal model.
- corresponding descriptions include, but are not limited to, at least one of the following attributes:
- training parameters learning rate, #epoch, batch-size, fine-tuned from public models, etc.
- Model 520 is learned/distilled from model 510
- Model X is fine-tuned from model 520
- Model Y is quantized or binarized from model 520
- a neural network When a neural network is used to make decision for given tasks, e.g. medical diagnosis, users (for example a doctor or a patient in medical applications) may make a request about target performance and interpretability levels of the neural network. Neural network representations that fulfil the query criterion will be searched and returned. Users may also request a particular type of representation (e.g. decision tree) with the level of interpretabilities above certain thresholds.
- a particular type of representation e.g. decision tree
- remote parties may request to retrieve/download a neural network representation that fulfils certain criteria, e.g. with the model size below certain thresholds and interpretability above certain levels.
- requests may be used e.g. for mobile healthcare applications, with which people make medical diagnosis for family members or themselves.
- a client application or device may send a request to a local or remote server for obtaining one or more neural networks corresponding to one or more interpretability levels.
- the request may comprise an indication of type of interpretability measure and/or one or more criteria associated with an interpretability measure, for example a threshold for a level of interpretability and/or an indication that the level of interpretability is to be determined based on a number of tests associated with a decision tree.
- a server may respond by sending a neural network or a set encapsulated neural networks.
- the response communication may include signaling associated with the provided neural network(s), as discussed above.
- a server may determine that it does not possess a neural network fulfilling the requested interpretability criteria, and in response, initiate generation of one or more new distilled neural networks. Having a particular target interpretability level set by the request from the client, the server may adjust the distillation method in order to achieve the target interpretability. For example, the distillation process may be constrained to a lower amount of nodes or filters in a soft-decision tree or fuzziness of logic predicates may be limited.
- the server may communicate the new neural network along with the associated interpretability level to the client.
- Figure 7 is a flowchart illustrating a method according to an embodiment.
- a method comprises obtaining a neural network 701 ; distilling the neural network into at least one inference model 702; based on the inference model, determining an interpretability measure, said interpretability measure indicating a level of interpretability of the inference model 703; and outputting the interpretability measure 704.
- An apparatus comprises means for obtaining a neural network; means for distilling the neural network into at least one inference model; means for determining an interpretability measure based on the inference model, said interpretability measure indicating a level of interpretability of the inference model; and means for outputting the interpretability measure.
- the means comprises a processor, a memory, and a computer program code residing in the memory, wherein the processor may further comprise processor circuitry.
- the interpretable neural network encapsulation has high accuracy in decision-making for given tasks and provides explainable logic of inferences involved in the decision making.
- the encapsulation also allows numeric interpretability measures to be used for specific decision-making models.
- the various embodiments of the invention can be implemented with the help of computer program code that resides in a memory and causes the relevant apparatuses to carry out the invention.
- a device may comprise circuitry and electronics for handling, receiving and transmitting data, computer program code in a memory, and a processor that, when running the computer program code, causes the device to carry out the features of an embodiment.
- a network device like a server may comprise circuitry and electronics for handling, receiving and transmitting data, computer program code in a memory, and a processor that, when running the computer program code, causes the network device to carry out the features of an embodiment.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Software Systems (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Medical Informatics (AREA)
- Public Health (AREA)
- Pathology (AREA)
- Databases & Information Systems (AREA)
- Epidemiology (AREA)
- Primary Health Care (AREA)
- Automation & Control Theory (AREA)
- Fuzzy Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| FI20185265 | 2018-03-21 | ||
| PCT/FI2019/050209 WO2019180310A1 (en) | 2018-03-21 | 2019-03-12 | A method, an apparatus and a computer program product for an interpretable neural network representation |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| EP3769270A1 true EP3769270A1 (en) | 2021-01-27 |
| EP3769270A4 EP3769270A4 (en) | 2021-12-29 |
Family
ID=67988278
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| EP19771931.3A Pending EP3769270A4 (en) | 2018-03-21 | 2019-03-12 | A method, an apparatus and a computer program product for an interpretable neural network representation |
Country Status (2)
| Country | Link |
|---|---|
| EP (1) | EP3769270A4 (en) |
| WO (1) | WO2019180310A1 (en) |
Families Citing this family (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN113537666B (en) * | 2020-04-16 | 2024-05-03 | 马上消费金融股份有限公司 | Evaluation model training method, evaluation and business auditing method, device and equipment |
| US11977990B2 (en) | 2020-04-28 | 2024-05-07 | International Business Machines Corporation | Decision tree interface for neural networks |
| CN111723810B (en) * | 2020-05-11 | 2022-09-16 | 北京航空航天大学 | Interpretability method of scene recognition task model |
| CN113947178B (en) * | 2020-07-15 | 2025-06-17 | 华为技术有限公司 | An interpretability method based on model dynamic behavior and related device |
| WO2022046077A1 (en) * | 2020-08-28 | 2022-03-03 | Siemens Aktiengesellschaft | Incremental learning for anomaly detection and localization in images |
| CN112434790B (en) * | 2020-11-10 | 2024-03-29 | 西安理工大学 | Self-interpretation method for distinguishing part of black box problem of convolutional neural network |
| CN112580781B (en) * | 2020-12-14 | 2025-06-27 | 深圳前海微众银行股份有限公司 | Processing method, device, equipment and storage medium of deep learning model |
| EP4216110A1 (en) * | 2022-01-21 | 2023-07-26 | Fujitsu Limited | A quantization method to improve the fidelity of rule extraction algorithms for use with artificial neural networks |
| CN116384450B (en) * | 2023-04-21 | 2025-11-25 | 吉林大学 | Deep Convolutional Fuzzy Neural Networks for Medical Data and Their Training Methods |
| CN119705504B (en) * | 2025-03-03 | 2025-08-05 | 吉林大学 | An explainable autonomous driving decision-making method based on causal knowledge |
Family Cites Families (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6564198B1 (en) * | 2000-02-16 | 2003-05-13 | Hrl Laboratories, Llc | Fuzzy expert system for interpretable rule extraction from neural networks |
| US10339465B2 (en) * | 2014-06-30 | 2019-07-02 | Amazon Technologies, Inc. | Optimized decision tree based models |
| EP3291146A1 (en) * | 2016-09-05 | 2018-03-07 | Fujitsu Limited | Knowledge extraction from a convolutional neural network |
-
2019
- 2019-03-12 EP EP19771931.3A patent/EP3769270A4/en active Pending
- 2019-03-12 WO PCT/FI2019/050209 patent/WO2019180310A1/en not_active Ceased
Also Published As
| Publication number | Publication date |
|---|---|
| WO2019180310A1 (en) | 2019-09-26 |
| EP3769270A4 (en) | 2021-12-29 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| EP3769270A1 (en) | A method, an apparatus and a computer program product for an interpretable neural network representation | |
| US20230281298A1 (en) | Using multimodal model consistency to detect adversarial attacks | |
| Wang et al. | Inferring salient objects from human fixations | |
| Aakur et al. | A perceptual prediction framework for self supervised event segmentation | |
| US10891524B2 (en) | Method and an apparatus for evaluating generative machine learning model | |
| US10438091B2 (en) | Method and apparatus for recognizing image content | |
| US11803971B2 (en) | Generating improved panoptic segmented digital images based on panoptic segmentation neural networks that utilize exemplar unknown object classes | |
| US12164599B1 (en) | Multi-view image analysis using neural networks | |
| CN119474383B (en) | Computational method and system for unstructured text data | |
| Oppong et al. | A Novel Computer Vision Model for Medicinal Plant Identification Using Log‐Gabor Filters and Deep Learning Algorithms | |
| CN114358188A (en) | Feature extraction model processing method, feature extraction model processing device, sample retrieval method, sample retrieval device and computer equipment | |
| Xiao et al. | Using convolution control block for Chinese sentiment analysis | |
| CN114048862B (en) | Model interpretation method and device | |
| CN113591529A (en) | Action segmentation model processing method and device, computer equipment and storage medium | |
| Shehu et al. | Lateralized approach for robustness against attacks in emotion categorization from images | |
| CN114612246B (en) | Object set identification method, device, computer equipment and storage medium | |
| CN114328904B (en) | Content processing method, device, computer equipment and storage medium | |
| CN117556004A (en) | A knowledge question and answer method, device and storage medium based on food engineering | |
| Nida et al. | Spatial deep feature augmentation technique for FER using genetic algorithm | |
| Lim et al. | A scene image is nonmutually exclusive—a fuzzy qualitative scene understanding | |
| CN114898426A (en) | Synonym label aggregation method, device, equipment and storage medium | |
| CN119889649A (en) | Disease prediction method, device, equipment and medium based on implicit knowledge enhancement | |
| Sathya et al. | An adaptive fuzzy ensemble model for facial expression recognition using poplar optimization and CRNN | |
| CN116961969A (en) | Methods and apparatus for determining the security of network links | |
| Xu et al. | Learning to generalize aerial person re‐identification using the meta‐transfer method |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
| PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
| STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
| 17P | Request for examination filed |
Effective date: 20201021 |
|
| AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
| AX | Request for extension of the european patent |
Extension state: BA ME |
|
| DAV | Request for validation of the european patent (deleted) | ||
| DAX | Request for extension of the european patent (deleted) | ||
| REG | Reference to a national code |
Ref country code: DE Ref legal event code: R079 Free format text: PREVIOUS MAIN CLASS: G06N0005040000 Ipc: G06N0003080000 |
|
| A4 | Supplementary search report drawn up and despatched |
Effective date: 20211130 |
|
| RIC1 | Information provided on ipc code assigned before grant |
Ipc: G06N 3/04 20060101ALN20211124BHEP Ipc: G16H 50/20 20180101ALI20211124BHEP Ipc: G06N 20/20 20190101ALI20211124BHEP Ipc: G06N 5/04 20060101ALI20211124BHEP Ipc: G06N 5/02 20060101ALI20211124BHEP Ipc: G06N 3/08 20060101AFI20211124BHEP |
|
| STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
| 17Q | First examination report despatched |
Effective date: 20230508 |