WO2023161044A1 - A method to prevent capturing of an ai module and an ai system thereof - Google Patents
A method to prevent capturing of an ai module and an ai system thereof Download PDFInfo
- Publication number
- WO2023161044A1 WO2023161044A1 PCT/EP2023/053355 EP2023053355W WO2023161044A1 WO 2023161044 A1 WO2023161044 A1 WO 2023161044A1 EP 2023053355 W EP2023053355 W EP 2023053355W WO 2023161044 A1 WO2023161044 A1 WO 2023161044A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- module
- frequency domain
- output
- domain transformation
- submodule
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/10—Protecting distributed programs or content, e.g. vending or licensing of copyrighted material ; Digital rights management [DRM]
- G06F21/12—Protecting executable software
- G06F21/14—Protecting executable software against software analysis or reverse engineering, e.g. by obfuscation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/50—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
- G06F21/55—Detecting local intrusion or implementing counter-measures
- G06F21/554—Detecting local intrusion or implementing counter-measures involving event detection and direct action
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2221/00—Indexing scheme relating to security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F2221/03—Indexing scheme relating to G06F21/50, monitoring users, programs or devices to maintain the integrity of platforms
- G06F2221/034—Test or assess a computer or a system
Definitions
- the present disclosure relates to a method to prevent capturing of an Al module and an Al system thereof.
- Al based systems receive large amounts of data and process the data to train Al models. Trained Al models generate output based on the use cases requested by the user.
- Al systems are used in the fields of computer vision, speech recognition, natural language processing, audio recognition, healthcare, autonomous driving, manufacturing, robotics etc. where they process data to generate required output based on certain rules/intelligence acquired through training.
- the Al systems use various models/algorithms which are trained using the training data. Once the Al system is trained using the training data, the Al systems use the models to analyze the real time data and generate appropriate result. The models may be fine-tuned in real-time based on the results. The models in the Al systems form the core of the system. Lots of effort, resources (tangible and intangible), and knowledge goes into developing these models.
- Figure 1 depicts an Al system (10);
- Figure 2 depicts a submodule with the Al system (10);
- Figure 3 illustrates method steps of (200) of training a submodule (14) in an
- Figure 4 illustrates method steps (300) to prevent capturing of an Al module (12) in the Al system (10).
- Al artificial intelligence
- Al artificial intelligence
- Al artificial intelligence
- Al module may include many components.
- An Al module with reference to this disclosure can be explained as a component which runs a model.
- a model can be defined as reference or an inference set of data, which is use different forms of correlation matrices. Using these models and the data from these models, correlations can be established between different types of data to arrive at some logical understanding of the data.
- a person skilled in the art would be aware of the different types of Al models such as linear regression, naive bayes classifier, support vector machine, neural networks and the like.
- Some of the typical tasks performed by Al systems are classification, clustering, regression etc.
- Majority of classification tasks depend upon labeled datasets; that is, the data sets are labelled manually in order for a neural network to learn the correlation between labels and data. This is known as supervised learning.
- Some of the typical applications of classifications are: face recognition, object identification, gesture recognition, voice recognition etc.
- Clustering or grouping is the detection of similarities in the inputs. The cluster learning techniques do not require labels to detect similarities. Learning without labels is called unsupervised learning.
- Unlabeled data is the majority of data in the world. One law of machine learning is: the more data an algorithm can train on, the more accurate it will be. Therefore, unsupervised learning models/algorithms has the potential to produce accurate models as training dataset size grows.
- the module needs to be protected against attacks. Attackers attempt to attack the model within the Al module and steal information from the Al module.
- the attack is initiated through an attack vector.
- a vector may be defined as a method in which a malicious code/virus data uses to propagate itself such as to infect a computer, a computer system or a computer network.
- an attack vector is defined a path or means by which a hacker can gain access to a computer or a network in order to deliver a payload or a malicious outcome.
- a model stealing attack uses a kind of attack vector that can make a digital twin/replica/copy of an Al module.
- the attacker typically generates random queries of the size and shape of the input specifications and starts querying the model with these arbitrary queries. This querying produces input-output pairs for random queries and generates a secondary dataset that is inferred from the pre-trained model. The attacker then take this I/O pairs and trains the new model from scratch using this secondary dataset.
- This black box model attack vector where no prior knowledge of original model is required. As the prior information regarding model is available and increasing, attacker moves towards more intelligent attacks. The attacker chooses relevant dataset at his disposal to extract model more efficiently. This is domain intelligence model-based attack vector. With these approaches, it is possible to demonstrate model stealing attack across different models and datasets.
- FIG. 1 depicts an Al system (10).
- the Al system (10) comprises an input interface (11 ), a blocker module (18), an Al module (12), a submodule (14), a blocker notification module (20), an information gain module (16) and at least an output interface (22).
- the input interface (11) receives input data from at least one user.
- the input interface (11) is a hardware interface wherein a user can enter his query for the Al module (12).
- a module with respect to this disclosure can either be a logic circuitry or a software programs that respond to and processes logical instructions to get a meaningful result.
- a module is implemented as any or a combination of: one or more microchips or integrated circuits interconnected using a parent board, hardwired logic, software stored by a memory device and executed by a microprocessor, microcontrollers, firmware, an application specific integrated circuit (ASIC), and/or a field programmable gate array (FPGA).
- these various modules can either be a software embedded in a single chip or a combination of software and hardware where each module and its functionality is executed by separate independent chips connected to each other to function as the system.
- a neural network in an embodiment the Al module
- Such neural network chips are specialized silicon chips, which incorporate Al technology and are used for machine learning.
- the blocker module (18) is configured to block a user when the information gain exceeds a predefined threshold. Information gain is calculated based on input attack queries exceeds a predefined threshold value. The blocker module (18) is further configured to modify a first output generated by an Al module (12). This is done only when the input is identified as an attack vector. [0019] The Al module (12) to process said input data and generate the first output data corresponding to said input. The Al module (12) executes a first model (M) based on the input to generate a first output.
- the first model could be any one from those mentioned above such as linear regression, naive bayes classifier, support vector machine or neural networks and the like.
- the submodule (14) is configured to identify an attack vector from the received input.
- the submodule comprises a computation module (141), a memory (142) and at least a comparator module (143).
- the computation module (141) is configured to at least derive an instantaneous frequency domain transformation signature of the received input.
- the memory (142) is configured to store a set of pre-derived frequency domain transformation signatures.
- the set of pre-derived Frequency domain transformation signatures comprise Frequency domain transformation signatures for known inputs comprising a range of non-attack vectors.
- the comparator module (143) is configured to compare the instantaneous Frequency domain transformation signature with the set of pre-derived frequency domain transformation signatures.
- the comparator module (143) can be a conventional electronic comparator or specialized electronic comparator either embedded with neural networks or executing another Al model to enhance their functions.
- the above-mentioned components of the submodule can either be implemented in a single chip or as any or a combination of: one or more microchips or integrated circuits interconnected using a parent board, hardwired logic, software stored by a memory device and executed by a microprocessor, firmware, an application specific integrated circuit (ASIC), and/or a field programmable gate array (FPGA).
- the blocker notification module (20) transmits a notification to the owner of said Al system (10) on detecting an attack vector.
- the notification could be transmitted in any audio/visual/textual form.
- the information gain module (16) is configured to calculate an information gain and send the information gain value to the blocker module (18). The information gain is calculated using the information gain methodology. In one embodiment, if the information gain extracted exceeds a pre-defined threshold, the Al system (10) is configured to lock out the user from the system. The locking out the system is initiated if the cumulative information gain extracted by plurality of users exceeds a pre-defined threshold.
- the output interface (22) sends output to said at least one user.
- the output sent by the output interface (22) comprises the first output data when the submodule (14) doesn’t identify an attack vector from the received input.
- the output sent by the output interface (22) comprises a modified output received from the blocker module (18), when an attack vector is detected from the input.
- FIG 3 illustrates method steps of (200) of training a submodule (14) in an Al system (10).
- said Al system (10) comprises at least an Al module (12), a dataset used to train the Al module (12).
- a frequency domain transformation on the dataset is computed to derive a set of pre-derived Frequency domain transformation signatures.
- the dataset here comprises a range of valid inputs which are not attack vectors.
- one of the frequency domain transformations that is computed is the Fast Fourier Transform (FFT).
- FFT Fast Fourier Transform
- the set of pre-derived Frequency domain transformation signatures is stored in a memory of the submodule.
- the idea is that we can transform inputs in to frequency domains using FFT and then create aggregation of FFT (using various methods such as averaging method, other aggregation measures like median, inter quartile range, weighted average and the like) to represent all frequency components and its variations across the range of valid inputs.
- Figure 4 illustrates method steps (300) to prevent capturing of an Al module (12) in an Al system (10).
- the Al system (10) and its components have been explained in the preceding paragraphs by means of figures 1 and 3.
- a person skilled in the art will understand that the submodule (14) trained by the method steps (200) is now used in real time for preventing capture of an Al module (12) in an Al system (10).
- input interface (11) receives input data from at least one user.
- this input data is transmitted through a blocker module (18) to an Al module (12).
- the Al module (12) computes a first output based on the input data.
- step 304 input is processed by submodule (14) to identify an attack vector from the input data, the identification information of the attack vector is sent to the information gain module (16).
- Processing of the input data further comprises computing an instantaneous Frequency domain transformation signature of the received input by means of a computational module. This is followed by comparing the instantaneous Frequency domain transformation signature with a set of pre-derived Frequency domain transformation signatures by means of a comparator module (143). Finally, identifying an attack vector based on said comparison.
- the set of pre-derived Frequency domain transformation signatures comprise Frequency domain transformation signatures for known inputs comprising a range of non-attack vectors.
- FFT Fast Fourier Transform
- the FFT of training data is computed and aggregate in the memory of the submodule (14) as mentioned in the training method for the submodule.
- the trained submodule (14) calculates FFT for each queries (individually or in batch). Than we use the aggregated FFT signature and match FFT signature of queries. If the signature does not match with given threshold then the queries are rejected (individually or batch). This aggregation serves as a baseline to check if incoming input frequency signature is within the predefined aggregation limits. If input frequency signature generated using FFT is not within limits of aggregated frequency signature then it is considered as an attack vector.
- an output is sent to a user by means of the output interface (22).
- the output sent by the output interface (22) comprises the first output data when the submodule (14) doesn’t identify an attack vector from the received input.
- an information gain is calculated.
- the information gain is sent to the blocker module (18). In an embodiment, if the information gain exceeds a pre-defined threshold, the user is blocked, and the notification is sent the owner of the Al system (10) using blocker notification module (20). If the information gain is below a pre-defined threshold, although an attack vector was detected, the blocker module (18) may modify the first output generated by the Al module (12) to send it to the output interface (22).
- the user profile may be used to determine whether the user is habitual attacker or was it one time attack or was it only incidental attack etc.
- the steps for unlocking of the system may be determined. If it was first time attacker, the user may be locked out temporarily. If the attacker is habitual attacker, then a stricter locking steps may be suggested and so on.
- a person skilled in the art will appreciate that while these method steps describe only a series of steps to accomplish the objectives, these methodologies may be implemented with slight modification to the Al system (10) described herein. This idea to develop a method to prevent capturing of an Al module (12) and an Al system (10) thereof is quite useful for time series inputs where time series can be sliced and FFT or other frequency domain transformations can be generated.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Computer Security & Cryptography (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Medical Informatics (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Technology Law (AREA)
- Complex Calculations (AREA)
- Character Discrimination (AREA)
Abstract
Description
Claims
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| EP23704946.5A EP4483278A1 (en) | 2022-02-25 | 2023-02-10 | A method to prevent capturing of an ai module and an ai system thereof |
| US18/841,215 US20250165593A1 (en) | 2022-02-25 | 2023-02-10 | A Method to Prevent Capturing of an AI Module and an AI System Thereof |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| IN202241010164 | 2022-02-25 | ||
| IN202241010164 | 2022-02-25 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2023161044A1 true WO2023161044A1 (en) | 2023-08-31 |
Family
ID=85227247
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/EP2023/053355 Ceased WO2023161044A1 (en) | 2022-02-25 | 2023-02-10 | A method to prevent capturing of an ai module and an ai system thereof |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US20250165593A1 (en) |
| EP (1) | EP4483278A1 (en) |
| WO (1) | WO2023161044A1 (en) |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10657262B1 (en) * | 2014-09-28 | 2020-05-19 | Red Balloon Security, Inc. | Method and apparatus for securing embedded device firmware |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190095629A1 (en) | 2017-09-25 | 2019-03-28 | International Business Machines Corporation | Protecting Cognitive Systems from Model Stealing Attacks |
| WO2022029753A1 (en) * | 2020-08-06 | 2022-02-10 | Robert Bosch Gmbh | A method of training a submodule and preventing capture of an ai module |
Family Cites Families (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US11575700B2 (en) * | 2020-01-27 | 2023-02-07 | Xm Cyber Ltd. | Systems and methods for displaying an attack vector available to an attacker of a networked system |
| US11818147B2 (en) * | 2020-11-23 | 2023-11-14 | Fair Isaac Corporation | Overly optimistic data patterns and learned adversarial latent features |
| US20250274480A1 (en) * | 2024-02-26 | 2025-08-28 | Bank Of America Corporation | Intelligent Attack Vector Analysis and Mitigation System |
-
2023
- 2023-02-10 EP EP23704946.5A patent/EP4483278A1/en active Pending
- 2023-02-10 US US18/841,215 patent/US20250165593A1/en active Pending
- 2023-02-10 WO PCT/EP2023/053355 patent/WO2023161044A1/en not_active Ceased
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20190095629A1 (en) | 2017-09-25 | 2019-03-28 | International Business Machines Corporation | Protecting Cognitive Systems from Model Stealing Attacks |
| WO2022029753A1 (en) * | 2020-08-06 | 2022-02-10 | Robert Bosch Gmbh | A method of training a submodule and preventing capture of an ai module |
Non-Patent Citations (3)
| Title |
|---|
| AMIR MAHDI SADEGHZADEH ET AL: "Hardness of Samples Is All You Need: Protecting Deep Learning Models Using Hardness of Samples", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 21 June 2021 (2021-06-21), XP081993083 * |
| HARDER PAULA ET AL: "SpectralDefense: Detecting Adversarial Attacks on CNNs in the Fourier Domain", 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), IEEE, 18 July 2021 (2021-07-18), pages 1 - 8, XP033974557, DOI: 10.1109/IJCNN52387.2021.9533442 * |
| YI ZENG ET AL: "Rethinking the Backdoor Attacks' Triggers: A Frequency Perspective", ARXIV.ORG, CORNELL UNIVERSITY LIBRARY, 201 OLIN LIBRARY CORNELL UNIVERSITY ITHACA, NY 14853, 7 April 2021 (2021-04-07), XP091124282 * |
Also Published As
| Publication number | Publication date |
|---|---|
| US20250165593A1 (en) | 2025-05-22 |
| EP4483278A1 (en) | 2025-01-01 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20230306107A1 (en) | A Method of Training a Submodule and Preventing Capture of an AI Module | |
| US20210224688A1 (en) | Method of training a module and method of preventing capture of an ai module | |
| US20230289436A1 (en) | A Method of Training a Submodule and Preventing Capture of an AI Module | |
| US20250165593A1 (en) | A Method to Prevent Capturing of an AI Module and an AI System Thereof | |
| US20230376752A1 (en) | A Method of Training a Submodule and Preventing Capture of an AI Module | |
| US20230050484A1 (en) | Method of Training a Module and Method of Preventing Capture of an AI Module | |
| US20240386111A1 (en) | A Method of Training a Submodule and Preventing Capture of an AI Module | |
| EP4423648A1 (en) | A method of training a submodule and preventing capture of an ai module | |
| US20230267200A1 (en) | A Method of Training a Submodule and Preventing Capture of an AI Module | |
| EP4007979A1 (en) | A method to prevent capturing of models in an artificial intelligence based system | |
| US20240061932A1 (en) | A Method of Training a Submodule and Preventing Capture of an AI Module | |
| US12212682B2 (en) | Method of preventing capture of an AI module and an AI system thereof | |
| US20250272390A1 (en) | A Method to Prevent Exploitation of AI Module in an AI System | |
| US20250272423A1 (en) | A Method to Prevent Exploitation of an AI Module in an AI System | |
| US12032688B2 (en) | Method of training a module and method of preventing capture of an AI module | |
| EP4627485A1 (en) | A method to prevent exploitation of an ai module in an ai system | |
| US20250139241A1 (en) | A Method of Preventing Capture of an AI Module and an AI System Thereof | |
| WO2024160680A1 (en) | A method of assessing vulnerability of an ai model and a framework thereof | |
| EP4007978A1 (en) | A method to prevent capturing of models in an artificial intelligence based system |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 23704946 Country of ref document: EP Kind code of ref document: A1 |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 18841215 Country of ref document: US |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 2023704946 Country of ref document: EP |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| ENP | Entry into the national phase |
Ref document number: 2023704946 Country of ref document: EP Effective date: 20240925 |
|
| WWP | Wipo information: published in national office |
Ref document number: 18841215 Country of ref document: US |