US20160012225A1 - System and method for the detection of malware - Google Patents
System and method for the detection of malware Download PDFInfo
- Publication number
- US20160012225A1 US20160012225A1 US14/862,570 US201514862570A US2016012225A1 US 20160012225 A1 US20160012225 A1 US 20160012225A1 US 201514862570 A US201514862570 A US 201514862570A US 2016012225 A1 US2016012225 A1 US 2016012225A1
- Authority
- US
- United States
- Prior art keywords
- instruction sequence
- knowledge base
- expert system
- sequence
- threatening
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F21/00—Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
- G06F21/50—Monitoring users, programs or devices to maintain the integrity of platforms, e.g. of processors, firmware or operating systems
- G06F21/55—Detecting local intrusion or implementing counter-measures
- G06F21/56—Computer malware detection or handling, e.g. anti-virus arrangements
- G06F21/562—Static detection
- G06F21/563—Static detection by source code analysis
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/02—Knowledge representation; Symbolic representation
Definitions
- a binary file is often transferred between many computing devices.
- a computing device that receives a binary file is usually not aware of the origin of the file or whether the code that it receives is safe.
- a binary file can be disassembled to determine if the file contains malware such as viruses, worms, Trojan Horses and/or the like.
- a disassembler translates a binary file from machine language into assembly language.
- Some disassemblers are interactive and allow an expert programmer to make annotations, corrections, clarifications or decisions regarding how the disassembler analyzes a file. For example, a disassembler may signal when a new function or particular section of code appears. When an identified action occurs, a particular section of the code may be labeled for future reference.
- analysis of unknown executables can be a time consuming process that is usually performed manually by specially trained personnel, or automatically by the use of statistical methods.
- a method of automatically identifying malware may include receiving, by an expert system knowledge base, an assembly language sequence from a binary file, identifying an instruction sequence from the received assembly language sequence, and classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system knowledge base to the instruction sequence. If the instruction sequence is classified as threatening, information may be transmitted to a code analysis component and a user may be notified that the binary file includes malware. The information may include one or more of the following: the instruction sequence, a label comprising an indication that the instruction sequence is threatening, and a request that one or more other assembly language sequences from the binary file be searched for at least a portion of the instruction sequence.
- a method of automatically identifying malware may include receiving, by an expert system knowledge base, an assembly language sequence from a binary file, identifying an instruction sequence from the received assembly language sequence, and classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system knowledge base to the instruction sequence. If the instruction sequence is classified as non-threatening, information may be transmitted to a code analysis component and a second instruction sequence may be requested. The information may include one or more of the following: the instruction sequence and a label comprising an indication that the instruction sequence is non-threatening.
- a method of automatically identifying malware may include receiving, by an expert system knowledge base, an assembly language sequence from a binary file, identifying an instruction sequence from the received assembly language sequence, and classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system to the instruction sequence. If the instruction sequence is classified as non-classifiable, the method may include transmitting a request to a code analysis component that the assembly language sequence be reanalyzed, receiving a new instruction sequence corresponding to the reanalyzed assembly language sequence, and classifying the new instruction sequence as threatening, non-threatening or non-classifiable.
- a method of automatically identifying malware may include analyzing, by a code analysis component, a binary file to generate an assembly language sequence and a corresponding instruction sequence, transmitting the instruction sequence to an expert system knowledge base and receiving, from the expert system knowledge base, classification information associated with the instruction sequence. If the classification information identifies the instruction sequence as threatening, the method may include identifying one or more other assembly language sequences from the binary file that comprise at least a portion of the instruction sequence, and transmitting at least one of the identified assembly language sequences to the expert system knowledge base. If the classification information identifies the instruction sequence as non-threatening, the method may include transmitting a second instruction sequence to the expert system knowledge base. If the classification information identifies the instruction sequence as non-classifiable, the method may include reanalyzing the assembly language sequence to produce a new instruction sequence, and transmitting the new instruction sequence to the expert system knowledge base.
- a system for automatically identifying malware may include a code analysis component configured to identify an assembly language sequence including one or more instruction sequences from a binary file, and an expert system knowledge base in communication with the code analysis component.
- the expert system knowledge base may be configured to classify the instruction sequence as threatening, non-threatening or non-classifiable using one or more rules.
- FIG. 1 illustrates an exemplary malware detection system according to an embodiment.
- FIG. 2 illustrates an exemplary expert system knowledge base according to an embodiment.
- FIG. 3 illustrates a flowchart of an exemplary method for detecting and analyzing malware according to an embodiment.
- FIG. 4 illustrates a block diagram of an exemplary system that may be used to contain or implement program instructions according to an embodiment.
- FIGS. 5 and 6 illustrate exemplary instruction sequences according to an embodiment.
- node refers to a sequence of instructions within an assembly language sequence that is executed by a processor.
- An “assembly language” refers to a computer programming language that implements a symbolic representation of numeric machine codes.
- assembly language sequence refers to a sequence of nodes written in assembly language.
- a “binary file” refers to a computer file that includes data encoded in binary format.
- An executable file is a type of binary file.
- Malware is malicious software designed to disrupt, infiltrate or damage a computer system. Examples of malware include viruses, worms, trojan horses, adware, spyware, root kits and/or the like.
- An “expert system” is artificial intelligence software and/or firmware that is designed to simulate the decision making process of a human in a specific problem domain.
- FIG. 1 illustrates a malware detection system according to an embodiment.
- a malware detection system may include a code analysis component 100 , an expert system knowledge base 200 and/or a connector logic component 150 .
- the code analysis component 100 , expert system knowledge base 200 and/or connector logic component 150 may be implemented using software, hardware or a combination of software and hardware.
- the code analysis component 100 , expert system knowledge base 200 and/or connector logic component 150 may reside on the same computing device.
- the code analysis component 100 , expert system knowledge base 200 and/or connector logic component 150 may reside on different computing devices in communication with one another.
- a code analysis component 100 may analyze binary files such as, but not limited to, executables.
- a code analysis component 100 may statically or dynamically analyze binary files. Static analysis may include analyzing a binary file that is not currently being executed. In comparison, dynamic analysis may include analyzing a binary file while the binary file is being executed.
- a code analysis component may be implemented using software, hardware or a combination of software and hardware.
- a code analysis component 100 may include a disassembler, a debugger, a decompiler and/or the like.
- the code analysis component 100 may be a disassembler, such as IDA Pro.
- a code analysis component may analyze a binary file to create an assembly language sequence.
- the assembly language sequence may include a human-readable representation of the binary file.
- the code analysis component 100 may include internal rules and/or operations which may be used to create an assembly language sequence from the binary file.
- the code analysis component 100 may analyze the assembly language sequence to determine an instruction sequence.
- a code analysis component 100 may interact with external devices to analyze a binary file.
- the code analysis component 100 may communicate with an expert system knowledge base 200 .
- the malware detection system may include an expert system knowledge base 200 .
- an expert system knowledge base 200 may include a representation of a human's expertise in a particular area.
- an expert system knowledge base 200 may include information, data, rules and/or the like to model the knowledge and practices of an experienced computer analyst.
- the expert system knowledge base 200 may be implemented using the C Language Integrated Production System (“CLIPS”).
- CLIPS is a programming language and software tool that may be used to create expert systems.
- FIG. 2 illustrates an expert system knowledge base according to an embodiment.
- the expert system knowledge base 200 may include internal rules and/or operations. In an embodiment, these internal rules and/or operations may be applied to an instruction sequence from an assembly language sequence to determine whether the assembly language sequence contains malware. In an embodiment, the internal rules and/or operations may represent the encoding of human expertise.
- a domain expert 205 may populate the expert system knowledge base 200 .
- a domain expert may be, but is not limited to, a human being who has expertise in analyzing malware.
- a domain expert 205 may be a computing device configured to provide the expert system knowledge base 200 with internal rules and/or operations that may represent the encoding of human expertise. For example, a computing device may automatically provide the expert system knowledge base 200 with updates, enhancements or the like for one or more internal rules and/or operations.
- the expert system knowledge base 200 may be populated with binary file structures 210 .
- a binary file structure may be a template that depicts one or more portions of a binary file and/or a sequence of the portions in a binary file.
- the Binary file structures 210 may be used to analyze whether a file structure is proper. For example, a binary file structure 210 may be analyzed to determine if the header on the file conforms to a protocol.
- the expert system knowledge base 200 may be populated with worm defining operations 215 .
- Worm defining operations 215 may identify instruction sequences which replicate an assembly language sequence.
- the expert system knowledge base 200 may be populated with Trojan Horse defining operations 220 .
- Trojan Horse defining operations 220 may identify instruction sequences in an assembly language sequence that are associated with one or more Trojan Horses.
- the expert system knowledge base 200 may be populated with virus defining operations 225 .
- Virus defining operations 225 may identify self-replicating instruction sequences in an assembly language sequence. Additional and/or alternative operations may be included in the expert system knowledge base 200 .
- the malware detection system may include a connector logic component 150 .
- a connector logic component 150 may enable communication between the code analysis component 100 and the expert system knowledge base 200 .
- the assembly language sequence sent from the code analysis component 100 may be in a format which cannot be directly processed by the expert system knowledge base 200 .
- the code analysis component 100 may communicate the assembly language sequence to the connector logic component 150 .
- the connector logic component 150 may convert the instruction sequence into a format that the expert system knowledge base 200 can process.
- the connector logic component 150 may send the newly converted instruction sequence to the expert system knowledge base 200 .
- the connector logic component may obtain information from the expert system knowledge base 200 .
- the connector logic component may convert the information from the expert system knowledge base 200 into a format that is readable by the code analysis component 100 and transmit the converted information to the code analysis component.
- FIG. 3 depicts a flowchart of a method for detecting and analyzing malware according to an embodiment.
- a binary file may be received by the code analysis component.
- the code analysis component may analyze the file to obtain an assembly language sequence and an instruction sequence.
- the code analysis component may send the assembly language sequence with the instruction sequence to the expert system knowledge base via the connector logic component.
- the expert system knowledge base may receive 300 the assembly language sequence. In an embodiment, the expert system knowledge base may identify 305 the instruction sequence from the assembly language sequence.
- the expert system knowledge base may apply internal operations and/or rules to classify 315 the instruction sequence.
- the classification may be used to determine if the instruction sequence contains malware.
- the expert system knowledge base may classify the instruction sequence as non-threatening 315 , threatening 330 or non-classifiable 345 . Additional and/or alternate classifications may be used within the scope of this disclosure.
- the expert system knowledge base may traverse through the nodes and branches of a received instruction sequence using one or more internal rules and/or operations.
- the expert system knowledge base apply a group of precedential rules to the received instruction sequence.
- Each rule in the set of precedential rules may have a ranking with respect to the other precedential rules in the set.
- the rules may be ranked based on the number of matches between each rule and the instruction sequence. For example, the instruction sequences that are most similar to the match criteria of a rule may cause that rule to be given a highest priority for a given traversal. Alternatively, the instruction sequences that are least similar to the match criteria of a rule may cause that rule to be given a lowest priority for a given traversal.
- CLIPS provides conflict resolution strategies such as a complexity strategy and a simplicity strategy which give precedence to the most and least specific matches, respectively.
- such strategies may be employed to rank the rules as to those which most specifically match the instruction sequence.
- the expert system knowledge base may apply the rule associated with the highest precedence to the instruction sequence.
- one or more additional precedential rules from the group may be applied, in the order of their precedence, to the instruction sequence until the instruction sequence is classified or until all precedential rules have been applied.
- FIG. 5 illustrates an exemplary instruction sequence according to an embodiment. If the expert system knowledge base is able to traverse the entire instruction sequence 500 from start (Instruction 1 505 ) to finish (Instruction 8 510 ), then the instruction sequence 500 may be classified as non-threatening.
- the expert system knowledge base may transmit 320 information signifying that the instruction sequence is non-threatening to the code analysis component.
- the information may include a label attached to the instruction sequence indicating that the instruction sequence is non-threatening.
- the expert system knowledge base may request 325 a new assembly sequence with a new instruction sequence to analyze from the code analysis component.
- the expert system knowledge base may classify an instruction sequence as threatening 330 if the expert system knowledge base is unable to traverse each instruction of the instruction sequence.
- the expert system knowledge base may analyze the instruction sequence by traversing the instructions of the instruction sequence to determine if there is malware. For example, a loop may be an indicator of malware. If during the traversal, the expert system knowledge base arrives at an instruction that it already analyzed, the expert system knowledge base may determine that the instruction sequence forms a loop.
- the expert system knowledge base may classify an instruction sequence having one or more loops as threatening.
- FIG. 6 illustrates an exemplary instruction sequence according to an embodiment. As illustrated by FIG. 6 , the instruction sequence 600 may be classified as threatening because it includes a loop from Instruction 6 605 to Instruction 4 610 .
- other activities that may be indicative of malware or other nefarious behaviors may include encryption/decryption routines, replicating code, key logging, independent initiation of network communication, communication with known hostile or suspicious network hosts and/or the like.
- an instruction sequence that includes one or more of these activities may be classified as threatening. Additional and/or alternate activities may be indicative of malware within the scope of this disclosure.
- the expert system knowledge base may transmit 335 information signifying that the instruction sequence is threatening to the code analysis component.
- the information may be sent to the code analysis component via the connector logic component, which may translate the information into a form readable by the code analysis component.
- the information may include a label attached to the instruction sequence indicating that the instruction sequence is threatening.
- the information may include a request that the code analysis component search other assembly language sequences for at least a portion of an instruction sequence that was previously analyzed 340 .
- the code analysis component may search other assembly language sequences for the loop discussed in the previous example.
- the code analysis component may use its internal operations and/or rules to translate and/or analyze the information to determine whether at least a portion of an instruction sequence is present inside the assembly language sequences. If the code analysis component finds the same instruction sequence or portion thereof, the code analysis component may send the relevant assembly language sequence and instruction sequence to the expert system knowledge base.
- the expert system knowledge base may determine 345 whether an instruction sequence is non-classifiable.
- An instruction sequence may be identified as being non-classifiable if the expert system knowledge base is unable to determine whether the instruction sequence is threatening.
- a programmer who created a binary file may have intentionally used methods to obfuscate the workings of the file prevent the code analysis component from issuing the correct instruction sequence.
- the code analysis component may send an incomplete or nonsensical instruction sequence to the expert system knowledge base via the connector logic component.
- the expert system knowledge base may analyze each node of the instruction sequence using its internal rules and/or operations. Based on its analysis, the expert system knowledge base may transmit 350 a request to the code analysis component to reinterpret a particular node or series of nodes. For example, the expert system knowledge base may request that the code analysis component generate a new instruction sequence for a particular node.
- the request may include alternate considerations for the code analysis component in analyzing the assembly sequence.
- the code analysis component may not be able to properly analyze an assembly sequence.
- the expert system knowledge base may detect that an incorrect instruction sequence should be altered or ignored to allow the analysis to continue. In an embodiment, this information may be included in a request to the code analysis component.
- the code analysis component may use its internal rules and/or operations reanalyze the assembly language sequence and instruction sequence.
- the expert system knowledge base may receive 345 the reanalyzed assembly language sequence and new instruction sequence from the code analysis component via the connector logic component.
- the expert system knowledge base may traverse the new instruction sequence to determine its classification.
- FIG. 4 depicts a block diagram of an exemplary system that may be used to contain or implement program instructions according to an embodiment.
- a bus 400 serves as the main information highway interconnecting the other illustrated components of the hardware.
- CPU 405 is the central processing unit of the system, performing calculations and logic operations required to execute a program.
- Read only memory (ROM) 410 and random access memory (RAM) 415 constitute exemplary memory devices or storage media.
- a disk controller 420 interfaces with one or more optional disk drives to the system bus 400 .
- These disk drives may include, for example, external or internal DVD drives 425 , CD ROM drives 430 or hard drives 435 . As indicated previously, these various disk drives and disk controllers are optional devices.
- Program instructions may be stored in the ROM 410 and/or the RAM 415 .
- program instructions may be stored on a computer readable storage medium, such as a hard drive, a compact disk, a digital disk, a memory or any other tangible recording medium.
- An optional display interface 440 may permit information from the bus 400 to be displayed on the display 445 in audio, graphic or alphanumeric format. Communication with external devices may occur using various communication ports 450 .
- the hardware may also include an interface 455 which allows for receipt of data from input devices such as a keyboard 460 or other input device 465 such as a mouse, remote control, touch pad or screen, pointer and/or joystick.
- input devices such as a keyboard 460 or other input device 465 such as a mouse, remote control, touch pad or screen, pointer and/or joystick.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Computer Security & Cryptography (AREA)
- General Engineering & Computer Science (AREA)
- Software Systems (AREA)
- Computer Hardware Design (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Virology (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Mathematical Physics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Stored Programmes (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
- Machine Translation (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
A method of automatically identifying malware may include receiving, by an expert system knowledge base, an assembly language sequence from a binary file, identifying an instruction sequence from the received assembly language sequence, and classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system knowledge base to the instruction sequence. If the instruction sequence is classified as threatening, information may be transmitted to a code analysis component and a user may be notified that the binary file includes malware. The information may include one or more of the following: the instruction sequence, a label comprising an indication that the instruction sequence is threatening, and a request that one or more other assembly language sequences from the binary file be searched for at least a portion of the instruction sequence.
Description
- This application is a continuation of U.S. application Ser. No. 12/550,025 filed Aug. 28, 2009 which claims the benefit of the filing date of Provisional Patent Application No. 61/092,848 filed Aug. 29, 2008 both of which are incorporated by reference in their entirety.
- A binary file is often transferred between many computing devices. A computing device that receives a binary file is usually not aware of the origin of the file or whether the code that it receives is safe. To ensure the security of a computing device, a binary file can be disassembled to determine if the file contains malware such as viruses, worms, Trojan Horses and/or the like.
- Typically, a disassembler translates a binary file from machine language into assembly language. Some disassemblers are interactive and allow an expert programmer to make annotations, corrections, clarifications or decisions regarding how the disassembler analyzes a file. For example, a disassembler may signal when a new function or particular section of code appears. When an identified action occurs, a particular section of the code may be labeled for future reference. However, analysis of unknown executables can be a time consuming process that is usually performed manually by specially trained personnel, or automatically by the use of statistical methods.
- Before the present methods are described, it is to be understood that this invention is not limited to the particular systems, methodologies or protocols described, as these may vary. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to limit the scope of the present disclosure which will be limited only by the appended claims.
- It must be noted that as used herein and in the appended claims, the singular forms “a,” “an,” and “the” include plural reference unless the context clearly dictates otherwise. Unless defined otherwise, all technical and scientific terms used herein have the same meanings as commonly understood by one of ordinary skill in the art. As used herein, the term “comprising” means “including, but not limited to.”
- In an embodiment, a method of automatically identifying malware may include receiving, by an expert system knowledge base, an assembly language sequence from a binary file, identifying an instruction sequence from the received assembly language sequence, and classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system knowledge base to the instruction sequence. If the instruction sequence is classified as threatening, information may be transmitted to a code analysis component and a user may be notified that the binary file includes malware. The information may include one or more of the following: the instruction sequence, a label comprising an indication that the instruction sequence is threatening, and a request that one or more other assembly language sequences from the binary file be searched for at least a portion of the instruction sequence.
- In an embodiment, a method of automatically identifying malware may include receiving, by an expert system knowledge base, an assembly language sequence from a binary file, identifying an instruction sequence from the received assembly language sequence, and classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system knowledge base to the instruction sequence. If the instruction sequence is classified as non-threatening, information may be transmitted to a code analysis component and a second instruction sequence may be requested. The information may include one or more of the following: the instruction sequence and a label comprising an indication that the instruction sequence is non-threatening.
- In an embodiment, a method of automatically identifying malware may include receiving, by an expert system knowledge base, an assembly language sequence from a binary file, identifying an instruction sequence from the received assembly language sequence, and classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system to the instruction sequence. If the instruction sequence is classified as non-classifiable, the method may include transmitting a request to a code analysis component that the assembly language sequence be reanalyzed, receiving a new instruction sequence corresponding to the reanalyzed assembly language sequence, and classifying the new instruction sequence as threatening, non-threatening or non-classifiable.
- In an embodiment, a method of automatically identifying malware may include analyzing, by a code analysis component, a binary file to generate an assembly language sequence and a corresponding instruction sequence, transmitting the instruction sequence to an expert system knowledge base and receiving, from the expert system knowledge base, classification information associated with the instruction sequence. If the classification information identifies the instruction sequence as threatening, the method may include identifying one or more other assembly language sequences from the binary file that comprise at least a portion of the instruction sequence, and transmitting at least one of the identified assembly language sequences to the expert system knowledge base. If the classification information identifies the instruction sequence as non-threatening, the method may include transmitting a second instruction sequence to the expert system knowledge base. If the classification information identifies the instruction sequence as non-classifiable, the method may include reanalyzing the assembly language sequence to produce a new instruction sequence, and transmitting the new instruction sequence to the expert system knowledge base.
- In an embodiment, a system for automatically identifying malware may include a code analysis component configured to identify an assembly language sequence including one or more instruction sequences from a binary file, and an expert system knowledge base in communication with the code analysis component. The expert system knowledge base may be configured to classify the instruction sequence as threatening, non-threatening or non-classifiable using one or more rules.
- Aspects, features, benefits and advantages of the embodiments described herein will be apparent with regard to the following description, appended claims, and accompanying drawings where:
-
FIG. 1 illustrates an exemplary malware detection system according to an embodiment. -
FIG. 2 illustrates an exemplary expert system knowledge base according to an embodiment. -
FIG. 3 illustrates a flowchart of an exemplary method for detecting and analyzing malware according to an embodiment. -
FIG. 4 illustrates a block diagram of an exemplary system that may be used to contain or implement program instructions according to an embodiment. -
FIGS. 5 and 6 illustrate exemplary instruction sequences according to an embodiment. - Before the present methods and systems are described, it is to be understood that this invention is not limited to the particular systems, methodologies or protocols described, as these may vary. It is also to be understood that the terminology used herein is for the purpose of describing particular embodiments only, and is not intended to limit the scope of the present disclosure which will be limited only by the appended claims.
- As used herein and in the appended claims, the singular forms “a,” “an,” and “the” include the plural reference unless the context clearly dictates otherwise. Unless defined otherwise, all technical and scientific terms used herein have the same meanings as commonly understood by one of ordinary skill in the art. As used herein, the term “comprising” means “including, but not limited to.”
- For the purpose of the description below, a “node” refers to a sequence of instructions within an assembly language sequence that is executed by a processor.
- An “assembly language” refers to a computer programming language that implements a symbolic representation of numeric machine codes.
- An “assembly language sequence” refers to a sequence of nodes written in assembly language.
- A “binary file” refers to a computer file that includes data encoded in binary format. An executable file is a type of binary file.
- “Malware” is malicious software designed to disrupt, infiltrate or damage a computer system. Examples of malware include viruses, worms, trojan horses, adware, spyware, root kits and/or the like.
- An “expert system” is artificial intelligence software and/or firmware that is designed to simulate the decision making process of a human in a specific problem domain.
-
FIG. 1 illustrates a malware detection system according to an embodiment. A malware detection system may include acode analysis component 100, an expertsystem knowledge base 200 and/or aconnector logic component 150. In an embodiment, thecode analysis component 100, expertsystem knowledge base 200 and/orconnector logic component 150 may be implemented using software, hardware or a combination of software and hardware. In an embodiment, thecode analysis component 100, expertsystem knowledge base 200 and/orconnector logic component 150 may reside on the same computing device. Alternatively, thecode analysis component 100, expertsystem knowledge base 200 and/orconnector logic component 150 may reside on different computing devices in communication with one another. - In an embodiment, a
code analysis component 100 may analyze binary files such as, but not limited to, executables. In an embodiment, acode analysis component 100 may statically or dynamically analyze binary files. Static analysis may include analyzing a binary file that is not currently being executed. In comparison, dynamic analysis may include analyzing a binary file while the binary file is being executed. - In an embodiment, a code analysis component may be implemented using software, hardware or a combination of software and hardware. In an embodiment a
code analysis component 100 may include a disassembler, a debugger, a decompiler and/or the like. For example, thecode analysis component 100 may be a disassembler, such as IDA Pro. - A code analysis component may analyze a binary file to create an assembly language sequence. In an embodiment, the assembly language sequence may include a human-readable representation of the binary file. The
code analysis component 100 may include internal rules and/or operations which may be used to create an assembly language sequence from the binary file. Thecode analysis component 100 may analyze the assembly language sequence to determine an instruction sequence. - In an embodiment, a
code analysis component 100 may interact with external devices to analyze a binary file. For example, as discussed below, thecode analysis component 100 may communicate with an expertsystem knowledge base 200. - As illustrated by
FIG. 1 , the malware detection system may include an expertsystem knowledge base 200. In an embodiment, an expertsystem knowledge base 200 may include a representation of a human's expertise in a particular area. For example, an expertsystem knowledge base 200 may include information, data, rules and/or the like to model the knowledge and practices of an experienced computer analyst. - In an embodiment, the expert
system knowledge base 200 may be implemented using the C Language Integrated Production System (“CLIPS”). CLIPS is a programming language and software tool that may be used to create expert systems. -
FIG. 2 illustrates an expert system knowledge base according to an embodiment. The expertsystem knowledge base 200 may include internal rules and/or operations. In an embodiment, these internal rules and/or operations may be applied to an instruction sequence from an assembly language sequence to determine whether the assembly language sequence contains malware. In an embodiment, the internal rules and/or operations may represent the encoding of human expertise. - In an embodiment, a
domain expert 205 may populate the expertsystem knowledge base 200. A domain expert may be, but is not limited to, a human being who has expertise in analyzing malware. In an embodiment, adomain expert 205 may be a computing device configured to provide the expertsystem knowledge base 200 with internal rules and/or operations that may represent the encoding of human expertise. For example, a computing device may automatically provide the expertsystem knowledge base 200 with updates, enhancements or the like for one or more internal rules and/or operations. - In an embodiment, the expert
system knowledge base 200 may be populated withbinary file structures 210. A binary file structure may be a template that depicts one or more portions of a binary file and/or a sequence of the portions in a binary file. TheBinary file structures 210 may be used to analyze whether a file structure is proper. For example, abinary file structure 210 may be analyzed to determine if the header on the file conforms to a protocol. - In an embodiment, the expert
system knowledge base 200 may be populated withworm defining operations 215.Worm defining operations 215 may identify instruction sequences which replicate an assembly language sequence. - In an embodiment, the expert
system knowledge base 200 may be populated with TrojanHorse defining operations 220. TrojanHorse defining operations 220 may identify instruction sequences in an assembly language sequence that are associated with one or more Trojan Horses. - In an embodiment, the expert
system knowledge base 200 may be populated with virus defining operations 225. Virus defining operations 225 may identify self-replicating instruction sequences in an assembly language sequence. Additional and/or alternative operations may be included in the expertsystem knowledge base 200. - Referring back to
FIG. 1 , the malware detection system may include aconnector logic component 150. Aconnector logic component 150 may enable communication between thecode analysis component 100 and the expertsystem knowledge base 200. - In an embodiment, the assembly language sequence sent from the
code analysis component 100 may be in a format which cannot be directly processed by the expertsystem knowledge base 200. Thecode analysis component 100 may communicate the assembly language sequence to theconnector logic component 150. Theconnector logic component 150 may convert the instruction sequence into a format that the expertsystem knowledge base 200 can process. Theconnector logic component 150 may send the newly converted instruction sequence to the expertsystem knowledge base 200. - Similarly, the connector logic component may obtain information from the expert
system knowledge base 200. The connector logic component may convert the information from the expertsystem knowledge base 200 into a format that is readable by thecode analysis component 100 and transmit the converted information to the code analysis component. -
FIG. 3 depicts a flowchart of a method for detecting and analyzing malware according to an embodiment. A binary file may be received by the code analysis component. The code analysis component may analyze the file to obtain an assembly language sequence and an instruction sequence. The code analysis component may send the assembly language sequence with the instruction sequence to the expert system knowledge base via the connector logic component. - The expert system knowledge base may receive 300 the assembly language sequence. In an embodiment, the expert system knowledge base may identify 305 the instruction sequence from the assembly language sequence.
- The expert system knowledge base may apply internal operations and/or rules to classify 315 the instruction sequence. In an embodiment, the classification may be used to determine if the instruction sequence contains malware. For example, in an embodiment, the expert system knowledge base may classify the instruction sequence as non-threatening 315, threatening 330 or non-classifiable 345. Additional and/or alternate classifications may be used within the scope of this disclosure.
- In an embodiment, the expert system knowledge base may traverse through the nodes and branches of a received instruction sequence using one or more internal rules and/or operations. In an embodiment, the expert system knowledge base apply a group of precedential rules to the received instruction sequence. Each rule in the set of precedential rules may have a ranking with respect to the other precedential rules in the set. In an embodiment, the rules may be ranked based on the number of matches between each rule and the instruction sequence. For example, the instruction sequences that are most similar to the match criteria of a rule may cause that rule to be given a highest priority for a given traversal. Alternatively, the instruction sequences that are least similar to the match criteria of a rule may cause that rule to be given a lowest priority for a given traversal.
- CLIPS provides conflict resolution strategies such as a complexity strategy and a simplicity strategy which give precedence to the most and least specific matches, respectively. In an embodiment, such strategies may be employed to rank the rules as to those which most specifically match the instruction sequence.
- In an embodiment, the expert system knowledge base may apply the rule associated with the highest precedence to the instruction sequence. In an embodiment, one or more additional precedential rules from the group may be applied, in the order of their precedence, to the instruction sequence until the instruction sequence is classified or until all precedential rules have been applied.
- If, when applying a rule or rules, the expert system knowledge base traverses the instruction sequence from start to finish, then the instruction sequence may be classified as non-threatening 315. For example,
FIG. 5 illustrates an exemplary instruction sequence according to an embodiment. If the expert system knowledge base is able to traverse theentire instruction sequence 500 from start (Instruction 1 505) to finish (Instruction 8 510), then theinstruction sequence 500 may be classified as non-threatening. - In an embodiment, the expert system knowledge base may transmit 320 information signifying that the instruction sequence is non-threatening to the code analysis component. In an embodiment, the information may include a label attached to the instruction sequence indicating that the instruction sequence is non-threatening.
- In an embodiment, in response to classifying an instruction sequence as non-threatening, the expert system knowledge base may request 325 a new assembly sequence with a new instruction sequence to analyze from the code analysis component.
- In an embodiment, the expert system knowledge base may classify an instruction sequence as threatening 330 if the expert system knowledge base is unable to traverse each instruction of the instruction sequence. For example, the expert system knowledge base may analyze the instruction sequence by traversing the instructions of the instruction sequence to determine if there is malware. For example, a loop may be an indicator of malware. If during the traversal, the expert system knowledge base arrives at an instruction that it already analyzed, the expert system knowledge base may determine that the instruction sequence forms a loop. In an embodiment, the expert system knowledge base may classify an instruction sequence having one or more loops as threatening.
FIG. 6 illustrates an exemplary instruction sequence according to an embodiment. As illustrated byFIG. 6 , theinstruction sequence 600 may be classified as threatening because it includes a loop fromInstruction 6 605 toInstruction 4 610. - In an embodiment, other activities that may be indicative of malware or other nefarious behaviors may include encryption/decryption routines, replicating code, key logging, independent initiation of network communication, communication with known hostile or suspicious network hosts and/or the like. As such, an instruction sequence that includes one or more of these activities may be classified as threatening. Additional and/or alternate activities may be indicative of malware within the scope of this disclosure.
- In an embodiment, the expert system knowledge base may transmit 335 information signifying that the instruction sequence is threatening to the code analysis component. The information may be sent to the code analysis component via the connector logic component, which may translate the information into a form readable by the code analysis component. In an embodiment, the information may include a label attached to the instruction sequence indicating that the instruction sequence is threatening.
- In an embodiment, the information may include a request that the code analysis component search other assembly language sequences for at least a portion of an instruction sequence that was previously analyzed 340. For example, the code analysis component may search other assembly language sequences for the loop discussed in the previous example. In an embodiment, the code analysis component may use its internal operations and/or rules to translate and/or analyze the information to determine whether at least a portion of an instruction sequence is present inside the assembly language sequences. If the code analysis component finds the same instruction sequence or portion thereof, the code analysis component may send the relevant assembly language sequence and instruction sequence to the expert system knowledge base.
- In an embodiment, the expert system knowledge base may determine 345 whether an instruction sequence is non-classifiable. An instruction sequence may be identified as being non-classifiable if the expert system knowledge base is unable to determine whether the instruction sequence is threatening. For example, a programmer who created a binary file may have intentionally used methods to obfuscate the workings of the file prevent the code analysis component from issuing the correct instruction sequence. As such, the code analysis component may send an incomplete or nonsensical instruction sequence to the expert system knowledge base via the connector logic component.
- The expert system knowledge base may analyze each node of the instruction sequence using its internal rules and/or operations. Based on its analysis, the expert system knowledge base may transmit 350 a request to the code analysis component to reinterpret a particular node or series of nodes. For example, the expert system knowledge base may request that the code analysis component generate a new instruction sequence for a particular node.
- In an embodiment, the request may include alternate considerations for the code analysis component in analyzing the assembly sequence. For example, in some instances, the code analysis component may not be able to properly analyze an assembly sequence. As such, it may be necessary for the expert system knowledge base to provide information to the code analysis component that will allow the analysis to continue. For example, the expert system knowledge base may detect that an incorrect instruction sequence should be altered or ignored to allow the analysis to continue. In an embodiment, this information may be included in a request to the code analysis component.
- In an embodiment, the code analysis component may use its internal rules and/or operations reanalyze the assembly language sequence and instruction sequence. The expert system knowledge base may receive 345 the reanalyzed assembly language sequence and new instruction sequence from the code analysis component via the connector logic component. The expert system knowledge base may traverse the new instruction sequence to determine its classification.
-
FIG. 4 depicts a block diagram of an exemplary system that may be used to contain or implement program instructions according to an embodiment. Referring toFIG. 4 , abus 400 serves as the main information highway interconnecting the other illustrated components of the hardware.CPU 405 is the central processing unit of the system, performing calculations and logic operations required to execute a program. Read only memory (ROM) 410 and random access memory (RAM) 415 constitute exemplary memory devices or storage media. - A
disk controller 420 interfaces with one or more optional disk drives to thesystem bus 400. These disk drives may include, for example, external or internal DVD drives 425, CD ROM drives 430 orhard drives 435. As indicated previously, these various disk drives and disk controllers are optional devices. - Program instructions may be stored in the
ROM 410 and/or theRAM 415. Optionally, program instructions may be stored on a computer readable storage medium, such as a hard drive, a compact disk, a digital disk, a memory or any other tangible recording medium. - An
optional display interface 440 may permit information from thebus 400 to be displayed on thedisplay 445 in audio, graphic or alphanumeric format. Communication with external devices may occur usingvarious communication ports 450. - In addition to the standard computer-type components, the hardware may also include an
interface 455 which allows for receipt of data from input devices such as akeyboard 460 orother input device 465 such as a mouse, remote control, touch pad or screen, pointer and/or joystick. - It will be appreciated that various of the above-disclosed and other features and functions, or alternatives thereof, may be desirably combined into many other different systems or applications. Also that various presently unforeseen or unanticipated alternatives, modifications, variations or improvements therein may be subsequently made by those skilled in the art which are also intended to be encompassed by the following embodiments.
Claims (22)
1. A method of automatically identifying malware, the method comprising:
receiving, by an expert system knowledge base, an assembly language sequence from a binary file;
identifying an instruction sequence from the received assembly language sequence;
classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system knowledge base to the instruction sequence;
if the instruction sequence is classified as threatening, transmitting information to a code analysis component, wherein the information comprises one or more of the following:
the instruction sequence,
a label comprising an indication that the instruction sequence is threatening, and
a request that one or more other assembly language sequences from the binary file be searched for at least a portion of the instruction sequence; and
notifying a user that the binary file includes malware.
2. The method of claim 1 , wherein applying one or more rules comprises applying one or more rules written in C Language Integrated Production System language.
3. The method of claim 1 , wherein classifying the instruction sequence comprises one or more of the following:
applying one or more rules to the instruction sequence to determine whether a binary file structure of the binary file is proper;
applying one or more worm defining operations to determine whether the instruction sequence comprises one or more instructions that replicate the assembly language sequence;
applying one or more Trojan Horse defining operations to determine whether the instruction sequence comprises one or more instructions associated with one or more Trojan Horses; and
applying one or more virus defining operations to determine whether the instruction sequence comprises one or more self-replicating instructions.
4. The method of claim 1 , wherein applying one or more rules comprises:
applying a set of precedential rules to the instruction sequence, wherein the set of precedential rules comprises a plurality of precedential rules, wherein each precedential rule is associated with a precedence with respect to the other precedential rules in the set.
5. The method of claim 4 , wherein applying a set of precedential rules comprises applying the precedential rules to the instruction sequence, in order of precedence, until the instruction sequence is classified or each precedential rule has been applied.
6. The method of claim 4 , wherein applying a set of precedential rules comprises ranking the precedential rules by giving precedence to rules having a higher number of matches to the instruction sequence.
7. The method of claim 1 , wherein classifying the instruction sequence comprises classifying the instruction sequence as threatening if the instruction sequence is unable to be traversed from start to finish.
8. The method of claim 1 , wherein classifying the instructions sequence comprises, for each node in the instruction sequence:
traversing the node;
determining whether the node has previously been traversed; and
if so, classifying the instruction sequence as threatening.
9. The method of claim 1 , wherein classifying the instruction sequence comprises classifying the instruction sequence as threatening if it includes one or more of the following:
encryption routines;
decryption routines; and
one or more instructions for replicating at least a portion of the instruction sequence.
10. A method of automatically identifying malware, the method comprising:
receiving, by an expert system knowledge base, an assembly language sequence from a binary file;
identifying an instruction sequence from the received assembly language sequence;
classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system knowledge base to the instruction sequence;
if the instruction sequence is classified as non-threatening, transmitting information to a code analysis component, wherein the information comprises one or more of the following:
the instruction sequence, and
a label comprising an indication that the instruction sequence is non-threatening; and
requesting a second instruction sequence.
11. The method of claim 10 , wherein classifying the instruction sequence comprises classifying the instruction sequence as non-threatening if the expert system traverses the instruction sequence in its entirety.
12. A method of automatically identifying malware, the method comprising:
receiving, by an expert system knowledge base, an assembly language sequence from a binary file;
identifying an instruction sequence from the received assembly language sequence;
classifying, by the expert system knowledge base, the instruction sequence as threatening, non-threatening or non-classifiable by applying one or more rules of the expert system to the instruction sequence; and
if the instruction sequence is classified as non-classifiable:
transmitting a request to a code analysis component that the assembly language sequence be reanalyzed,
receiving a new instruction sequence corresponding to the reanalyzed assembly language sequence, and
classifying the new instruction sequence as threatening, non-threatening or non-classifiable.
13. A method of automatically identifying malware, the method comprising:
analyzing, by a code analysis component, a binary file to generate an assembly language sequence and a corresponding instruction sequence;
transmitting the instruction sequence to an expert system knowledge base;
receiving, from the expert system knowledge base, classification information associated with the instruction sequence;
if the classification information identifies the instruction sequence as threatening:
identifying one or more other assembly language sequences from the binary file that comprise at least a portion of the instruction sequence, and
transmitting at least one of the identified assembly language sequences to the expert system knowledge base;
if the classification information identifies the instruction sequence as non-threatening, transmitting a second instruction sequence to the expert system knowledge base; and
if the classification information identifies the instruction sequence as non-classifiable:
reanalyzing the assembly language sequence to produce a new instruction sequence, and
transmitting the new instruction sequence to the expert system knowledge base.
14. The method of claim 13 , wherein analyzing a binary file comprises one or more of statically analyzing the binary file and dynamically analyzing the binary file.
15. A system for automatically identifying malware, the system comprising:
a code analysis component configured to identify an assembly language sequence from a binary file, wherein the assembly language sequence comprises one or more instruction sequences; and
an expert system knowledge base in communication with the code analysis component, wherein the expert system knowledge base is configured to classify the instruction sequence as threatening, non-threatening or non-classifiable using one or more rules.
16. The system of claim 15 , further comprising a connector logic component in communication with the code analysis component and the expert system knowledge base, wherein the connector logic component is configured to enable communication between the code analysis component and the expert system knowledge base.
17. The system of claim 16 , wherein the connector logic component is configured to perform one or more of the following:
convert the instruction sequence into a format that the expert system knowledge base can process; and
convert information received from the expert system knowledge base into a format that the code analysis component can process.
18. The system of claim 14 , wherein the expert system knowledge base is populated with one or more of the following:
C Language Integrated Production System rules;
binary file structures;
worm defining operations;
Trojan Horse defining operations; and
virus defining operations.
19. The system of claim 14 , wherein the expert system knowledge base is configured to classify the instruction sequence by one or more of the following:
applying one or more rules to the instruction sequence to determine whether a binary file structure of the binary file is proper;
applying one or more worm defining operations to determine whether the instruction sequence comprises one or more instructions that replicate the assembly language sequence;
applying one or more Trojan Horse defining operations to determine whether the instruction sequence comprises one or more instructions associated with one or more Trojan Horses; and
applying one or more virus defining operations to determine whether the instruction sequence comprises one or more self-replicating instructions.
20. The system of claim 14 , wherein the expert system knowledge base is configured to apply a set of precedential rules to the instruction sequence, wherein the set of precedential rules comprises a plurality of precedential rules, wherein each precedential rule is associated with a precedence with respect to the other precedential rules in the set.
21. The system of claim 20 , wherein the expert system knowledge base is further configured to apply the precedential rules to the instruction sequence, in order of precedence, until the instruction sequence is classified or each precedential rule has been applied.
22. The system of claim 20 , wherein the expert system knowledge base is further configured to rank the precedential rules by giving precedence to rules having a higher number of matches to the instruction sequence.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US14/862,570 US20160012225A1 (en) | 2008-08-29 | 2015-09-23 | System and method for the detection of malware |
Applications Claiming Priority (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US9284808P | 2008-08-29 | 2008-08-29 | |
| US12/550,025 US20100058474A1 (en) | 2008-08-29 | 2009-08-28 | System and method for the detection of malware |
| US14/862,570 US20160012225A1 (en) | 2008-08-29 | 2015-09-23 | System and method for the detection of malware |
Related Parent Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/550,025 Continuation US20100058474A1 (en) | 2008-08-29 | 2009-08-28 | System and method for the detection of malware |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20160012225A1 true US20160012225A1 (en) | 2016-01-14 |
Family
ID=41721978
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/550,025 Abandoned US20100058474A1 (en) | 2008-08-29 | 2009-08-28 | System and method for the detection of malware |
| US14/862,570 Abandoned US20160012225A1 (en) | 2008-08-29 | 2015-09-23 | System and method for the detection of malware |
Family Applications Before (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/550,025 Abandoned US20100058474A1 (en) | 2008-08-29 | 2009-08-28 | System and method for the detection of malware |
Country Status (12)
| Country | Link |
|---|---|
| US (2) | US20100058474A1 (en) |
| EP (1) | EP2340488A4 (en) |
| JP (1) | JP5562961B2 (en) |
| CN (1) | CN102203791A (en) |
| AU (1) | AU2009287433B2 (en) |
| BR (1) | BRPI0913145A2 (en) |
| CA (1) | CA2735600C (en) |
| MY (1) | MY165418A (en) |
| RU (1) | RU2497189C2 (en) |
| SG (1) | SG193808A1 (en) |
| WO (1) | WO2010025453A1 (en) |
| ZA (1) | ZA201101745B (en) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180066129A1 (en) * | 2015-03-31 | 2018-03-08 | Dow Global Technologies Llc | Flooding compounds for telecommunication cables |
| WO2021046811A1 (en) * | 2019-09-12 | 2021-03-18 | 奇安信安全技术(珠海)有限公司 | Attack behavior determination method and apparatus, and computer storage medium |
| CN114301725A (en) * | 2021-12-24 | 2022-04-08 | 珠海格力电器股份有限公司 | Device control method, device, electronic device and storage medium |
Families Citing this family (178)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7051322B2 (en) | 2002-12-06 | 2006-05-23 | @Stake, Inc. | Software analysis framework |
| US8539582B1 (en) | 2004-04-01 | 2013-09-17 | Fireeye, Inc. | Malware containment and security analysis on connection |
| US8566946B1 (en) | 2006-04-20 | 2013-10-22 | Fireeye, Inc. | Malware containment on connection |
| US8793787B2 (en) | 2004-04-01 | 2014-07-29 | Fireeye, Inc. | Detecting malicious network content using virtual environment components |
| US7587537B1 (en) | 2007-11-30 | 2009-09-08 | Altera Corporation | Serializer-deserializer circuits formed from input-output circuit registers |
| US8881282B1 (en) | 2004-04-01 | 2014-11-04 | Fireeye, Inc. | Systems and methods for malware attack detection and identification |
| US8528086B1 (en) | 2004-04-01 | 2013-09-03 | Fireeye, Inc. | System and method of detecting computer worms |
| US8549638B2 (en) | 2004-06-14 | 2013-10-01 | Fireeye, Inc. | System and method of containing computer worms |
| US9106694B2 (en) | 2004-04-01 | 2015-08-11 | Fireeye, Inc. | Electronic message analysis for malware detection |
| US8171553B2 (en) | 2004-04-01 | 2012-05-01 | Fireeye, Inc. | Heuristic based capture with replay to virtual machine |
| US8613080B2 (en) | 2007-02-16 | 2013-12-17 | Veracode, Inc. | Assessment and analysis of software security flaws in virtual machines |
| US8732455B2 (en) * | 2008-07-25 | 2014-05-20 | Infotect Security Pte Ltd | Method and system for securing against leakage of source code |
| US8997219B2 (en) | 2008-11-03 | 2015-03-31 | Fireeye, Inc. | Systems and methods for detecting malicious PDF network content |
| US8850571B2 (en) | 2008-11-03 | 2014-09-30 | Fireeye, Inc. | Systems and methods for detecting malicious network content |
| US8832829B2 (en) * | 2009-09-30 | 2014-09-09 | Fireeye, Inc. | Network-based binary file extraction and analysis for malware detection |
| EP2610776B1 (en) | 2011-09-16 | 2019-08-21 | Veracode, Inc. | Automated behavioural and static analysis using an instrumented sandbox and machine learning classification for mobile security |
| RU2011138462A (en) * | 2011-09-20 | 2013-04-10 | Закрытое акционерное общество "Лаборатория Касперского" | USE OF USER SOLUTIONS TO DETECT UNKNOWN COMPUTER THREATS |
| US8533836B2 (en) * | 2012-01-13 | 2013-09-10 | Accessdata Group, Llc | Identifying software execution behavior |
| US9286063B2 (en) | 2012-02-22 | 2016-03-15 | Veracode, Inc. | Methods and systems for providing feedback and suggested programming methods |
| US9519782B2 (en) | 2012-02-24 | 2016-12-13 | Fireeye, Inc. | Detecting malicious network content |
| CN102663281B (en) * | 2012-03-16 | 2015-03-18 | 华为数字技术(成都)有限公司 | Method and device for detecting malicious software |
| CN103065090B (en) * | 2012-12-20 | 2016-03-23 | 广东欧珀移动通信有限公司 | A kind of application program malice Ad blocking method and device |
| TWI461952B (en) * | 2012-12-26 | 2014-11-21 | Univ Nat Taiwan Science Tech | Method and system for detecting malware applications |
| US10572665B2 (en) | 2012-12-28 | 2020-02-25 | Fireeye, Inc. | System and method to create a number of breakpoints in a virtual machine via virtual machine trapping events |
| US9176843B1 (en) | 2013-02-23 | 2015-11-03 | Fireeye, Inc. | Framework for efficient security coverage of mobile software applications |
| US9367681B1 (en) | 2013-02-23 | 2016-06-14 | Fireeye, Inc. | Framework for efficient security coverage of mobile software applications using symbolic execution to reach regions of interest within an application |
| US9824209B1 (en) | 2013-02-23 | 2017-11-21 | Fireeye, Inc. | Framework for efficient security coverage of mobile software applications that is usable to harden in the field code |
| US9009822B1 (en) | 2013-02-23 | 2015-04-14 | Fireeye, Inc. | Framework for multi-phase analysis of mobile applications |
| US8990944B1 (en) | 2013-02-23 | 2015-03-24 | Fireeye, Inc. | Systems and methods for automatically detecting backdoors |
| US9009823B1 (en) | 2013-02-23 | 2015-04-14 | Fireeye, Inc. | Framework for efficient security coverage of mobile software applications installed on mobile devices |
| US9195829B1 (en) | 2013-02-23 | 2015-11-24 | Fireeye, Inc. | User interface with real-time visual playback along with synchronous textual analysis log display and event/time index for anomalous behavior detection in applications |
| US9159035B1 (en) | 2013-02-23 | 2015-10-13 | Fireeye, Inc. | Framework for computer application analysis of sensitive information tracking |
| US9355247B1 (en) | 2013-03-13 | 2016-05-31 | Fireeye, Inc. | File extraction from memory dump for malicious content analysis |
| US9104867B1 (en) | 2013-03-13 | 2015-08-11 | Fireeye, Inc. | Malicious content analysis using simulated user interaction without user involvement |
| US9565202B1 (en) | 2013-03-13 | 2017-02-07 | Fireeye, Inc. | System and method for detecting exfiltration content |
| US9626509B1 (en) | 2013-03-13 | 2017-04-18 | Fireeye, Inc. | Malicious content analysis with multi-version application support within single operating environment |
| US9311479B1 (en) | 2013-03-14 | 2016-04-12 | Fireeye, Inc. | Correlation and consolidation of analytic data for holistic view of a malware attack |
| US9430646B1 (en) | 2013-03-14 | 2016-08-30 | Fireeye, Inc. | Distributed systems and methods for automatically detecting unknown bots and botnets |
| US10713358B2 (en) | 2013-03-15 | 2020-07-14 | Fireeye, Inc. | System and method to extract and utilize disassembly features to classify software intent |
| US9251343B1 (en) | 2013-03-15 | 2016-02-02 | Fireeye, Inc. | Detecting bootkits resident on compromised computers |
| US9413781B2 (en) | 2013-03-15 | 2016-08-09 | Fireeye, Inc. | System and method employing structured intelligence to verify and contain threats at endpoints |
| RU2531861C1 (en) * | 2013-04-26 | 2014-10-27 | Закрытое акционерное общество "Лаборатория Касперского" | System and method of assessment of harmfullness of code executed in addressing space of confidential process |
| US9495180B2 (en) | 2013-05-10 | 2016-11-15 | Fireeye, Inc. | Optimized resource allocation for virtual machines within a malware content detection system |
| US9635039B1 (en) | 2013-05-13 | 2017-04-25 | Fireeye, Inc. | Classifying sets of malicious indicators for detecting command and control communications associated with malware |
| US10133863B2 (en) | 2013-06-24 | 2018-11-20 | Fireeye, Inc. | Zero-day discovery system |
| US9536091B2 (en) | 2013-06-24 | 2017-01-03 | Fireeye, Inc. | System and method for detecting time-bomb malware |
| US9300686B2 (en) | 2013-06-28 | 2016-03-29 | Fireeye, Inc. | System and method for detecting malicious links in electronic messages |
| US9888016B1 (en) | 2013-06-28 | 2018-02-06 | Fireeye, Inc. | System and method for detecting phishing using password prediction |
| US9294501B2 (en) | 2013-09-30 | 2016-03-22 | Fireeye, Inc. | Fuzzy hash of behavioral results |
| US9628507B2 (en) | 2013-09-30 | 2017-04-18 | Fireeye, Inc. | Advanced persistent threat (APT) detection center |
| US10089461B1 (en) | 2013-09-30 | 2018-10-02 | Fireeye, Inc. | Page replacement code injection |
| US10192052B1 (en) | 2013-09-30 | 2019-01-29 | Fireeye, Inc. | System, apparatus and method for classifying a file as malicious using static scanning |
| US9736179B2 (en) | 2013-09-30 | 2017-08-15 | Fireeye, Inc. | System, apparatus and method for using malware analysis results to drive adaptive instrumentation of virtual machines to improve exploit detection |
| US9690936B1 (en) | 2013-09-30 | 2017-06-27 | Fireeye, Inc. | Multistage system and method for analyzing obfuscated content for malware |
| US10515214B1 (en) | 2013-09-30 | 2019-12-24 | Fireeye, Inc. | System and method for classifying malware within content created during analysis of a specimen |
| US9171160B2 (en) | 2013-09-30 | 2015-10-27 | Fireeye, Inc. | Dynamically adaptive framework and method for classifying malware using intelligent static, emulation, and dynamic analyses |
| US9189627B1 (en) | 2013-11-21 | 2015-11-17 | Fireeye, Inc. | System, apparatus and method for conducting on-the-fly decryption of encrypted objects for malware detection |
| US9747446B1 (en) | 2013-12-26 | 2017-08-29 | Fireeye, Inc. | System and method for run-time object classification |
| US9756074B2 (en) | 2013-12-26 | 2017-09-05 | Fireeye, Inc. | System and method for IPS and VM-based detection of suspicious objects |
| US9292686B2 (en) | 2014-01-16 | 2016-03-22 | Fireeye, Inc. | Micro-virtualization architecture for threat-aware microvisor deployment in a node of a network environment |
| US9262635B2 (en) | 2014-02-05 | 2016-02-16 | Fireeye, Inc. | Detection efficacy of virtual machine-based analysis with application specific events |
| US9241010B1 (en) | 2014-03-20 | 2016-01-19 | Fireeye, Inc. | System and method for network behavior detection |
| US10242185B1 (en) | 2014-03-21 | 2019-03-26 | Fireeye, Inc. | Dynamic guest image creation and rollback |
| US9591015B1 (en) | 2014-03-28 | 2017-03-07 | Fireeye, Inc. | System and method for offloading packet processing and static analysis operations |
| US9223972B1 (en) | 2014-03-31 | 2015-12-29 | Fireeye, Inc. | Dynamically remote tuning of a malware content detection system |
| US9432389B1 (en) | 2014-03-31 | 2016-08-30 | Fireeye, Inc. | System, apparatus and method for detecting a malicious attack based on static analysis of a multi-flow object |
| US9594912B1 (en) | 2014-06-06 | 2017-03-14 | Fireeye, Inc. | Return-oriented programming detection |
| US20150363197A1 (en) * | 2014-06-13 | 2015-12-17 | The Charles Stark Draper Laboratory Inc. | Systems And Methods For Software Analytics |
| US10084813B2 (en) | 2014-06-24 | 2018-09-25 | Fireeye, Inc. | Intrusion prevention and remedy system |
| US9398028B1 (en) | 2014-06-26 | 2016-07-19 | Fireeye, Inc. | System, device and method for detecting a malicious attack based on communcations between remotely hosted virtual machines and malicious web servers |
| US10805340B1 (en) | 2014-06-26 | 2020-10-13 | Fireeye, Inc. | Infection vector and malware tracking with an interactive user display |
| US10002252B2 (en) | 2014-07-01 | 2018-06-19 | Fireeye, Inc. | Verification of trusted threat-aware microvisor |
| US10360378B2 (en) | 2014-08-22 | 2019-07-23 | Nec Corporation | Analysis device, analysis method and computer-readable recording medium |
| US10671726B1 (en) | 2014-09-22 | 2020-06-02 | Fireeye Inc. | System and method for malware analysis using thread-level event monitoring |
| US9773112B1 (en) | 2014-09-29 | 2017-09-26 | Fireeye, Inc. | Exploit detection of malware and malware families |
| US10027689B1 (en) | 2014-09-29 | 2018-07-17 | Fireeye, Inc. | Interactive infection visualization for improved exploit detection and signature generation for malware and malware families |
| US9197665B1 (en) * | 2014-10-31 | 2015-11-24 | Cyberpoint International Llc | Similarity search and malware prioritization |
| US9690933B1 (en) | 2014-12-22 | 2017-06-27 | Fireeye, Inc. | Framework for classifying an object as malicious with machine learning for deploying updated predictive models |
| US10075455B2 (en) | 2014-12-26 | 2018-09-11 | Fireeye, Inc. | Zero-day rotating guest image profile |
| US9934376B1 (en) | 2014-12-29 | 2018-04-03 | Fireeye, Inc. | Malware detection appliance architecture |
| US9838417B1 (en) | 2014-12-30 | 2017-12-05 | Fireeye, Inc. | Intelligent context aware user interaction for malware detection |
| US10148693B2 (en) | 2015-03-25 | 2018-12-04 | Fireeye, Inc. | Exploit detection system |
| US9690606B1 (en) | 2015-03-25 | 2017-06-27 | Fireeye, Inc. | Selective system call monitoring |
| US9438613B1 (en) | 2015-03-30 | 2016-09-06 | Fireeye, Inc. | Dynamic content activation for automated analysis of embedded objects |
| US9483644B1 (en) | 2015-03-31 | 2016-11-01 | Fireeye, Inc. | Methods for detecting file altering malware in VM based analysis |
| US10474813B1 (en) | 2015-03-31 | 2019-11-12 | Fireeye, Inc. | Code injection technique for remediation at an endpoint of a network |
| US10417031B2 (en) | 2015-03-31 | 2019-09-17 | Fireeye, Inc. | Selective virtualization for security threat detection |
| US9654485B1 (en) | 2015-04-13 | 2017-05-16 | Fireeye, Inc. | Analytics-based security monitoring system and method |
| US9594904B1 (en) | 2015-04-23 | 2017-03-14 | Fireeye, Inc. | Detecting malware based on reflection |
| CN104869170B (en) * | 2015-05-29 | 2018-11-13 | 四川效率源信息安全技术股份有限公司 | For the decryption method of UC browser data file encryptions |
| US9516055B1 (en) * | 2015-05-29 | 2016-12-06 | Trend Micro Incorporated | Automatic malware signature extraction from runtime information |
| US10726127B1 (en) | 2015-06-30 | 2020-07-28 | Fireeye, Inc. | System and method for protecting a software component running in a virtual machine through virtual interrupts by the virtualization layer |
| US10454950B1 (en) | 2015-06-30 | 2019-10-22 | Fireeye, Inc. | Centralized aggregation technique for detecting lateral movement of stealthy cyber-attacks |
| US11113086B1 (en) | 2015-06-30 | 2021-09-07 | Fireeye, Inc. | Virtual system and method for securing external network connectivity |
| US10642753B1 (en) | 2015-06-30 | 2020-05-05 | Fireeye, Inc. | System and method for protecting a software component running in virtual machine using a virtualization layer |
| US10715542B1 (en) | 2015-08-14 | 2020-07-14 | Fireeye, Inc. | Mobile application risk analysis |
| US10176321B2 (en) | 2015-09-22 | 2019-01-08 | Fireeye, Inc. | Leveraging behavior-based rules for malware family classification |
| US10033747B1 (en) | 2015-09-29 | 2018-07-24 | Fireeye, Inc. | System and method for detecting interpreter-based exploit attacks |
| US10210329B1 (en) | 2015-09-30 | 2019-02-19 | Fireeye, Inc. | Method to detect application execution hijacking using memory protection |
| US9825976B1 (en) | 2015-09-30 | 2017-11-21 | Fireeye, Inc. | Detection and classification of exploit kits |
| US9825989B1 (en) | 2015-09-30 | 2017-11-21 | Fireeye, Inc. | Cyber attack early warning system |
| US10706149B1 (en) | 2015-09-30 | 2020-07-07 | Fireeye, Inc. | Detecting delayed activation malware using a primary controller and plural time controllers |
| US10817606B1 (en) | 2015-09-30 | 2020-10-27 | Fireeye, Inc. | Detecting delayed activation malware using a run-time monitoring agent and time-dilation logic |
| US10601865B1 (en) | 2015-09-30 | 2020-03-24 | Fireeye, Inc. | Detection of credential spearphishing attacks using email analysis |
| US10284575B2 (en) | 2015-11-10 | 2019-05-07 | Fireeye, Inc. | Launcher for setting analysis environment variations for malware detection |
| RU2613535C1 (en) * | 2015-11-20 | 2017-03-16 | Илья Самуилович Рабинович | Method for detecting malicious software and elements |
| US10447728B1 (en) | 2015-12-10 | 2019-10-15 | Fireeye, Inc. | Technique for protecting guest processes using a layered virtualization architecture |
| US10846117B1 (en) | 2015-12-10 | 2020-11-24 | Fireeye, Inc. | Technique for establishing secure communication between host and guest processes of a virtualization architecture |
| US10108446B1 (en) | 2015-12-11 | 2018-10-23 | Fireeye, Inc. | Late load technique for deploying a virtualization layer underneath a running operating system |
| US10565378B1 (en) | 2015-12-30 | 2020-02-18 | Fireeye, Inc. | Exploit of privilege detection framework |
| US10133866B1 (en) | 2015-12-30 | 2018-11-20 | Fireeye, Inc. | System and method for triggering analysis of an object for malware in response to modification of that object |
| US10621338B1 (en) | 2015-12-30 | 2020-04-14 | Fireeye, Inc. | Method to detect forgery and exploits using last branch recording registers |
| US10050998B1 (en) | 2015-12-30 | 2018-08-14 | Fireeye, Inc. | Malicious message analysis system |
| US11552986B1 (en) | 2015-12-31 | 2023-01-10 | Fireeye Security Holdings Us Llc | Cyber-security framework for application of virtual features |
| US10581874B1 (en) | 2015-12-31 | 2020-03-03 | Fireeye, Inc. | Malware detection system with contextual analysis |
| US9824216B1 (en) | 2015-12-31 | 2017-11-21 | Fireeye, Inc. | Susceptible environment detection system |
| US10671721B1 (en) | 2016-03-25 | 2020-06-02 | Fireeye, Inc. | Timeout management services |
| US10601863B1 (en) | 2016-03-25 | 2020-03-24 | Fireeye, Inc. | System and method for managing sensor enrollment |
| US10785255B1 (en) | 2016-03-25 | 2020-09-22 | Fireeye, Inc. | Cluster configuration within a scalable malware detection system |
| US10616266B1 (en) | 2016-03-25 | 2020-04-07 | Fireeye, Inc. | Distributed malware detection system and submission workflow thereof |
| US10826933B1 (en) | 2016-03-31 | 2020-11-03 | Fireeye, Inc. | Technique for verifying exploit/malware at malware detection appliance through correlation with endpoints |
| US10893059B1 (en) | 2016-03-31 | 2021-01-12 | Fireeye, Inc. | Verification and enhancement using detection systems located at the network periphery and endpoint devices |
| US10169585B1 (en) | 2016-06-22 | 2019-01-01 | Fireeye, Inc. | System and methods for advanced malware detection through placement of transition events |
| US10462173B1 (en) | 2016-06-30 | 2019-10-29 | Fireeye, Inc. | Malware detection verification and enhancement by coordinating endpoint and malware detection systems |
| US10592678B1 (en) | 2016-09-09 | 2020-03-17 | Fireeye, Inc. | Secure communications between peers using a verified virtual trusted platform module |
| US10491627B1 (en) | 2016-09-29 | 2019-11-26 | Fireeye, Inc. | Advanced malware detection using similarity analysis |
| US10795991B1 (en) | 2016-11-08 | 2020-10-06 | Fireeye, Inc. | Enterprise search |
| US10587647B1 (en) | 2016-11-22 | 2020-03-10 | Fireeye, Inc. | Technique for malware detection capability comparison of network security devices |
| US10552610B1 (en) | 2016-12-22 | 2020-02-04 | Fireeye, Inc. | Adaptive virtual machine snapshot update framework for malware behavioral analysis |
| US10581879B1 (en) | 2016-12-22 | 2020-03-03 | Fireeye, Inc. | Enhanced malware detection for generated objects |
| US10523609B1 (en) | 2016-12-27 | 2019-12-31 | Fireeye, Inc. | Multi-vector malware detection and analysis |
| US10904286B1 (en) | 2017-03-24 | 2021-01-26 | Fireeye, Inc. | Detection of phishing attacks using similarity analysis |
| US10902119B1 (en) | 2017-03-30 | 2021-01-26 | Fireeye, Inc. | Data extraction system for malware analysis |
| US10848397B1 (en) | 2017-03-30 | 2020-11-24 | Fireeye, Inc. | System and method for enforcing compliance with subscription requirements for cyber-attack detection service |
| US10798112B2 (en) | 2017-03-30 | 2020-10-06 | Fireeye, Inc. | Attribute-controlled malware detection |
| US10791138B1 (en) | 2017-03-30 | 2020-09-29 | Fireeye, Inc. | Subscription-based malware detection |
| US10855700B1 (en) | 2017-06-29 | 2020-12-01 | Fireeye, Inc. | Post-intrusion detection of cyber-attacks during lateral movement within networks |
| US10601848B1 (en) | 2017-06-29 | 2020-03-24 | Fireeye, Inc. | Cyber-security system and method for weak indicator detection and correlation to generate strong indicators |
| US10503904B1 (en) | 2017-06-29 | 2019-12-10 | Fireeye, Inc. | Ransomware detection and mitigation |
| US10893068B1 (en) | 2017-06-30 | 2021-01-12 | Fireeye, Inc. | Ransomware file modification prevention technique |
| WO2019035313A1 (en) * | 2017-08-18 | 2019-02-21 | 日本電信電話株式会社 | Intrusion prevention device, intrusion prevention method, and program |
| US10747872B1 (en) | 2017-09-27 | 2020-08-18 | Fireeye, Inc. | System and method for preventing malware evasion |
| US10805346B2 (en) | 2017-10-01 | 2020-10-13 | Fireeye, Inc. | Phishing attack detection |
| US11108809B2 (en) | 2017-10-27 | 2021-08-31 | Fireeye, Inc. | System and method for analyzing binary code for malware classification using artificial neural network techniques |
| US11005860B1 (en) | 2017-12-28 | 2021-05-11 | Fireeye, Inc. | Method and system for efficient cybersecurity analysis of endpoint events |
| US11271955B2 (en) | 2017-12-28 | 2022-03-08 | Fireeye Security Holdings Us Llc | Platform and method for retroactive reclassification employing a cybersecurity-based global data store |
| US11240275B1 (en) | 2017-12-28 | 2022-02-01 | Fireeye Security Holdings Us Llc | Platform and method for performing cybersecurity analyses employing an intelligence hub with a modular architecture |
| US10826931B1 (en) | 2018-03-29 | 2020-11-03 | Fireeye, Inc. | System and method for predicting and mitigating cybersecurity system misconfigurations |
| US11003773B1 (en) | 2018-03-30 | 2021-05-11 | Fireeye, Inc. | System and method for automatically generating malware detection rule recommendations |
| US10956477B1 (en) | 2018-03-30 | 2021-03-23 | Fireeye, Inc. | System and method for detecting malicious scripts through natural language processing modeling |
| US11558401B1 (en) | 2018-03-30 | 2023-01-17 | Fireeye Security Holdings Us Llc | Multi-vector malware detection data sharing system for improved detection |
| US11314859B1 (en) | 2018-06-27 | 2022-04-26 | FireEye Security Holdings, Inc. | Cyber-security system and method for detecting escalation of privileges within an access token |
| US11075930B1 (en) | 2018-06-27 | 2021-07-27 | Fireeye, Inc. | System and method for detecting repetitive cybersecurity attacks constituting an email campaign |
| US11228491B1 (en) | 2018-06-28 | 2022-01-18 | Fireeye Security Holdings Us Llc | System and method for distributed cluster configuration monitoring and management |
| US11316900B1 (en) | 2018-06-29 | 2022-04-26 | FireEye Security Holdings Inc. | System and method for automatically prioritizing rules for cyber-threat detection and mitigation |
| US11182473B1 (en) | 2018-09-13 | 2021-11-23 | Fireeye Security Holdings Us Llc | System and method for mitigating cyberattacks against processor operability by a guest process |
| US11763004B1 (en) | 2018-09-27 | 2023-09-19 | Fireeye Security Holdings Us Llc | System and method for bootkit detection |
| US12074887B1 (en) | 2018-12-21 | 2024-08-27 | Musarubra Us Llc | System and method for selectively processing content after identification and removal of malicious content |
| US11743290B2 (en) | 2018-12-21 | 2023-08-29 | Fireeye Security Holdings Us Llc | System and method for detecting cyberattacks impersonating legitimate sources |
| US11176251B1 (en) | 2018-12-21 | 2021-11-16 | Fireeye, Inc. | Determining malware via symbolic function hash analysis |
| US11368475B1 (en) | 2018-12-21 | 2022-06-21 | Fireeye Security Holdings Us Llc | System and method for scanning remote services to locate stored objects with malware |
| US11601444B1 (en) | 2018-12-31 | 2023-03-07 | Fireeye Security Holdings Us Llc | Automated system for triage of customer issues |
| US11310238B1 (en) | 2019-03-26 | 2022-04-19 | FireEye Security Holdings, Inc. | System and method for retrieval and analysis of operational data from customer, cloud-hosted virtual resources |
| US11677786B1 (en) | 2019-03-29 | 2023-06-13 | Fireeye Security Holdings Us Llc | System and method for detecting and protecting against cybersecurity attacks on servers |
| US11636198B1 (en) | 2019-03-30 | 2023-04-25 | Fireeye Security Holdings Us Llc | System and method for cybersecurity analyzer update and concurrent management system |
| US11258806B1 (en) | 2019-06-24 | 2022-02-22 | Mandiant, Inc. | System and method for automatically associating cybersecurity intelligence to cyberthreat actors |
| US11556640B1 (en) | 2019-06-27 | 2023-01-17 | Mandiant, Inc. | Systems and methods for automated cybersecurity analysis of extracted binary string sets |
| US12445458B1 (en) | 2019-06-28 | 2025-10-14 | Google Llc | System and method for identifying malicious hosts prior to commencement of a cyber-attack |
| US11392700B1 (en) | 2019-06-28 | 2022-07-19 | Fireeye Security Holdings Us Llc | System and method for supporting cross-platform data verification |
| US12200013B2 (en) | 2019-08-07 | 2025-01-14 | Musarubra Us Llc | System and method for detecting cyberattacks impersonating legitimate sources |
| US11886585B1 (en) | 2019-09-27 | 2024-01-30 | Musarubra Us Llc | System and method for identifying and mitigating cyberattacks through malicious position-independent code execution |
| US11637862B1 (en) | 2019-09-30 | 2023-04-25 | Mandiant, Inc. | System and method for surfacing cyber-security threats with a self-learning recommendation engine |
| CN110704068B (en) * | 2019-10-18 | 2023-02-17 | 安徽中科国创高可信软件有限公司 | Processing method and system for cross-file collaborative program analysis based on database |
| US11838300B1 (en) | 2019-12-24 | 2023-12-05 | Musarubra Us Llc | Run-time configurable cybersecurity system |
| US11436327B1 (en) | 2019-12-24 | 2022-09-06 | Fireeye Security Holdings Us Llc | System and method for circumventing evasive code for cyberthreat detection |
| US11522884B1 (en) | 2019-12-24 | 2022-12-06 | Fireeye Security Holdings Us Llc | Subscription and key management system |
| US11528294B2 (en) * | 2021-02-18 | 2022-12-13 | SecureworksCorp. | Systems and methods for automated threat detection |
| CN114884686B (en) * | 2022-03-17 | 2024-03-08 | 新华三信息安全技术有限公司 | PHP threat identification method and device |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6347374B1 (en) * | 1998-06-05 | 2002-02-12 | Intrusion.Com, Inc. | Event detection |
| US20050022018A1 (en) * | 2003-06-30 | 2005-01-27 | Symantec Corporation | Signature extraction system and method |
| US20070094734A1 (en) * | 2005-09-29 | 2007-04-26 | Mangione-Smith William H | Malware mutation detector |
| US20090271867A1 (en) * | 2005-12-30 | 2009-10-29 | Peng Zhang | Virtual machine to detect malicious code |
| US8528086B1 (en) * | 2004-04-01 | 2013-09-03 | Fireeye, Inc. | System and method of detecting computer worms |
Family Cites Families (13)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5960170A (en) * | 1997-03-18 | 1999-09-28 | Trend Micro, Inc. | Event triggered iterative virus detection |
| RU22718U1 (en) * | 2001-12-28 | 2002-04-20 | Кулик Сергей Дмитриевич | DEVICE FOR IMPLEMENTATION AND MODELING OF COMPUTER VIRUS OF MUTANT |
| RU2248608C1 (en) * | 2003-07-22 | 2005-03-20 | Павлов Владимир Павлович | Processor |
| US20050086526A1 (en) * | 2003-10-17 | 2005-04-21 | Panda Software S.L. (Sociedad Unipersonal) | Computer implemented method providing software virus infection information in real time |
| RU2271613C1 (en) * | 2004-09-15 | 2006-03-10 | Военный университет связи | Method for protecting computer networks against unauthorized attack |
| EP1854019A4 (en) * | 2004-09-22 | 2010-12-22 | Cyberdefender Corp | Threat protection network |
| US7636856B2 (en) * | 2004-12-06 | 2009-12-22 | Microsoft Corporation | Proactive computer malware protection through dynamic translation |
| JP2006201845A (en) * | 2005-01-18 | 2006-08-03 | Hitachi Software Eng Co Ltd | Computer preventing virus infection and secret information disclosure |
| GB0513375D0 (en) * | 2005-06-30 | 2005-08-03 | Retento Ltd | Computer security |
| US8312545B2 (en) * | 2006-04-06 | 2012-11-13 | Juniper Networks, Inc. | Non-signature malware detection system and method for mobile platforms |
| US8261344B2 (en) * | 2006-06-30 | 2012-09-04 | Sophos Plc | Method and system for classification of software using characteristics and combinations of such characteristics |
| JP2008129714A (en) * | 2006-11-17 | 2008-06-05 | Univ Of Tsukuba | Abnormality detection method, abnormality detection device, abnormality detection program, and learning model generation method |
| IL181426A (en) * | 2007-02-19 | 2011-06-30 | Deutsche Telekom Ag | Automatic extraction of signatures for malware |
-
2009
- 2009-08-28 US US12/550,025 patent/US20100058474A1/en not_active Abandoned
- 2009-08-31 CA CA2735600A patent/CA2735600C/en active Active
- 2009-08-31 MY MYPI2011000836A patent/MY165418A/en unknown
- 2009-08-31 CN CN2009801429308A patent/CN102203791A/en active Pending
- 2009-08-31 EP EP09810716A patent/EP2340488A4/en not_active Withdrawn
- 2009-08-31 RU RU2011111719/08A patent/RU2497189C2/en active
- 2009-08-31 BR BRPI0913145A patent/BRPI0913145A2/en not_active IP Right Cessation
- 2009-08-31 JP JP2011525271A patent/JP5562961B2/en active Active
- 2009-08-31 WO PCT/US2009/055524 patent/WO2010025453A1/en not_active Ceased
- 2009-08-31 SG SG2013063151A patent/SG193808A1/en unknown
- 2009-08-31 AU AU2009287433A patent/AU2009287433B2/en not_active Ceased
-
2011
- 2011-03-07 ZA ZA2011/01745A patent/ZA201101745B/en unknown
-
2015
- 2015-09-23 US US14/862,570 patent/US20160012225A1/en not_active Abandoned
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6347374B1 (en) * | 1998-06-05 | 2002-02-12 | Intrusion.Com, Inc. | Event detection |
| US20050022018A1 (en) * | 2003-06-30 | 2005-01-27 | Symantec Corporation | Signature extraction system and method |
| US8528086B1 (en) * | 2004-04-01 | 2013-09-03 | Fireeye, Inc. | System and method of detecting computer worms |
| US20070094734A1 (en) * | 2005-09-29 | 2007-04-26 | Mangione-Smith William H | Malware mutation detector |
| US20090271867A1 (en) * | 2005-12-30 | 2009-10-29 | Peng Zhang | Virtual machine to detect malicious code |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180066129A1 (en) * | 2015-03-31 | 2018-03-08 | Dow Global Technologies Llc | Flooding compounds for telecommunication cables |
| WO2021046811A1 (en) * | 2019-09-12 | 2021-03-18 | 奇安信安全技术(珠海)有限公司 | Attack behavior determination method and apparatus, and computer storage medium |
| CN113632432A (en) * | 2019-09-12 | 2021-11-09 | 奇安信安全技术(珠海)有限公司 | Method, device and computer storage medium for determining attack behavior |
| CN114301725A (en) * | 2021-12-24 | 2022-04-08 | 珠海格力电器股份有限公司 | Device control method, device, electronic device and storage medium |
Also Published As
| Publication number | Publication date |
|---|---|
| JP5562961B2 (en) | 2014-07-30 |
| AU2009287433A1 (en) | 2010-03-04 |
| EP2340488A4 (en) | 2012-07-11 |
| MY165418A (en) | 2018-03-21 |
| CA2735600A1 (en) | 2010-03-04 |
| SG193808A1 (en) | 2013-10-30 |
| CN102203791A (en) | 2011-09-28 |
| RU2011111719A (en) | 2012-10-10 |
| US20100058474A1 (en) | 2010-03-04 |
| JP2012501504A (en) | 2012-01-19 |
| AU2009287433B2 (en) | 2014-06-05 |
| ZA201101745B (en) | 2012-01-25 |
| WO2010025453A1 (en) | 2010-03-04 |
| CA2735600C (en) | 2018-08-21 |
| RU2497189C2 (en) | 2013-10-27 |
| EP2340488A1 (en) | 2011-07-06 |
| BRPI0913145A2 (en) | 2019-09-24 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| AU2009287433B2 (en) | System and method for detection of malware | |
| EP2472425B1 (en) | System and method for detecting unknown malware | |
| JP5972401B2 (en) | Attack analysis system, linkage device, attack analysis linkage method, and program | |
| US11455400B2 (en) | Method, system, and storage medium for security of software components | |
| US8291500B1 (en) | Systems and methods for automated malware artifact retrieval and analysis | |
| US8732836B2 (en) | System and method for correcting antivirus records to minimize false malware detections | |
| US20150186649A1 (en) | Function Fingerprinting | |
| EP3800570B1 (en) | Methods and systems for genetic malware analysis and classification using code reuse patterns | |
| US11657899B2 (en) | Computing device | |
| Ladisa et al. | On the feasibility of cross-language detection of malicious packages in npm and pypi | |
| Naidu et al. | A syntactic approach for detecting viral polymorphic malware variants | |
| WO2025049586A1 (en) | Generative sequence processing models for cybersecurity | |
| Ugarte-Pedrero et al. | On the adoption of anomaly detection for packed executable filtering | |
| JP6258189B2 (en) | Specific apparatus, specific method, and specific program | |
| CN119538266A (en) | A safety detection method, device, equipment, medium and product | |
| JP2011034377A (en) | Information processor, information processing method and program | |
| TWI715647B (en) | System and method for ip fingerprinting and ip dna analysis | |
| HK1162708A (en) | System and method for detection of malware | |
| Hennelová et al. | The Impact of Anti-forensic Techniques on Data-Driven Digital Forensics: Anomaly Detection Case Study | |
| US20250165724A1 (en) | Cyber security system for electronic communications | |
| US12306942B2 (en) | Identifying security events in programming code for logging | |
| Labied et al. | A Deep Learning-Based Model for Malware Detection Using Transformer Architecture | |
| KR102016967B1 (en) | Method of processing vulnerability/risk through data correlation/association analysis of system information for system and processing the vulnerability/risk of system and apparatus therefor | |
| JP2008293273A (en) | Information processing apparatus and information processing method |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: AVAST SOFTWARE B.V., NETHERLANDS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:AVG NETHERLANDS B.V.;REEL/FRAME:043603/0008 Effective date: 20170901 |
|
| STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |