[go: up one dir, main page]

WO2016169193A1 - Procédé et appareil pour détecter des clics de triche - Google Patents

Procédé et appareil pour détecter des clics de triche Download PDF

Info

Publication number
WO2016169193A1
WO2016169193A1 PCT/CN2015/089545 CN2015089545W WO2016169193A1 WO 2016169193 A1 WO2016169193 A1 WO 2016169193A1 CN 2015089545 W CN2015089545 W CN 2015089545W WO 2016169193 A1 WO2016169193 A1 WO 2016169193A1
Authority
WO
WIPO (PCT)
Prior art keywords
suspicious
content
click
cheating
group
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2015/089545
Other languages
English (en)
Chinese (zh)
Inventor
庄馨
田天
朱军
夏粉
张潼
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Baidu Online Network Technology Beijing Co Ltd
Original Assignee
Baidu Online Network Technology Beijing Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Baidu Online Network Technology Beijing Co Ltd filed Critical Baidu Online Network Technology Beijing Co Ltd
Publication of WO2016169193A1 publication Critical patent/WO2016169193A1/fr
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor

Definitions

  • the present application relates to the field of network technologies, and in particular, to a method and apparatus for detecting click cheating.
  • the existing methods for detecting cheats in advertising crowds include: 1 discovering suspicious clickers by establishing rules describing the characteristics of individual clickers, and then judging cheating behavior.
  • the drawback of this technique is that because the source of crowdsourcing cheats is real user clicks rather than machines, the behavior is highly random and difficult to judge based on rules established by individual users or advertisers. 2 Determine whether there is a cheating behavior against it by observing the click traffic of an advertiser.
  • the drawback of this technique is that since crowdsourcing cheats come from real users, when they feel that cheating is detected, they can quickly adjust their behavior, thus invalidating the previous rules.
  • the present application provides a method and apparatus for detecting click cheating, which solves the technical problem of low detection efficiency and low detection precision for cheating clicks in the prior art.
  • the present application provides a method for detecting a cheating of a click, the method comprising: determining a suspicious click based on a number of times the predetermined content is clicked by the user within a predetermined time period; determining at least one group based on the suspicious click a suspicious user group suspected of cheating; determining a non-cheating user group to be excluded according to a keyword of the suspicious content clicked by the suspicious user group in the predetermined time period; and excluding non-cheating in the suspicious user group User groups to identify groups of cheating users.
  • the determining the suspicious click based on the number of times the predetermined content is clicked by the user within the predetermined time period comprises: obtaining the number of times each predetermined content is clicked by the user within the predetermined time period; determining each of the Whether the number of times the predetermined content is clicked satisfies the predetermined condition; the click corresponding to the predetermined content whose number of clicks satisfies the predetermined condition is determined as a suspicious click.
  • determining whether the number of clicks satisfies a predetermined condition comprises: determining whether the number of clicks is greater than or equal to a first predetermined threshold, and less than or equal to a second predetermined threshold; if yes, determining the The number of times clicked satisfies the predetermined condition.
  • the determining, according to the suspicious click, at least one group of suspicious users suspected of cheating comprises: obtaining related information of the suspicious clicks; determining at least one set of suspicious users based on the related information A group in which each group of suspicious user groups clicks on the same set of content during the same time period.
  • the information related to the suspicious click includes at least one of the following: identification information of the user corresponding to the suspicious click; identification information of the content corresponding to the suspicious click; and corresponding to the suspicious click time.
  • the determining the at least one set of suspicious user groups based on the related information comprises: clustering the suspicious clicks based on the related information, so that a user group corresponding to each cluster center is The same group of content is clicked in the same time period; the user group corresponding to each cluster center is determined as a group of suspicious users.
  • the determining a non-cheating user group to be excluded according to a keyword of the suspicious content clicked by the group of the suspicious user within the predetermined time period includes: acquiring a keyword of the suspicious content clicked by the suspicious user group in the predetermined time period; determining, according to the keyword, whether the suspicious content is a similar content; if yes, corresponding to the suspicious content
  • the suspicious user community is identified as a non-cheating user group to be excluded.
  • the determining whether the suspicious content is a similar content based on the keyword comprises: determining whether a proportion of a similar keyword in the keyword is greater than or equal to a predetermined ratio; if yes, determining the Suspicious content is similar content.
  • the present application provides an apparatus for detecting a cheating of a click, the apparatus comprising: a first determining unit, configured to determine a suspicious click based on a number of times the predetermined content is clicked by the user within a predetermined time period; a unit, configured to determine, according to the suspicious click, at least one group of suspicious users suspected of cheating; and a third determining unit, configured to: according to the suspicious content clicked by each group of the suspicious user group in the predetermined time period
  • the keyword determines a non-cheating user group to be excluded; and a fourth determining unit, configured to exclude the non-cheating user group in the suspicious user group to determine a cheating user group.
  • the first determining unit includes: an obtaining subunit, configured to acquire a number of times each predetermined content is clicked by the user in the predetermined time period; and a determining subunit, configured to determine each of the predetermined Whether the number of times the content is clicked satisfies a predetermined condition; the determining subunit is configured to determine a click corresponding to the predetermined content whose number of clicks meets the predetermined condition as a suspicious click.
  • the determining subunit is configured to: determine whether the number of clicks is greater than or equal to a first predetermined threshold, and less than or equal to a second predetermined threshold; if yes, determine that the number of clicks is satisfied Predetermined conditions.
  • the second determining unit includes: an information acquiring subunit, configured to acquire related information of the suspicious click; and a user group determining subunit, configured to determine at least one set of suspicious based on the related information A group of users, in which each group of suspicious users clicks on the same set of content during the same time period.
  • the information related to the suspicious click includes at least one of the following: identification information of the user corresponding to the suspicious click; identification information of the content corresponding to the suspicious click; and corresponding to the suspicious click time.
  • the user group determining subunit configuration is configured to: The related information clusters the suspicious clicks, so that the user groups corresponding to each cluster center click the same group of content in the same time period; the user groups corresponding to each cluster center are determined as a group of suspicious Client.
  • the third determining unit includes: a keyword obtaining subunit, configured to acquire a keyword of the suspicious content clicked by the group of the suspicious user within the predetermined time period; the category judging a unit, configured to determine, according to the keyword, whether the suspicious content is a similar content; the group determining sub-unit to be excluded, configured to determine, in response to the suspicious content as a similar content, the suspicious user group corresponding to the suspicious content as being to be excluded Non-cheating user groups.
  • the category determining subunit is configured to: determine whether a proportion of the same type of keywords in the keyword is greater than or equal to a predetermined ratio; if yes, determine that the suspicious content is a similar content.
  • the method and device for detecting click cheating provided by the present application determine the suspicious user group suspected of cheating by narrowing the detection range, and the non-cheating user in the suspicious user group according to the keyword of the suspicious content clicked by the suspicious user group Group exclusion, which enables monitoring of clicks on scheduled content, improves the efficiency and detection accuracy of detecting cheating clicks, and reduces waste of time and resources.
  • FIG. 1 is a flowchart of an embodiment of a method for detecting click cheating provided by an embodiment of the present application
  • FIG. 2 is a flowchart of an embodiment of a method for determining a suspicious click provided by an embodiment of the present application
  • FIG. 3 is a flowchart of an embodiment of a method for determining at least one suspicious user group suspected of cheating according to a suspicious click according to an embodiment of the present application
  • FIG. 4 is a flowchart of an embodiment of a method for determining a non-cheating user group to be excluded according to keywords of suspicious content clicked by each group of suspicious user groups in the predetermined time period according to an embodiment of the present application;
  • FIG. 5 is a schematic structural diagram of an embodiment of an apparatus for detecting click cheating provided by an embodiment of the present application
  • FIG. 6 is a schematic structural diagram of a computer system suitable for implementing a terminal device or a server of an embodiment of the present application.
  • FIG. 1 a flow 100 of one embodiment of a method for detecting click cheating is shown.
  • a suspicious click is determined based on the number of times the predetermined content is clicked by the user within a predetermined time period.
  • the predetermined content is content that may be involved in cheating clicks, such as content that can benefit from clicks (such as advertisements, voting, social networking sites, etc.), the greater the number of clicks on which the content is clicked, and The greater the benefits to its associated beneficiaries or units.
  • the beneficiaries associated with the scheduled content pass the task publisher, publish the task of clicking the scheduled content (such as advertising) on the crowdsourcing platform, and then the task publisher organizes a large number of netizens to receive the problem. Tasks, netizens complete tasks by clicking on the scheduled content, thereby obtaining rewards for completing the task.
  • the predetermined period of time may select a period of time in which the predetermined content is clicked.
  • the distribution of the click corresponding to the predetermined content on the time axis may be acquired, and the distribution density is greater than a predetermined threshold. A period of time is taken as a predetermined period of time.
  • the amount of clicks on the predetermined content at each time may also be acquired, and a period of time in which the amount of clicks each time is greater than a predetermined threshold is taken as the predetermined time period. It can be understood that there may be other implementation manners for selecting a predetermined time period, which is not limited in this application.
  • the suspicious click may be determined based on the number of times the predetermined content is clicked by the user within the predetermined time period.
  • step 102 at least one suspicious user group suspected of cheating is determined according to the suspicious click.
  • a cheating task publisher may publish a click task containing multiple predetermined content to a crowdsourcing platform at a time, for example, publishing a click task containing 10 advertisements to the crowdsourcing platform.
  • the cheating clicker After the cheating clicker receives the cheat click task, it usually clicks on a set of scheduled contents in the task within a certain period of time. Therefore, the user group performing the task (clicking on a predetermined set of contents corresponding to the task) can be determined by analyzing related information of the suspicious click (such as click time, clicked content, etc.). A group of users who clicked on the same set of predetermined content was identified as a suspicious user group suspected of cheating.
  • the above determined suspicious clicks may include multiple sets of predetermined content corresponding to different cheating click tasks.
  • the predetermined content corresponding to the same cheating click task is a set of predetermined content.
  • each set of scheduled content can also correspond to a group of suspicious users suspected of cheating. Therefore, the suspicious click determined above corresponds to at least one group of suspicious users suspected of cheating.
  • step 103 the non-cheating user group to be excluded is determined according to the keyword of the suspicious content clicked by each group of suspicious user groups within a predetermined time period.
  • some scheduled content may be more popular during a certain period of time, with a large amount of user clicks, or some users may be interested in a related batch of predetermined content within a certain period of time. For example, if a flu occurs in a certain area at a certain time, then there may be a large number of living in the area. People search online for anti-flu or flu-preventing drugs, and the ads or websites that these residents click on may overlap and are related to the flu. For another example, a certain season is more suitable for traveling to a certain area. In this season, a large number of users may search online and click on advertisements or websites related to travel to the area. The content clicked by these users may also partially overlap, and both Related to tourism in the area.
  • These predetermined content does not involve cheating clicks, and users who click on these predetermined content are not cheat user groups, but may be identified as cheating user groups suspected of cheating. Therefore, these non-cheating user groups need to be cheated. Find and exclude suspected suspicious user groups.
  • the suspicious content is a predetermined content that the suspicious user group clicks within the predetermined time period.
  • the non-cheating user group to be excluded may be determined according to the keywords of the suspicious content clicked by each group of suspicious user groups. If the keyword proximity of all suspicious content clicked by a certain group of suspicious users is high, it can be determined that the user group is a non-cheating user group.
  • step 104 the non-cheating user population in the suspicious user population is excluded to determine the cheating user population.
  • the non-cheating user group in the suspicious user group is excluded, and the remaining suspicious user groups are determined as the cheating user group.
  • the method provided by the foregoing embodiment of the present application determines a suspicious user group suspected of cheating by narrowing the detection range, and excludes the non-cheating user group in the suspicious user group according to the keyword of the suspicious content clicked by the suspicious user group, thereby
  • the monitoring of the clicks of the predetermined content is realized, the efficiency of detecting cheating clicks and the detection precision are improved, and the waste of time and resources is reduced.
  • FIG. 2 a flow 200 of one embodiment of a method of determining suspicious clicks is illustrated.
  • step 201 the number of times each predetermined content is clicked by the user within the predetermined time period is acquired.
  • the number of times each predetermined content is clicked by the user in the predetermined time period can be obtained by the click log of the predetermined content. It can be understood that each predetermined content in the predetermined time period can be obtained by other means. The number of times the user clicked. Ben Shen Please do not limit this aspect.
  • step 202 it is determined whether the number of times each predetermined content is clicked satisfies a predetermined condition.
  • a predetermined threshold a the likelihood that the predetermined content involves cheating is small. Because the purpose of cheating is to increase the amount of clicks, if the predetermined content involves cheating, the number of times it is clicked must not be too small.
  • the probability that the predetermined content involves cheating will be small. Because cheating can increase the amount of clicks, the scale of organizational cheating is usually limited, and it is impossible to reach an excessive level. For example, suppose a cheating click can increase the number of clicks 1000, and if the number of times a predetermined content is clicked is 10,000, it can be judged that the predetermined content must not involve cheating. Because even if the predetermined content involves cheating, then the corresponding normal click volume is closer to an order of magnitude larger than the amount of clicks that can be increased by cheating clicks. Therefore, it is not meaningful to increase the amount of clicks by cheating.
  • the predetermined condition is a condition that the predetermined content that may be involved in cheating is satisfied by the number of clicks, and firstly, it is determined whether the number of times the predetermined content is clicked within the predetermined time period is greater than or equal to a first predetermined threshold, and less than or equal to the second predetermined. Threshold. If the number of times is greater than or equal to the first predetermined threshold and less than or equal to the second predetermined threshold, it is determined that the number of times satisfies the predetermined condition.
  • step 203 the click corresponding to the predetermined content whose number of clicks satisfies the predetermined condition is determined as a suspicious click.
  • the predetermined content is likely to involve cheating, and the corresponding click is determined as a suspicious click. It should be noted that suspicious clicks do not necessarily mean cheating clicks, because even if a content involves cheating, the content will also be clicked normally by non-cheating users.
  • a flow 300 of one embodiment of a method of determining at least one suspicious user population suspected of cheating based on suspicious clicks is illustrated.
  • step 301 information related to suspicious clicks is obtained.
  • the related information of the suspicious click may include at least one of the following: The identification information of the user corresponding to the suspicious click; the identification information of the content corresponding to the suspicious click; and the time corresponding to the suspicious click.
  • the identifier information of the user corresponding to the suspicious click may be the MAC address of the user who performs the suspicious click, or the IP address, or the serial number of the terminal device (such as a mobile phone, a computer, etc.), and the present application has suspicious clicks.
  • the specific content and form of the corresponding user identification information are not limited.
  • the content identification information corresponding to the suspicious click may be the name of the content clicked by the suspicious click, or the information such as the number used to identify or distinguish the content, and the specific content and form of the content identification information corresponding to the suspicious click in the present application.
  • the time corresponding to the suspicious click may be the time corresponding to the user performing the suspicious click described above.
  • information related to suspicious clicks can be obtained from the click log. It can be understood that the related information of the suspicious click can also be obtained by other means, and the manner in which the present application obtains the relevant information of the suspicious click is not limited.
  • step 302 at least one set of suspicious user groups is determined based on the related information, wherein each set of suspicious user groups clicks on the same set of content within the same time period.
  • the suspicious user group may be determined based on the foregoing related information, wherein there may be one or more groups of suspicious user groups, and each group of suspicious user groups clicks the same group of content in the same time period.
  • a non-parametric clustering algorithm may be used to determine a suspicious user group. Specifically, first, all suspicious clicks are clustered based on the above related information, so that each cluster center The corresponding user group clicks on the same set of content in the same time period. The user group corresponding to each cluster center is then determined as a group of suspicious users.
  • FIG. 4 there is shown a flow 400 of one embodiment of a method for determining a non-cheating user population to be excluded based on keywords of suspicious content clicked by each set of suspicious user groups during the predetermined time period.
  • step 401 keywords of suspicious content clicked by each group of suspicious user groups within the predetermined time period are acquired.
  • the suspicious content is a predetermined content that the suspicious user group clicks within the predetermined time period. It should be noted that users in the suspicious user group may also click other content that does not involve cheating within the above predetermined time period, but these do not involve cheating. The content is not related to suspicious clicks and therefore will not be judged as suspicious.
  • the keyword of the suspicious content is a word that best reflects various characteristics of the suspicious content.
  • the keyword may be a category of an advertisement product (drug), a name of a disease that the drug can treat, a name of a pharmaceutical factory that produces the drug, and the most important chemical component of the drug. Name and so on.
  • the content of the suspicious content may be parsed to obtain related keywords.
  • the related keywords may also be obtained from the name or identification information of the above suspicious content. It can be understood that there may be other ways of obtaining keywords related to suspicious content, and the method for obtaining keywords related to suspicious content is not limited in the present application.
  • step 402 it is determined whether the suspicious content is the same content based on the keyword.
  • different suspicious content may be determined according to keywords corresponding to different suspicious content. Specifically, firstly, among the keywords of a group of suspicious content clicked by each group of suspicious user groups in the predetermined time period, whether the proportion of the similar keywords is greater than or equal to a predetermined ratio. If the proportion of the similar keywords is greater than or equal to a predetermined ratio, it is determined that the suspicious content is the same content.
  • step 403 the suspicious user group corresponding to the suspicious content is determined as a non-cheating user group to be excluded.
  • FIG. 5 a block diagram of one embodiment of an apparatus for detecting click cheating in accordance with the present application is shown.
  • the apparatus 500 of this embodiment includes: a first determining unit 501, a second determining unit 502, a third determining unit 503, and a fourth determining unit 504.
  • the first determining unit 501 is configured to determine, according to the number of times the predetermined content is clicked by the user within a predetermined time period. Suspicious clicks.
  • the second determining unit 502 is configured to determine, according to the suspicious click, at least one group of suspicious users suspected of cheating.
  • the third determining unit 503 is configured to determine, according to the keyword of the suspicious content clicked by the group of the suspicious user within the predetermined time period, the non-cheating user group to be excluded.
  • the fourth determining unit 504 is configured to exclude the non-cheating user group in the suspicious user group to determine a cheating user group.
  • the first determining unit 501 includes an obtaining subunit, a determining subunit, and a determining subunit (not shown).
  • the obtaining subunit is configured to acquire the number of times each predetermined content is clicked by the user in the predetermined time period.
  • the judging subunit is configured to judge whether the number of times each of the predetermined contents is clicked satisfies a predetermined condition.
  • the determining subunit is configured to determine a click corresponding to the predetermined content whose number of clicks meets the predetermined condition as a suspicious click.
  • the determining subunit is configured to: determine whether the number of clicks is greater than or equal to a first predetermined threshold, and less than or equal to a second predetermined threshold. If so, it is determined that the number of times the click is made satisfies a predetermined condition.
  • the second determining unit 502 includes an information acquisition subunit and a user population determination subunit (not shown).
  • the information obtaining subunit is configured to acquire related information of the suspicious click.
  • the user group determining subunit is configured to determine at least one set of suspicious user groups based on the related information, wherein each group of suspicious user groups clicks the same set of content within the same time period.
  • the information related to the suspicious click includes at least one of the following: identification information of the user corresponding to the suspicious click; identification information of the content corresponding to the suspicious click; and corresponding to the suspicious click Moment.
  • the user group determining subunit is configured to: cluster the suspicious clicks based on the related information, so that a user group corresponding to each cluster center clicks in the same time period The same group of content; the user group corresponding to each cluster center is determined as a group of suspicious users.
  • the third determining unit 503 includes a keyword acquisition subunit, a category determination subunit, and a to-be-excluded group determination subunit (not shown).
  • the keyword acquisition subunit is configured to acquire keywords of the suspicious content clicked by the group of the suspicious user group within the predetermined time period.
  • a category determining subunit for determining the ok based on the keyword Whether the content is similar.
  • the to-be-excluded group determining sub-unit is configured to determine the suspicious user group corresponding to the suspicious content as the non-cheating user group to be excluded in response to the suspicious content being the same content.
  • the category determining subunit is configured to: determine whether a proportion of the same type of keywords in the keyword is greater than or equal to a predetermined ratio; if yes, determine that the suspicious content is a similar content.
  • the device 500 may be preset in the server, or may be loaded into the server by downloading or the like. Corresponding units in device 500 can cooperate with units in the server to implement a scheme for detecting click cheating.
  • FIG. 6 a block diagram of a computer system 600 suitable for use in implementing a terminal device or server of an embodiment of the present application is shown.
  • computer system 600 includes a central processing unit (CPU) 601 that can be loaded into a program in random access memory (RAM) 603 according to a program stored in read only memory (ROM) 602 or from storage portion 608. And perform various appropriate actions and processes.
  • RAM random access memory
  • ROM read only memory
  • RAM random access memory
  • various programs and data required for the operation of the system 600 are also stored.
  • the CPU 601, the ROM 602, and the RAM 603 are connected to each other through a bus 604.
  • An input/output (I/O) interface 605 is also coupled to bus 604.
  • the following components are connected to the I/O interface 605: an input portion 606 including a keyboard, a mouse, etc.; an output portion 607 including, for example, a cathode ray tube (CRT), a liquid crystal display (LCD), and the like, and a storage portion 608 including a hard disk or the like. And a communication portion 609 including a network interface card such as a LAN card, a modem, or the like. The communication section 609 performs communication processing via a network such as the Internet.
  • Driver 610 is also coupled to I/O interface 605 as needed.
  • a removable medium 611 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory or the like, is mounted on the drive 610 as needed so that a computer program read therefrom is installed into the storage portion 608 as needed.
  • an embodiment of the present disclosure includes a computer program A sequential product comprising a computer program tangibly embodied on a machine readable medium, the computer program comprising program code for executing the method illustrated in the flowchart.
  • the computer program can be downloaded and installed from the network via communication portion 609, and/or installed from removable media 611.
  • each block of the flowchart or block diagrams can represent a module, a program segment, or a portion of code that includes one or more logic for implementing the specified.
  • Functional executable instructions can also occur in a different order than that illustrated in the drawings. For example, two successively represented blocks may in fact be executed substantially in parallel, and they may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowcharts, and combinations of blocks in the block diagrams and/or flowcharts can be implemented in a dedicated hardware-based system that performs the specified function or operation. Or it can be implemented by a combination of dedicated hardware and computer instructions.
  • the unit modules described in the embodiments of the present application may be implemented by software or by hardware.
  • the described unit modules may also be provided in the processor, for example, which may be described as a processor comprising an operation first determining unit, a second determining unit, a third determining unit and a fourth determining unit.
  • the names of the unit modules do not constitute a limitation on the unit modules themselves in some cases.
  • the first determining unit may also be described as “for determining the number of times the predetermined content is clicked by the user within a predetermined time period. Unit of suspicious clicks.”
  • the present application further provides a computer readable storage medium, which may be a computer readable storage medium included in the apparatus described in the foregoing embodiment, or may exist separately, not A computer readable storage medium that is assembled into a terminal.
  • the computer readable storage medium stores one or more programs that are used by one or more processors to perform the methods described herein for detecting click cheating.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Information Transfer Between Computers (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

L'invention porte sur un procédé et un appareil pour détecter des clics de triche. Le procédé comprend : la détermination, sur la base du nombre de clics réalisés sur des contenus prédéterminés par un utilisateur dans une période de temps prédéterminée, de clics suspicieux (101) ; la détermination, selon les clics suspicieux, d'au moins un groupe d'utilisateurs suspicieux suspectés de tricher (102) ; la détermination, selon des mots-clés des contenus prédéterminés cliqués par chaque groupe d'utilisateurs suspicieux dans la période de temps prédéterminée, de groupes d'utilisateurs ne trichant pas à éliminer (103) ; et l'élimination des groupes d'utilisateurs ne trichant pas dans les groupes d'utilisateurs suspicieux afin de déterminer des groupes d'utilisateurs trichant (104). Des clics sur des contenus prédéterminés peuvent être surveillés, améliorant ainsi l'efficacité de détection et la précision de détection de clics de triche, et réduisant le gaspillage de temps et de ressources.
PCT/CN2015/089545 2015-04-24 2015-09-14 Procédé et appareil pour détecter des clics de triche Ceased WO2016169193A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201510202474.9A CN104765874B (zh) 2015-04-24 2015-04-24 用于检测点击作弊的方法及装置
CN201510202474.9 2015-04-24

Publications (1)

Publication Number Publication Date
WO2016169193A1 true WO2016169193A1 (fr) 2016-10-27

Family

ID=53647701

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/089545 Ceased WO2016169193A1 (fr) 2015-04-24 2015-09-14 Procédé et appareil pour détecter des clics de triche

Country Status (2)

Country Link
CN (1) CN104765874B (fr)
WO (1) WO2016169193A1 (fr)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109034867A (zh) * 2018-06-21 2018-12-18 腾讯科技(深圳)有限公司 点击流量检测方法、装置及存储介质
CN110210886A (zh) * 2018-05-31 2019-09-06 腾讯科技(深圳)有限公司 识别虚假操作方法、装置、服务器、可读存储介质、系统
CN110827094A (zh) * 2019-11-15 2020-02-21 湖南快乐阳光互动娱乐传媒有限公司 广告投放的反作弊方法及系统
CN113592036A (zh) * 2021-08-25 2021-11-02 北京沃东天骏信息技术有限公司 流量作弊行为识别方法、装置及存储介质和电子设备
CN114926221A (zh) * 2022-05-31 2022-08-19 北京奇艺世纪科技有限公司 作弊用户识别方法、装置、电子设备及存储介质
CN120258906A (zh) * 2025-05-22 2025-07-04 北京美数信息科技有限公司 面向互联网的广告投放虚假流量监测方法及系统

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104765874B (zh) * 2015-04-24 2019-03-26 百度在线网络技术(北京)有限公司 用于检测点击作弊的方法及装置
CN106445796B (zh) * 2015-08-04 2021-01-19 腾讯科技(深圳)有限公司 作弊渠道的自动检测方法及装置
CN105354721B (zh) * 2015-09-29 2019-09-06 北京金山安全软件有限公司 一种识别机器操作行为的方法及装置
CN106998336B (zh) * 2016-01-22 2020-07-31 腾讯科技(深圳)有限公司 渠道中的用户检测方法和装置
CN106649527B (zh) * 2016-10-20 2021-02-09 重庆邮电大学 基于Spark Streaming的广告点击异常检测系统及检测方法
CN107168854B (zh) * 2017-06-01 2020-06-30 北京京东尚科信息技术有限公司 互联网广告异常点击检测方法、装置、设备及可读存储介质
CN107229557B (zh) * 2017-06-26 2020-10-20 微鲸科技有限公司 异常点击检测方法及装置、点击量统计方法及装置
CN107566897B (zh) * 2017-07-19 2019-10-15 北京奇艺世纪科技有限公司 一种视频刷量的鉴别方法、装置及电子设备
CN107529093B (zh) * 2017-09-05 2020-05-22 北京奇艺世纪科技有限公司 一种视频文件播放量的检测方法及系统
CN110046910B (zh) * 2018-12-13 2023-04-14 蚂蚁金服(杭州)网络技术有限公司 判断客户通过电子支付平台所进行交易合法性的方法和设备
CN109842619B (zh) * 2019-01-08 2022-07-08 北京百度网讯科技有限公司 用户账号拦截方法和装置
CN110069691B (zh) * 2019-04-29 2021-05-28 百度在线网络技术(北京)有限公司 用于处理点击行为数据的方法和装置
CN112579907B (zh) * 2020-12-25 2023-08-11 北京百度网讯科技有限公司 一种异常任务检测方法、装置、电子设备和存储介质
CN112508630B (zh) * 2021-01-29 2021-05-25 腾讯科技(深圳)有限公司 异常会话群的检测方法、装置、计算机设备和存储介质
CN113179358B (zh) * 2021-04-09 2022-08-09 作业帮教育科技(北京)有限公司 一种题目解答的防作弊方法、装置及电子设备

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101093510A (zh) * 2007-07-25 2007-12-26 北京搜狗科技发展有限公司 一种针对网页作弊的反作弊方法及系统
US20090299967A1 (en) * 2008-06-02 2009-12-03 Microsoft Corporation User advertisement click behavior modeling
CN103390027A (zh) * 2013-06-25 2013-11-13 亿赞普(北京)科技有限公司 一种互联网广告反作弊方法和系统
CN104765874A (zh) * 2015-04-24 2015-07-08 百度在线网络技术(北京)有限公司 用于检测点击作弊的方法及装置

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102289756A (zh) * 2010-06-18 2011-12-21 百度在线网络技术(北京)有限公司 点击有效性的判断方法及其系统
CN103853839A (zh) * 2014-03-18 2014-06-11 北京博雅立方科技有限公司 一种评测广告页面恶意点击疑似度的方法及装置
CN103870572B (zh) * 2014-03-18 2017-07-04 北京博雅立方科技有限公司 一种防御恶意点击广告页面的方法及装置

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101093510A (zh) * 2007-07-25 2007-12-26 北京搜狗科技发展有限公司 一种针对网页作弊的反作弊方法及系统
US20090299967A1 (en) * 2008-06-02 2009-12-03 Microsoft Corporation User advertisement click behavior modeling
CN103390027A (zh) * 2013-06-25 2013-11-13 亿赞普(北京)科技有限公司 一种互联网广告反作弊方法和系统
CN104765874A (zh) * 2015-04-24 2015-07-08 百度在线网络技术(北京)有限公司 用于检测点击作弊的方法及装置

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110210886A (zh) * 2018-05-31 2019-09-06 腾讯科技(深圳)有限公司 识别虚假操作方法、装置、服务器、可读存储介质、系统
CN110210886B (zh) * 2018-05-31 2023-08-22 腾讯科技(深圳)有限公司 识别虚假操作方法、装置、服务器、可读存储介质、系统
CN109034867A (zh) * 2018-06-21 2018-12-18 腾讯科技(深圳)有限公司 点击流量检测方法、装置及存储介质
CN109034867B (zh) * 2018-06-21 2022-10-25 腾讯科技(深圳)有限公司 点击流量检测方法、装置及存储介质
CN110827094A (zh) * 2019-11-15 2020-02-21 湖南快乐阳光互动娱乐传媒有限公司 广告投放的反作弊方法及系统
CN110827094B (zh) * 2019-11-15 2023-05-23 湖南快乐阳光互动娱乐传媒有限公司 广告投放的反作弊方法及系统
CN113592036A (zh) * 2021-08-25 2021-11-02 北京沃东天骏信息技术有限公司 流量作弊行为识别方法、装置及存储介质和电子设备
CN114926221A (zh) * 2022-05-31 2022-08-19 北京奇艺世纪科技有限公司 作弊用户识别方法、装置、电子设备及存储介质
CN120258906A (zh) * 2025-05-22 2025-07-04 北京美数信息科技有限公司 面向互联网的广告投放虚假流量监测方法及系统

Also Published As

Publication number Publication date
CN104765874A (zh) 2015-07-08
CN104765874B (zh) 2019-03-26

Similar Documents

Publication Publication Date Title
WO2016169193A1 (fr) Procédé et appareil pour détecter des clics de triche
US10134058B2 (en) Methods and apparatus for identifying unique users for on-line advertising
JP7422823B2 (ja) 動的アプリケーションコンテンツ分析
US8515966B2 (en) Analyzing queries to generate product intention rules
CN104281622B (zh) 一种社交媒体中的信息推荐方法和装置
JP6386007B2 (ja) ネイティブ・アプリケーションのテスト
JP5577385B2 (ja) コンテンツ配信装置
US9122710B1 (en) Discovery of new business openings using web content analysis
CN109325179A (zh) 一种内容推广的方法及装置
WO2019169978A1 (fr) Procédé et dispositif de recommandation de ressources
CN107291755B (zh) 一种终端推送方法及装置
US12481571B2 (en) Contribution incrementality machine learning models
US11669870B2 (en) Using embedded elements for online content verification
JP2021508400A (ja) 高速かつ安全なコンテンツ提供のためのシステム
US20110238490A1 (en) Auction flighting
US20240037176A1 (en) Using embedded elements for online content verification
US9754211B2 (en) Incrementality modeling
TWI688870B (zh) 用於偵測欺詐的成對用戶-內容提供者之方法與系統
CN111325228B (zh) 一种模型训练方法及装置
US20190087879A1 (en) Marketplace listing analysis systems and methods
US20150302503A1 (en) System and method to customize user experience based on brand resilience data
US9749438B1 (en) Providing a content item for presentation with multiple applications
JP2022527671A (ja) プラットフォームをまたいだコンテンツのミュート
CN113892085B (zh) 限制在客户端设备上对冗余数字组件的提供和显示
US9767196B1 (en) Content selection

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15889680

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15889680

Country of ref document: EP

Kind code of ref document: A1