[go: up one dir, main page]

US11164563B2 - Wake word based on acoustic analysis - Google Patents

Wake word based on acoustic analysis Download PDF

Info

Publication number
US11164563B2
US11164563B2 US16/717,102 US201916717102A US11164563B2 US 11164563 B2 US11164563 B2 US 11164563B2 US 201916717102 A US201916717102 A US 201916717102A US 11164563 B2 US11164563 B2 US 11164563B2
Authority
US
United States
Prior art keywords
verbal
pws
sws
sound
circuitry
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/717,102
Other versions
US20210183370A1 (en
Inventor
Fabrice Saadoun
Nissim Gean
Kenneth A. Haas
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Motorola Solutions Inc
Original Assignee
Motorola Solutions Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Motorola Solutions Inc filed Critical Motorola Solutions Inc
Priority to US16/717,102 priority Critical patent/US11164563B2/en
Assigned to MOTOROLA SOLUTIONS INC. reassignment MOTOROLA SOLUTIONS INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HAAS, KENNETH A., SAADOUN, FABRICE, GEAN, Nissim
Priority to PCT/US2020/061735 priority patent/WO2021126475A1/en
Publication of US20210183370A1 publication Critical patent/US20210183370A1/en
Application granted granted Critical
Publication of US11164563B2 publication Critical patent/US11164563B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B13/00Burglar, theft or intruder alarms
    • G08B13/16Actuation by interference with mechanical vibrations in air or other fluid
    • G08B13/1654Actuation by interference with mechanical vibrations in air or other fluid using passive vibration detection systems
    • G08B13/1672Actuation by interference with mechanical vibrations in air or other fluid using passive vibration detection systems using sonic detecting means, e.g. a microphone operating in the audio frequency range
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B25/00Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems
    • G08B25/002Generating a prealarm to the central station
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B25/00Alarm systems in which the location of the alarm condition is signalled to a central station, e.g. fire or police telegraphic systems
    • G08B25/008Alarm setting and unsetting, i.e. arming or disarming of the security system
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • GPHYSICS
    • G08SIGNALLING
    • G08BSIGNALLING OR CALLING SYSTEMS; ORDER TELEGRAPHS; ALARM SYSTEMS
    • G08B29/00Checking or monitoring of signalling or alarm systems; Prevention or correction of operating errors, e.g. preventing unauthorised operation
    • G08B29/18Prevention or correction of operating errors
    • G08B29/185Signal analysis techniques for reducing or preventing false alarms or for enhancing the reliability of the system
    • G08B29/188Data fusion; cooperative systems, e.g. voting among different detectors
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L2015/088Word spotting

Definitions

  • acoustic analysis will be utilized to determine whether or not particular public-safety events happen.
  • a gunshot detection system called Shotspotter®, manufactured by SSI, Inc. attempts to determine whether or not gunshots were detected.
  • Such a system uses multiple acoustic sensors that detect when gunfire occurs, providing comprehensive outdoor coverage over complex urban geographies. When there is a shooting, sensors report this information and software triangulates and pinpoints the precise location of each round fired.
  • Shotspotter® system accurately detects and locates gunshots, it would be beneficial if a system existed that aided in detecting other types of public-safety events. Unfortunately, using a system like Shotspotter® for multi-event detection would require very high computation and power consumption, which would be prohibitively expensive for large, area-wide event detection.
  • FIG. 1 is a general operating environment for the present invention.
  • FIG. 2 is a block diagram of the server of FIG. 1 .
  • FIG. 3 is a block diagram of an acoustic monitoring device as shown in FIG. 1 .
  • FIG. 4 is a flow chart showing operation of the device of FIG. 3 and the server of FIG. 2 .
  • a sound-enabled detection system responds to a wakeup sound, and notifies a central server. Based on which wakeup sound is detected; the system's configuration is modified accordingly. More particularly, at least one sensor in the system has its wakeup sound reconfigured to “wake up” upon detection of a second wakeup sound when the first wakeup sound is detected.
  • utilizing wakeup “sounds” in low-power modes of operation are commonly known in the area of personal digital assistants.
  • Many virtual assistant systems perform continuous recognition that targets some particular wakeup phrase. This process is known as phrase spotting.
  • phrase spotting Typically, a low-power, low complexity phrase spotting technique is utilized to detect a particular phrase. The low complexity phrase spotting is required to prevent exes power consumption and to not behave intrusively by responding to utterances not intended for it.
  • devices upon detection of a wakeup sound, devices will send a notification that the sound has been detected, and then have their wakeup sound changed, and remaining/returning to a low-power, less-computationally-intensive mode.
  • This allows edge devices to aide in detecting public-safety events with more confidence without having to waste precious resources.
  • the system may also switch to a more computationally-intensive high-power mode, recognizing a much larger vocabulary of sounds to process more sophisticated analysis.
  • the more-computationally-intensive mode processes sounds for a predetermined period of time, at which point it reverts to the low-power sound-spotting state, typically after a short refractory period.
  • the device may change its wakeup sound.
  • PWS Primary Wakeup Sound
  • a gunshot may be considered a primary wakeup sound that a low-power, low-complexity detector is configured to wake upon.
  • a server Upon detecting the primary wakeup sound, a server will be notified, and multiple devices will be reconfigured to wake upon detection of a Secondary Wakeup Sound (SWS) used to aide in confirming that a particular public-safety event has happened.
  • SWS Secondary Wakeup Sound
  • Secondary Wakeup Sound a sound that acoustic detectors are configured to “wake” upon detecting.
  • a scream may be considered a secondary wakeup sound that a low-power, low-complexity detector is configured to wake upon.
  • a server Upon detecting the secondary wakeup sound, a server will be notified, and multiple devices will be reconfigured to wake upon detection of other Secondary Wakeup Sounds used to aide in confirming that a particular public-safety event has happened.
  • the SWS is a function of the PWS. In other words, the SWS is mapped to the PWS.
  • the mapping process preferably comprises an operation that associates each element of a given set (the domain) with one or more elements of a second set (the range).
  • the PWSs comprises the domain, while the SWSs comprise the range.
  • the mapping may be explicit based on predefined rules, or the mapping may be trained via machine learning. The mapping produces the SWSs.
  • a server will map the detected PWS to the SWS. More specifically, if PWS (x) is an element of a group of possible PWSs X (i.e., x ⁇ X), we say that f “takes the value” or “maps to” f(x) at x.
  • the set X is called the domain of the function f.
  • the mapping may be explicit based on predefined rules, or the mapping may be trained via, for example, machine learning.
  • acoustic monitoring devices will be configured to detect PWSs. Once detected, the information will be sent to a server, which can decide SWSs the device should detect. Various devices will be reconfigured to detect the SWSs. If there are more than one device at the same location, the server may configure devices to wake upon differing SWSs. For example, if a gunshot was detected as a PWS, then a first device may be configured to wake upon a human screaming, while another device may be configured to wake upon the utterance of “help”. Once woken, devices will report the detection of PWSs and SWSs to the server. As more and more PWSs and SWSs are detected by the edge devices, the situational awareness will be more accurate.
  • FIG. 1 is a general operational environment for the present invention.
  • the operating environment includes acoustic detection circuitry 105 which comprises any device capable of acoustic monitoring for wakeup sounds.
  • acoustic detection circuitry 105 which comprises any device capable of acoustic monitoring for wakeup sounds.
  • such devices include various microphones, smartphones, police radios, acoustic sensors mounted on vehicles, . . . , etc.
  • Devices 105 are preferably spread through a wide area (e.g., citywide) to form a wide-area acoustic detection system.
  • Devices 105 may also comprise stationary sensors located on, for example, light poles within an area. The sensors are specially designed to detect the report wakeup sounds being detected.
  • the operational environment additionally comprises server 109 .
  • Server 109 serves as an event determining system that provides information on the detection of particular wakeup sounds to a user, and configures devices with various wake sounds that may be dependent upon a prior wake sound being detected.
  • Network 106 is provided to facilitate communications between devices 105 and server 109 .
  • Network 106 may operate according to an industry standard land mobile radio (LMR) or cellular protocol such as, for example, the Project 25 (P25) standard defined by the Association of Public Safety Communications Officials International (APCO), the TETRA standard defined by the European Telecommunication Standards Institute (ETSI), the Digital Private Mobile Radio (dPMR) standard also defined by the ETSI, the Digital Mobile Radio (DMR) standard also defined by the ETSI, the Long Term Evolution (LTE) (including LTE-Advanced or LTE-Advanced Pro compliant with, for example, the 3GPP TS 36 specification series), or the 5G (including a network architecture compliant with, for example, the 3GPP TS 23 specification series and a new radio (NR) air interface compliant with the 3GPP TS 38 specification series) standard, among other possibilities, and over which multimedia broadcast multicast services (MBMS), single site point-to-multipoint (SC-PTM) services, or Mission Critical Push-
  • Acoustic detection devices 105 are configured to operate in a low-power, low complexity mode of operation, listening for a PWS. In order to accomplish this task, dedicated circuitry and/or software is utilized for sound detection. Various chip and software vendors provide such circuitry optimized for phrase spotting.
  • an event notification is sent through network 106 to server 109 .
  • the event notification may comprise the wakeup sound, or a description of the wakeup sound that caused the event notification to be sent to server 109 .
  • server 109 may reconfigure at least one device 105 to change its wakeup sound from a first wakeup sound to a second wakeup sound.
  • Devices that sent notifications may then enter a low-power state and listen for the second wakeup sound. If the second wakeup sound is detected, the device will again send an event notification through network 106 to server 109 , notifying server 109 that the second wakeup sound has been detected.
  • the devices may again be reconfigured with differing wakeup sounds.
  • FIG. 2 is a block diagram of server 109 .
  • server 109 comprises logic circuitry 201 , network interface 203 , database 204 , and graphical-user interface (GUI) 206 .
  • GUI graphical-user interface
  • server 109 may include more, fewer, or different components. Regardless, all components are connected via common data busses as known in the art.
  • Logic circuitry 201 comprises a digital signal processor (DSP), general purpose microprocessor, a programmable logic device, or application specific integrated circuit (ASIC) and is configured to receive an event notification from a device 105 and notify a user (via GUI) 206 that a particular wakeup sound has been detected.
  • Logic circuitry 201 is also configured to reconfigure devices 105 to listen for an SWS when a PWS has been detected.
  • Database 204 comprises standard memory (such as RAM, ROM, . . . , etc) and serves to store PWSs and their associated SWSs.
  • database 204 may comprise a table of PWSs.
  • the device when a device “hears” the PWS, the device may be reconfigured to “listen” for an associated SWS.
  • Multiple SWSs may be rank ordered and associated with a PWS, so, for example, if multiple devices 105 are at a particular incident scene, each detecting the PWS, the multiple devices may be reconfigured to listen to a unique SWS, so that each device on scene may listen for different SWSs.
  • GUI 206 provides a man/machine interface for providing information about detected PWSs and SWSs to a user and displaying information.
  • GUI 206 may provide a way of conveying (e.g., displaying) information received from processor 201 .
  • Part of this information may comprise a identification of a PWS/SWS detected, a location of the PWS/SWS detected, a recording of the PWS/SWS detected, an identification of the device which detected the PWS/SWS, . . . , etc.
  • GUI 506 may comprise any combination of a touch screen, a computer screen, a keyboard, speaker, or any other interface needed to provide information to the user.
  • network interface 203 provides processing, modulating, and transceiver elements that are operable in accordance with any one or more standard land mobile radio (LMR) or cellular protocols described above, wherein some of the functionality of the processing, modulating, and transceiver elements may be performed by means of the logic circuitry 201 through programmed logic such as software.
  • LMR standard land mobile radio
  • server 109 comprises a network interface receiving an indication that a primary wakeup sound (PWS) was detected by a device, and logic circuitry determining a secondary wakeup sound (SWS) based on receiving the PWS, and reconfiguring the device to wake upon the detection of the SWS.
  • PWS primary wakeup sound
  • SWS secondary wakeup sound
  • the SWS is based on the PWS.
  • a database may be provided comprising a plurality of PWSs and their associated SWSs, wherein the logic circuitry accesses the database to determine an SWS from a detected PWS.
  • a GUI is provided coupled to the logic circuitry and wherein the logic circuitry instructs the GUI to notify a user that the PWS was detected.
  • the logic circuitry may receive a location of the device and base the SWS is on the location of the device.
  • the logic circuitry may be configured to determine if other devices are near the device, and also base the SWS upon whether other devices are near the device.
  • FIG. 3 is a block diagram of a device 105 as shown in FIG. 1 .
  • device 105 comprises microphone 305 and sound-spotting circuitry 302 .
  • Sound-spotting circuitry 302 preferably comprises low-cost, low-complexity microprocessor specifically designed for detecting a wake sound, and “waking” logic circuitry 301 when a particular sound has been detected. Sound-spotting circuitry may be configured with the particular sound to alert, or wake logic circuitry 301 as described in U.S. Pat. No. 10,217,453 entitled, VIRTUAL ASSISTANT CONFIGURED BY SELECTION OF WAKE-UP PHRASE, incorporated by reference herein.
  • logic circuitry 301 and sound-spotting circuitry 302 may be co-located within a single processor that operates in a low-power mode until a wake sound is detected.
  • sound-spotting circuitry 302 will be “listening” for a particular PWS. Once the PWS is detected, sound-spotting circuitry 302 will notify (possibly waking) microprocessor 301 that the PWS was detected. This causes logic circuitry 301 to notify server 109 of the event via network interface 303 . In response, logic circuitry 301 may receive an SWS through interface 203 and reprogram/reconfigure circuitry 302 to “listen” for the SWS.
  • logic circuitry 301 may reconfigure circuitry 302 with a second wake sound (SWS) by accessing database 304 and determining an SWS associated with the detected PWS.
  • Database 304 is similar to the database described above with reference to FIG. 2 , and comprises a list of PWSs and their associated SWSs.
  • the PWS and the SWS may be based on a location.
  • devices may be configured with location-finding circuitry 306 in order to ascertain the device's location.
  • location-finding circuitry may comprise a standard Global Positioning System (GPS) receiver, however in alternate embodiments other forms of location circuitry may be utilized as well.
  • GPS Global Positioning System
  • Each device 105 may periodically provide their location to server 109 , as is known in the art.
  • sound-spotting circuitry 302 may be configured with a PWS or SWS based on a location of device 105 . So, for example, a first PWS may be used when device 105 is located in a first geographical area, and a second PWS may be used with device 105 is in a second geographical area.
  • circuitry 302 may be configured with a PWS or SWS that detects gunfire.
  • circuitry 302 may be configured with a PWS or SWS that detects people fighting.
  • the PWS or SWS (which may include wake words) may be configured based on public-safety statistics for a particular area, so that if a particular geographic area has a high incidence of a particular type of event, the PWS and SWS may be tailored in device 105 to detect the particular type of event.
  • circuitry 302 may be configured with a PWS or SWS so that no two devices share the same PWS or SWS. More particularly, if three devices 105 are located on a same block, a first device 105 may be configured with a first PWS, while a second device may be configured with a second PWS, and a third device may be configured with a third PWS. Thus, logic circuitry 301 , or logic circuitry 201 will determine if multiple devices 105 are within a particular geographic area, or are within a predetermined distance from each other. If this is the case, then devices 105 will be configured such that at least two devices 105 will have differing PWS or SWS.
  • the configuration of PWSs and SWSs based on location will require server 109 or devices 105 to know the location of all devices 105 . This can be accomplished by devices 105 sharing their locations with server 109 , and server 109 reporting the locations of each device to devices 105 .
  • devices 105 comprise sound-spotting circuitry configured with a primary wakeup sound (PWS), and logic circuitry configured to be woken upon the sound-spotting circuitry detecting the PWS, and notify a user that the PWS was detected.
  • the logic circuitry is also configured to reconfigure the sound-spotting circuitry with a secondary wakeup sound (SWS) when the PWS has been detected by the sound-spotting circuitry.
  • SWS secondary wakeup sound
  • a database may be provided that comprises a plurality of PWSs and their associated SWSs. The logic circuitry may access the database to determine an SWS from a detected PWS.
  • a network interface is provided so that the logic circuitry may notify the user that the PWS was detected by notifying a server via the network interface.
  • Location-finding circuitry may be provided to determine a location of a device, and the SWS is may also be based on the location of the device.
  • the logic circuitry can be configured to determine if other devices are near the device, and wherein the SWS may also be based upon whether other devices are near the device.
  • FIG. 4 is a flow chart showing operation of device 105 and server 109 .
  • the logic flow begins at step 401 where sound-spotting circuitry is configured within a device with a primary wakeup sound (PWS).
  • PWS primary wakeup sound
  • sound spotting circuitry wakes upon detecting the PWS.
  • SWS secondary wakeup sound
  • references to specific implementation embodiments such as “circuitry” may equally be accomplished via either on general purpose computing apparatus (e.g., CPU) or specialized processing apparatus (e.g., DSP) executing software instructions stored in non-transitory computer-readable memory.
  • general purpose computing apparatus e.g., CPU
  • specialized processing apparatus e.g., DSP
  • DSP digital signal processor
  • a includes . . . a”, “contains . . . a” does not, without more constraints, preclude the existence of additional identical elements in the process, method, article, or apparatus that comprises, has, includes, contains the element.
  • the terms “a” and “an” are defined as one or more unless explicitly stated otherwise herein.
  • the terms “substantially”, “essentially”, “approximately”, “about” or any other version thereof, are defined as being close to as understood by one of ordinary skill in the art, and in one non-limiting embodiment the term is defined to be within 10%, in another embodiment within 5%, in another embodiment within 1% and in another embodiment within 0.5%.
  • the term “coupled” as used herein is defined as connected, although not necessarily directly and not necessarily mechanically.
  • a device or structure that is “configured” in a certain way is configured in at least that way, but may also be configured in ways that are not listed.
  • processors such as microprocessors, digital signal processors, customized processors and field programmable gate arrays (FPGAs) and unique stored program instructions (including both software and firmware) that control the one or more processors to implement, in conjunction with certain non-processor circuits, some, most, or all of the functions of the method and/or apparatus described herein.
  • processors or “processing devices” such as microprocessors, digital signal processors, customized processors and field programmable gate arrays (FPGAs) and unique stored program instructions (including both software and firmware) that control the one or more processors to implement, in conjunction with certain non-processor circuits, some, most, or all of the functions of the method and/or apparatus described herein.
  • FPGAs field programmable gate arrays
  • unique stored program instructions including both software and firmware
  • an embodiment can be implemented as a computer-readable storage medium having computer readable code stored thereon for programming a computer (e.g., comprising a processor) to perform a method as described and claimed herein.
  • Examples of such computer-readable storage mediums include, but are not limited to, a hard disk, a CD-ROM, an optical storage device, a magnetic storage device, a ROM (Read Only Memory), a PROM (Programmable Read Only Memory), an EPROM (Erasable Programmable Read Only Memory), an EEPROM (Electrically Erasable Programmable Read Only Memory) and a Flash memory.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Acoustics & Sound (AREA)
  • Human Computer Interaction (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Business, Economics & Management (AREA)
  • Emergency Management (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Security & Cryptography (AREA)
  • Telephone Function (AREA)

Abstract

A sound-enabled detection system responds to a wakeup sound, and notifies a central server. Based on which wakeup sound is detected; the system's configuration is modified accordingly. More particularly, at least one sensor in the system has its wakeup sound reconfigured to “wake up” upon detection of a second wakeup sound when the first wakeup sound is detected.

Description

BACKGROUND OF THE INVENTION
Oftentimes acoustic analysis will be utilized to determine whether or not particular public-safety events happen. For example, a gunshot detection system called Shotspotter®, manufactured by SSI, Inc. attempts to determine whether or not gunshots were detected. Such a system uses multiple acoustic sensors that detect when gunfire occurs, providing comprehensive outdoor coverage over complex urban geographies. When there is a shooting, sensors report this information and software triangulates and pinpoints the precise location of each round fired.
While the Shotspotter® system accurately detects and locates gunshots, it would be beneficial if a system existed that aided in detecting other types of public-safety events. Unfortunately, using a system like Shotspotter® for multi-event detection would require very high computation and power consumption, which would be prohibitively expensive for large, area-wide event detection.
BRIEF DESCRIPTION OF THE SEVERAL VIEWS OF THE DRAWINGS
The accompanying figures where like reference numerals refer to identical or functionally similar elements throughout the separate views, and which together with the detailed description below are incorporated in and form part of the specification, serve to further illustrate various embodiments and to explain various principles and advantages all in accordance with the present invention.
FIG. 1 is a general operating environment for the present invention.
FIG. 2 is a block diagram of the server of FIG. 1.
FIG. 3 is a block diagram of an acoustic monitoring device as shown in FIG. 1.
FIG. 4 is a flow chart showing operation of the device of FIG. 3 and the server of FIG. 2.
Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions and/or relative positioning of some of the elements in the figures may be exaggerated relative to other elements to help to improve understanding of various embodiments of the present invention. Also, common but well-understood elements that are useful or necessary in a commercially feasible embodiment are often not depicted in order to facilitate a less obstructed view of these various embodiments of the present invention. It will further be appreciated that certain actions and/or steps may be described or depicted in a particular order of occurrence while those skilled in the art will understand that such specificity with respect to sequence is not actually required.
DETAILED DESCRIPTION
In order to address the above-mentioned need, a method and apparatus for predicting an event based on acoustic analysis is provided herein. A sound-enabled detection system responds to a wakeup sound, and notifies a central server. Based on which wakeup sound is detected; the system's configuration is modified accordingly. More particularly, at least one sensor in the system has its wakeup sound reconfigured to “wake up” upon detection of a second wakeup sound when the first wakeup sound is detected.
Expanding on the above, utilizing wakeup “sounds” in low-power modes of operation are commonly known in the area of personal digital assistants. Many virtual assistant systems perform continuous recognition that targets some particular wakeup phrase. This process is known as phrase spotting. Typically, a low-power, low complexity phrase spotting technique is utilized to detect a particular phrase. The low complexity phrase spotting is required to prevent exes power consumption and to not behave intrusively by responding to utterances not intended for it.
Various commercial implementations detect the phrases, “Hey, Siri”, “OK, Google”, and “Alexa”. Various chip and software vendors, provide devices and algorithms optimized for phrase spotting. When the system recognizes a wakeup phrase, the system switches to a more computationally-intensive high-power mode, recognizing a much larger vocabulary to process more sophisticated utterances. The more-computationally-intensive mode processes speech until it detects an end-of-utterance condition, at which point it reverts to the low-power phrase-spotting state, typically after a short refractory period.
In the preferred embodiment of the present invention, upon detection of a wakeup sound, devices will send a notification that the sound has been detected, and then have their wakeup sound changed, and remaining/returning to a low-power, less-computationally-intensive mode. This allows edge devices to aide in detecting public-safety events with more confidence without having to waste precious resources. When the system recognizes a wakeup sound, the system may also switch to a more computationally-intensive high-power mode, recognizing a much larger vocabulary of sounds to process more sophisticated analysis. The more-computationally-intensive mode processes sounds for a predetermined period of time, at which point it reverts to the low-power sound-spotting state, typically after a short refractory period. As mentioned, upon reverting to the low-power sound-spotting state, the device may change its wakeup sound.
In order to better explain operation of the above-described devices, the following definitions are provided:
Primary Wakeup Sound (PWS)—a sound that acoustic detectors are configured to “wake” upon detecting. For example, a gunshot may be considered a primary wakeup sound that a low-power, low-complexity detector is configured to wake upon. Upon detecting the primary wakeup sound, a server will be notified, and multiple devices will be reconfigured to wake upon detection of a Secondary Wakeup Sound (SWS) used to aide in confirming that a particular public-safety event has happened.
Secondary Wakeup Sound (SWS)—a sound that acoustic detectors are configured to “wake” upon detecting. For example, a scream may be considered a secondary wakeup sound that a low-power, low-complexity detector is configured to wake upon. Upon detecting the secondary wakeup sound, a server will be notified, and multiple devices will be reconfigured to wake upon detection of other Secondary Wakeup Sounds used to aide in confirming that a particular public-safety event has happened. The SWS is a function of the PWS. In other words, the SWS is mapped to the PWS.
The mapping process preferably comprises an operation that associates each element of a given set (the domain) with one or more elements of a second set (the range). The PWSs comprises the domain, while the SWSs comprise the range. The mapping may be explicit based on predefined rules, or the mapping may be trained via machine learning. The mapping produces the SWSs.
A server will map the detected PWS to the SWS. More specifically, if PWS (x) is an element of a group of possible PWSs X (i.e., x∈X), we say that f “takes the value” or “maps to” f(x) at x. The set X is called the domain of the function f. The set of possible outcomes of y=f(x) is called the range. The mapping may be explicit based on predefined rules, or the mapping may be trained via, for example, machine learning. The mapping process preferably comprises an operation that associates each element of a given set (the domain, PWSs) with one or more elements of a second set (the range, SWSs) such that y=f(x), or SWS=Function(PWS).
During operation, acoustic monitoring devices will be configured to detect PWSs. Once detected, the information will be sent to a server, which can decide SWSs the device should detect. Various devices will be reconfigured to detect the SWSs. If there are more than one device at the same location, the server may configure devices to wake upon differing SWSs. For example, if a gunshot was detected as a PWS, then a first device may be configured to wake upon a human screaming, while another device may be configured to wake upon the utterance of “help”. Once woken, devices will report the detection of PWSs and SWSs to the server. As more and more PWSs and SWSs are detected by the edge devices, the situational awareness will be more accurate.
FIG. 1 is a general operational environment for the present invention. As shown, the operating environment includes acoustic detection circuitry 105 which comprises any device capable of acoustic monitoring for wakeup sounds. As shown in FIG. 1, such devices include various microphones, smartphones, police radios, acoustic sensors mounted on vehicles, . . . , etc. Devices 105 are preferably spread through a wide area (e.g., citywide) to form a wide-area acoustic detection system. Devices 105 may also comprise stationary sensors located on, for example, light poles within an area. The sensors are specially designed to detect the report wakeup sounds being detected.
The operational environment additionally comprises server 109. Server 109 serves as an event determining system that provides information on the detection of particular wakeup sounds to a user, and configures devices with various wake sounds that may be dependent upon a prior wake sound being detected.
Network 106 is provided to facilitate communications between devices 105 and server 109. Network 106 may operate according to an industry standard land mobile radio (LMR) or cellular protocol such as, for example, the Project 25 (P25) standard defined by the Association of Public Safety Communications Officials International (APCO), the TETRA standard defined by the European Telecommunication Standards Institute (ETSI), the Digital Private Mobile Radio (dPMR) standard also defined by the ETSI, the Digital Mobile Radio (DMR) standard also defined by the ETSI, the Long Term Evolution (LTE) (including LTE-Advanced or LTE-Advanced Pro compliant with, for example, the 3GPP TS 36 specification series), or the 5G (including a network architecture compliant with, for example, the 3GPP TS 23 specification series and a new radio (NR) air interface compliant with the 3GPP TS 38 specification series) standard, among other possibilities, and over which multimedia broadcast multicast services (MBMS), single site point-to-multipoint (SC-PTM) services, or Mission Critical Push-to-talk (MCPTT) services may be provided, or over which an open mobile alliance (OMA) push to talk (PTT) over cellular (OMA-PoC), a voice over IP (VoIP), or a PTT over IP (PoIP) application may be implemented.
Acoustic detection devices 105 are configured to operate in a low-power, low complexity mode of operation, listening for a PWS. In order to accomplish this task, dedicated circuitry and/or software is utilized for sound detection. Various chip and software vendors provide such circuitry optimized for phrase spotting.
Once devices 105 detects a wake-up sound, an event notification is sent through network 106 to server 109. The event notification may comprise the wakeup sound, or a description of the wakeup sound that caused the event notification to be sent to server 109. In response, server 109 may reconfigure at least one device 105 to change its wakeup sound from a first wakeup sound to a second wakeup sound. Devices that sent notifications may then enter a low-power state and listen for the second wakeup sound. If the second wakeup sound is detected, the device will again send an event notification through network 106 to server 109, notifying server 109 that the second wakeup sound has been detected. The devices may again be reconfigured with differing wakeup sounds.
FIG. 2 is a block diagram of server 109. As shown, server 109 comprises logic circuitry 201, network interface 203, database 204, and graphical-user interface (GUI) 206. In other implementations, server 109 may include more, fewer, or different components. Regardless, all components are connected via common data busses as known in the art.
Logic circuitry 201 comprises a digital signal processor (DSP), general purpose microprocessor, a programmable logic device, or application specific integrated circuit (ASIC) and is configured to receive an event notification from a device 105 and notify a user (via GUI) 206 that a particular wakeup sound has been detected. Logic circuitry 201 is also configured to reconfigure devices 105 to listen for an SWS when a PWS has been detected.
Database 204 is provided. Database 204 comprises standard memory (such as RAM, ROM, . . . , etc) and serves to store PWSs and their associated SWSs. For example, database 204 may comprise a table of PWSs. For each PWS, there may comprise at least one SWS associated with the PWS. As discussed, when a device “hears” the PWS, the device may be reconfigured to “listen” for an associated SWS. Multiple SWSs may be rank ordered and associated with a PWS, so, for example, if multiple devices 105 are at a particular incident scene, each detecting the PWS, the multiple devices may be reconfigured to listen to a unique SWS, so that each device on scene may listen for different SWSs.
GUI 206 provides a man/machine interface for providing information about detected PWSs and SWSs to a user and displaying information. For example, GUI 206 may provide a way of conveying (e.g., displaying) information received from processor 201. Part of this information may comprise a identification of a PWS/SWS detected, a location of the PWS/SWS detected, a recording of the PWS/SWS detected, an identification of the device which detected the PWS/SWS, . . . , etc. In order to provide the above features (and additional features), GUI 506 may comprise any combination of a touch screen, a computer screen, a keyboard, speaker, or any other interface needed to provide information to the user.
Finally, network interface 203 provides processing, modulating, and transceiver elements that are operable in accordance with any one or more standard land mobile radio (LMR) or cellular protocols described above, wherein some of the functionality of the processing, modulating, and transceiver elements may be performed by means of the logic circuitry 201 through programmed logic such as software.
Thus, as shown, server 109 comprises a network interface receiving an indication that a primary wakeup sound (PWS) was detected by a device, and logic circuitry determining a secondary wakeup sound (SWS) based on receiving the PWS, and reconfiguring the device to wake upon the detection of the SWS. As discussed, the SWS is based on the PWS.
As discussed, the logic circuitry can determine the SWS by mapping the PWS to the SWS such that SWS=Function(PWS). Additionally, a database may be provided comprising a plurality of PWSs and their associated SWSs, wherein the logic circuitry accesses the database to determine an SWS from a detected PWS.
A GUI is provided coupled to the logic circuitry and wherein the logic circuitry instructs the GUI to notify a user that the PWS was detected.
As discussed, the logic circuitry may receive a location of the device and base the SWS is on the location of the device. In addition, the logic circuitry may be configured to determine if other devices are near the device, and also base the SWS upon whether other devices are near the device.
FIG. 3 is a block diagram of a device 105 as shown in FIG. 1. In addition to those components shown in FIG. 2, device 105 comprises microphone 305 and sound-spotting circuitry 302. Sound-spotting circuitry 302 preferably comprises low-cost, low-complexity microprocessor specifically designed for detecting a wake sound, and “waking” logic circuitry 301 when a particular sound has been detected. Sound-spotting circuitry may be configured with the particular sound to alert, or wake logic circuitry 301 as described in U.S. Pat. No. 10,217,453 entitled, VIRTUAL ASSISTANT CONFIGURED BY SELECTION OF WAKE-UP PHRASE, incorporated by reference herein.
It should be noted that in alternate embodiments of the present invention, logic circuitry 301 and sound-spotting circuitry 302 may be co-located within a single processor that operates in a low-power mode until a wake sound is detected.
During operation, utilizing microphone 305, sound-spotting circuitry 302 will be “listening” for a particular PWS. Once the PWS is detected, sound-spotting circuitry 302 will notify (possibly waking) microprocessor 301 that the PWS was detected. This causes logic circuitry 301 to notify server 109 of the event via network interface 303. In response, logic circuitry 301 may receive an SWS through interface 203 and reprogram/reconfigure circuitry 302 to “listen” for the SWS.
In an alternate embodiment of the present invention, logic circuitry 301 may reconfigure circuitry 302 with a second wake sound (SWS) by accessing database 304 and determining an SWS associated with the detected PWS. Database 304 is similar to the database described above with reference to FIG. 2, and comprises a list of PWSs and their associated SWSs.
It should be noted that in one embodiment of the present invention the PWS and the SWS may be based on a location. In this embodiment, devices may be configured with location-finding circuitry 306 in order to ascertain the device's location. Such location-finding circuitry may comprise a standard Global Positioning System (GPS) receiver, however in alternate embodiments other forms of location circuitry may be utilized as well. Each device 105 may periodically provide their location to server 109, as is known in the art.
Whether device 105 is determining the PWS or SWS, or whether server 109 is determining the PWS or SWS, sound-spotting circuitry 302 may be configured with a PWS or SWS based on a location of device 105. So, for example, a first PWS may be used when device 105 is located in a first geographical area, and a second PWS may be used with device 105 is in a second geographical area.
Expanding on the above, if device 105 is located in an area with a high amount of gunfire, then circuitry 302 may be configured with a PWS or SWS that detects gunfire. As a further example, if device 105 is located in an area where large crowds gather, then circuitry 302 may be configured with a PWS or SWS that detects people fighting. As is evident, the PWS or SWS (which may include wake words) may be configured based on public-safety statistics for a particular area, so that if a particular geographic area has a high incidence of a particular type of event, the PWS and SWS may be tailored in device 105 to detect the particular type of event.
Additionally, if multiple devices 105 are co-located in a same geographic area, circuitry 302 may be configured with a PWS or SWS so that no two devices share the same PWS or SWS. More particularly, if three devices 105 are located on a same block, a first device 105 may be configured with a first PWS, while a second device may be configured with a second PWS, and a third device may be configured with a third PWS. Thus, logic circuitry 301, or logic circuitry 201 will determine if multiple devices 105 are within a particular geographic area, or are within a predetermined distance from each other. If this is the case, then devices 105 will be configured such that at least two devices 105 will have differing PWS or SWS.
The configuration of PWSs and SWSs based on location will require server 109 or devices 105 to know the location of all devices 105. This can be accomplished by devices 105 sharing their locations with server 109, and server 109 reporting the locations of each device to devices 105.
With the above in mind, devices 105 comprise sound-spotting circuitry configured with a primary wakeup sound (PWS), and logic circuitry configured to be woken upon the sound-spotting circuitry detecting the PWS, and notify a user that the PWS was detected. The logic circuitry is also configured to reconfigure the sound-spotting circuitry with a secondary wakeup sound (SWS) when the PWS has been detected by the sound-spotting circuitry. As discussed, the SWS is based on the PWS.
As discussed, the logic circuitry may determine the SWS by mapping the PWS to the SWS such that SWS=Function(PWS). Also, a database may be provided that comprises a plurality of PWSs and their associated SWSs. The logic circuitry may access the database to determine an SWS from a detected PWS.
A network interface is provided so that the logic circuitry may notify the user that the PWS was detected by notifying a server via the network interface.
Location-finding circuitry may be provided to determine a location of a device, and the SWS is may also be based on the location of the device.
The logic circuitry can be configured to determine if other devices are near the device, and wherein the SWS may also be based upon whether other devices are near the device.
FIG. 4 is a flow chart showing operation of device 105 and server 109. The logic flow begins at step 401 where sound-spotting circuitry is configured within a device with a primary wakeup sound (PWS). At step 403 sound spotting circuitry wakes upon detecting the PWS. At step 405, the sound-spotting circuitry is reconfigured with a secondary wakeup sound (SWS) when the PWS has been detected by the sound-spotting circuitry.
In the foregoing specification, specific embodiments have been described. However, one of ordinary skill in the art appreciates that various modifications and changes can be made without departing from the scope of the invention as set forth in the claims below. Accordingly, the specification and figures are to be regarded in an illustrative rather than a restrictive sense, and all such modifications are intended to be included within the scope of present teachings.
Those skilled in the art will further recognize that references to specific implementation embodiments such as “circuitry” may equally be accomplished via either on general purpose computing apparatus (e.g., CPU) or specialized processing apparatus (e.g., DSP) executing software instructions stored in non-transitory computer-readable memory. It will also be understood that the terms and expressions used herein have the ordinary technical meaning as is accorded to such terms and expressions by persons skilled in the technical field as set forth above except where different specific meanings have otherwise been set forth herein.
The benefits, advantages, solutions to problems, and any element(s) that may cause any benefit, advantage, or solution to occur or become more pronounced are not to be construed as a critical, required, or essential features or elements of any or all the claims. The invention is defined solely by the appended claims including any amendments made during the pendency of this application and all equivalents of those claims as issued.
Moreover in this document, relational terms such as first and second, top and bottom, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. The terms “comprises,” “comprising,” “has”, “having,” “includes”, “including,” “contains”, “containing” or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises, has, includes, contains a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. An element proceeded by “comprises . . . a”, “has . . . a”, “includes . . . a”, “contains . . . a” does not, without more constraints, preclude the existence of additional identical elements in the process, method, article, or apparatus that comprises, has, includes, contains the element. The terms “a” and “an” are defined as one or more unless explicitly stated otherwise herein. The terms “substantially”, “essentially”, “approximately”, “about” or any other version thereof, are defined as being close to as understood by one of ordinary skill in the art, and in one non-limiting embodiment the term is defined to be within 10%, in another embodiment within 5%, in another embodiment within 1% and in another embodiment within 0.5%. The term “coupled” as used herein is defined as connected, although not necessarily directly and not necessarily mechanically. A device or structure that is “configured” in a certain way is configured in at least that way, but may also be configured in ways that are not listed.
It will be appreciated that some embodiments may be comprised of one or more generic or specialized processors (or “processing devices”) such as microprocessors, digital signal processors, customized processors and field programmable gate arrays (FPGAs) and unique stored program instructions (including both software and firmware) that control the one or more processors to implement, in conjunction with certain non-processor circuits, some, most, or all of the functions of the method and/or apparatus described herein. Alternatively, some or all functions could be implemented by a state machine that has no stored program instructions, or in one or more application specific integrated circuits (ASICs), in which each function or some combinations of certain of the functions are implemented as custom logic. Of course, a combination of the two approaches could be used.
Moreover, an embodiment can be implemented as a computer-readable storage medium having computer readable code stored thereon for programming a computer (e.g., comprising a processor) to perform a method as described and claimed herein. Examples of such computer-readable storage mediums include, but are not limited to, a hard disk, a CD-ROM, an optical storage device, a magnetic storage device, a ROM (Read Only Memory), a PROM (Programmable Read Only Memory), an EPROM (Erasable Programmable Read Only Memory), an EEPROM (Electrically Erasable Programmable Read Only Memory) and a Flash memory. Further, it is expected that one of ordinary skill, notwithstanding possibly significant effort and many design choices motivated by, for example, available time, current technology, and economic considerations, when guided by the concepts and principles disclosed herein will be readily capable of generating such software instructions and programs and ICs with minimal experimentation.
The Abstract of the Disclosure is provided to allow the reader to quickly ascertain the nature of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, it can be seen that various features are grouped together in various embodiments for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed embodiment. Thus the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separately claimed subject matter.

Claims (15)

What is claimed is:
1. An apparatus comprising:
a network interface configured to receive an indication that a non-verbal primary wakeup sound (PWS) was detected by a device;
logic circuitry configured to:
determine a non-verbal secondary wakeup sound (SWS) associated with the detected non-verbal PWS; and
reconfigure the device to change its wakeup sound from the non-verbal PWS to the non-verbal SWS upon a detection of the non-verbal PWS;
wherein the non-verbal PWS is a first sound type and the non-verbal SWS is a second sound type determined based on the non-verbal PWS.
2. The apparatus of claim 1 wherein the logic circuitry determines the non-verbal SWS by mapping the non-verbal PWS to the non-verbal SWS such that non-verbal SWS=Function(non-verbal PWS).
3. The apparatus of claim 1 further comprising:
a database comprising a plurality of non-verbal PWSs and their associated non-verbal SWSs; and
wherein the logic circuitry accesses the database to determine an non-verbal SWS from a detected non-verbal PWS.
4. The apparatus of claim 1 further comprising:
a graphical user interface (GUI) coupled to the logic circuitry; and
wherein the logic circuitry instructs the GUI to notify a user that the non-verbal PWS was detected.
5. The apparatus of claim 1 wherein:
the logic circuitry receives a location of the device; and
wherein the non-verbal SWS is also based on the location of the device.
6. The apparatus of claim 5 wherein:
the logic circuitry is configured to determine if other devices are near the device; and
wherein the non-verbal SWS is also based upon whether other devices are near the device.
7. An apparatus comprising:
sound-spotting circuitry configured with a non-verbal primary wakeup sound (PWS);
logic circuitry configured to be woken upon the sound-spotting circuitry detecting the non-verbal PWS, and notify a user that the non-verbal PWS was detected; and
wherein the logic circuitry is also configured to reconfigure the sound-spotting circuitry with a non-verbal secondary wakeup sound (SWS) when the non-verbal PWS has been detected by the sound-spotting circuitry;
wherein the non-verbal PWS is a first sound type and the non-verbal SWS is a second sound type determined based on the non-verbal PWS.
8. The apparatus of claim 7 wherein the logic circuitry determines the non-verbal SWS by mapping the non-verbal PWS to the non-verbal SWS such that non-verbal SWS=Function(non-verbal PWS).
9. The apparatus of claim 7 further comprising:
a database comprising a plurality of non-verbal PWSs and their associated non-verbal SWSs; and
wherein the logic circuitry accesses the database to determine a non-verbal SWS from a detected non-verbal PWS.
10. The apparatus of claim 7 further comprising:
a network interface; and
wherein the logic circuitry notifies the user that the non-verbal PWS was detected by notifying a server via the network interface.
11. The apparatus of claim 7 further comprising:
location-finding circuitry determining a location of a device; and
wherein the non-verbal SWS is also based on the location of the device.
12. The apparatus of claim 11 wherein:
the logic circuitry is configured to determine if other devices are near the device; and
wherein the non-verbal SWS is also based upon whether other devices are near the device.
13. A method comprising the steps of:
configuring sound-spotting circuitry within a device with a non-verbal primary wakeup sound (PWS);
waking upon the sound-spotting circuitry detecting the non-verbal PWS; and
reconfiguring the sound-spotting circuitry with a non-verbal secondary wakeup sound (SWS) when the non-verbal PWS has been detected by the sound-spotting circuitry;
wherein the non-verbal PWS is a first sound type and the non-verbal SWS is a second sound type determined based on the non-verbal PWS.
14. The method of claim 13 wherein the step of reconfiguring comprises the step of determining the non-verbal SWS by mapping the non-verbal PWS to the non-verbal SWS such that non-verbal SWS=Function(non-verbal PWS).
15. The method of claim 13 wherein the non-verbal SWS is also based on a location of the device.
US16/717,102 2019-12-17 2019-12-17 Wake word based on acoustic analysis Active 2040-06-09 US11164563B2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US16/717,102 US11164563B2 (en) 2019-12-17 2019-12-17 Wake word based on acoustic analysis
PCT/US2020/061735 WO2021126475A1 (en) 2019-12-17 2020-11-23 Wake word based on acoustic analysis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US16/717,102 US11164563B2 (en) 2019-12-17 2019-12-17 Wake word based on acoustic analysis

Publications (2)

Publication Number Publication Date
US20210183370A1 US20210183370A1 (en) 2021-06-17
US11164563B2 true US11164563B2 (en) 2021-11-02

Family

ID=73943328

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/717,102 Active 2040-06-09 US11164563B2 (en) 2019-12-17 2019-12-17 Wake word based on acoustic analysis

Country Status (2)

Country Link
US (1) US11164563B2 (en)
WO (1) WO2021126475A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102886539B1 (en) * 2020-12-28 2025-11-14 베이징 바이두 넷컴 사이언스 테크놀로지 컴퍼니 리미티드 Wakeup indicator monitoring method, apparatus and electronic device
FR3150284B1 (en) * 2023-06-22 2025-10-24 Univ De Toulon Method and system for detecting acoustic phenomena

Citations (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030173829A1 (en) * 2002-03-13 2003-09-18 Yu-Wen Zeng Sound-activated wake-up device for electronic input devices having a sleep-mode
US20140237273A1 (en) * 2013-02-15 2014-08-21 Sony Corporation Information processing apparatus, information processing method, and program
US20150063575A1 (en) * 2013-08-28 2015-03-05 Texas Instruments Incorporated Acoustic Sound Signature Detection Based on Sparse Features
US20160098921A1 (en) * 2014-10-02 2016-04-07 Knowles Electronics, Llc Low power acoustic apparatus and method of operation
US9570075B1 (en) * 2016-09-09 2017-02-14 Republic Wireless, Inc. Techniques for integrating voice control into an active telephony call
US20170162205A1 (en) * 2015-12-07 2017-06-08 Semiconductor Components Industries, Llc Method and apparatus for a low power voice trigger device
US20180046864A1 (en) * 2016-08-10 2018-02-15 Vivint, Inc. Sonic sensing
US9899021B1 (en) * 2013-12-20 2018-02-20 Amazon Technologies, Inc. Stochastic modeling of user interactions with a detection system
US20180108343A1 (en) * 2016-10-14 2018-04-19 Soundhound, Inc. Virtual assistant configured by selection of wake-up phrase
US10069976B1 (en) * 2017-06-13 2018-09-04 Harman International Industries, Incorporated Voice agent forwarding
US10083359B2 (en) 2016-04-28 2018-09-25 Motorola Solutions, Inc. Method and device for incident situation prediction
WO2018231133A1 (en) 2017-06-13 2018-12-20 Minut Ab Methods and devices for obtaining an event designation based on audio data
US20180376111A1 (en) 2016-03-15 2018-12-27 Motorola Solutions, Inc Method and apparatus for camera activation
US20190018635A1 (en) * 2017-07-11 2019-01-17 Roku, Inc. Controlling Visual Indicators In An Audio Responsive Electronic Device, and Capturing and Providing Audio Using an API, By Native and Non-Native Computing Devices and Services
US10257740B2 (en) 2017-06-13 2019-04-09 Motorola Solutions, Inc. Assigning priorities to portable communication devices based on roles associated with an incident profile
US10365883B2 (en) * 2007-04-09 2019-07-30 Staton Techiya, Llc Always on headwear recording system
US20190251960A1 (en) * 2018-02-13 2019-08-15 Roku, Inc. Trigger Word Detection With Multiple Digital Assistants
US10748529B1 (en) * 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
US11100918B2 (en) * 2018-08-27 2021-08-24 American Family Mutual Insurance Company, S.I. Event sensing system

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030173829A1 (en) * 2002-03-13 2003-09-18 Yu-Wen Zeng Sound-activated wake-up device for electronic input devices having a sleep-mode
US10365883B2 (en) * 2007-04-09 2019-07-30 Staton Techiya, Llc Always on headwear recording system
US20140237273A1 (en) * 2013-02-15 2014-08-21 Sony Corporation Information processing apparatus, information processing method, and program
US10748529B1 (en) * 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
US20150063575A1 (en) * 2013-08-28 2015-03-05 Texas Instruments Incorporated Acoustic Sound Signature Detection Based on Sparse Features
US9899021B1 (en) * 2013-12-20 2018-02-20 Amazon Technologies, Inc. Stochastic modeling of user interactions with a detection system
US20160098921A1 (en) * 2014-10-02 2016-04-07 Knowles Electronics, Llc Low power acoustic apparatus and method of operation
US20170162205A1 (en) * 2015-12-07 2017-06-08 Semiconductor Components Industries, Llc Method and apparatus for a low power voice trigger device
US20180376111A1 (en) 2016-03-15 2018-12-27 Motorola Solutions, Inc Method and apparatus for camera activation
US10083359B2 (en) 2016-04-28 2018-09-25 Motorola Solutions, Inc. Method and device for incident situation prediction
US20180046864A1 (en) * 2016-08-10 2018-02-15 Vivint, Inc. Sonic sensing
US9570075B1 (en) * 2016-09-09 2017-02-14 Republic Wireless, Inc. Techniques for integrating voice control into an active telephony call
US20180108343A1 (en) * 2016-10-14 2018-04-19 Soundhound, Inc. Virtual assistant configured by selection of wake-up phrase
US10217453B2 (en) 2016-10-14 2019-02-26 Soundhound, Inc. Virtual assistant configured by selection of wake-up phrase
US10069976B1 (en) * 2017-06-13 2018-09-04 Harman International Industries, Incorporated Voice agent forwarding
WO2018231133A1 (en) 2017-06-13 2018-12-20 Minut Ab Methods and devices for obtaining an event designation based on audio data
US10257740B2 (en) 2017-06-13 2019-04-09 Motorola Solutions, Inc. Assigning priorities to portable communication devices based on roles associated with an incident profile
US20190018635A1 (en) * 2017-07-11 2019-01-17 Roku, Inc. Controlling Visual Indicators In An Audio Responsive Electronic Device, and Capturing and Providing Audio Using an API, By Native and Non-Native Computing Devices and Services
US20190251960A1 (en) * 2018-02-13 2019-08-15 Roku, Inc. Trigger Word Detection With Multiple Digital Assistants
US11100918B2 (en) * 2018-08-27 2021-08-24 American Family Mutual Insurance Company, S.I. Event sensing system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
The International Search Report and the Written Opinion corresponding patent application No. PCT/US2020/061735 filed Nov. 23, 2020, dated Mar. 17, 2021, all pages.

Also Published As

Publication number Publication date
US20210183370A1 (en) 2021-06-17
WO2021126475A1 (en) 2021-06-24

Similar Documents

Publication Publication Date Title
US10064001B1 (en) Passive device monitoring using radio frequency signals
RU2602375C2 (en) Systems and methods for service discovery
US9521519B2 (en) Mobile communication devices and context-based geofence control methods thereof
US9232378B2 (en) Locating a victim via a first responder's device
KR101609075B1 (en) Baseband on-board location monitoring
US9307358B2 (en) Non-intrusive region notification
KR101599169B1 (en) Utilizing relationships between places of relevance
US10608929B2 (en) Method for routing communications from a mobile device to a target device
US9942384B2 (en) Method and apparatus for device mode detection
US11164563B2 (en) Wake word based on acoustic analysis
WO2017160170A1 (en) Method and apparatus for camera activation
CN114623843A (en) Object guide method, system, device, terminal and storage medium
US20250095640A1 (en) Improving detection of voice-based keywords using falsely rejected data
EP2654354B1 (en) Providing a current position while reducing an overall power consumption
US10338188B2 (en) Location assistance with a dynamically updated beacon payload from an electronic device
US11816631B2 (en) Voting based proximity detection and ranging
US10811011B2 (en) Correcting for impulse noise in speech recognition systems
AU2017442559B2 (en) Device, system and method for crowd control
CN105245708B (en) Method and system for automatically reminding position of intelligent terminal
US11664016B2 (en) Distinguish devices in PAN/VAN
US10424300B1 (en) Device identification within a personal area network
US20240168707A1 (en) Apparatus, system and method for voice controlled channel change of a radio
US20190190715A1 (en) Digital assistant water mark
US20230171562A1 (en) Systems and methods for tracking devices
CN110583009A (en) Ringing control method, device, storage medium and electronic device

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

AS Assignment

Owner name: MOTOROLA SOLUTIONS INC., ILLINOIS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SAADOUN, FABRICE;GEAN, NISSIM;HAAS, KENNETH A.;SIGNING DATES FROM 20200115 TO 20200323;REEL/FRAME:052190/0963

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4