US12198714B2 - Voice signal analysis method and device and chip design method and device - Google Patents
Voice signal analysis method and device and chip design method and device Download PDFInfo
- Publication number
- US12198714B2 US12198714B2 US17/850,906 US202217850906A US12198714B2 US 12198714 B2 US12198714 B2 US 12198714B2 US 202217850906 A US202217850906 A US 202217850906A US 12198714 B2 US12198714 B2 US 12198714B2
- Authority
- US
- United States
- Prior art keywords
- resolution
- signal
- model
- voice
- snore
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/30—Circuit design
- G06F30/39—Circuit design at the physical level
- G06F30/392—Floor-planning or layout, e.g. partitioning or placement
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/48—Other medical applications
- A61B5/4803—Speech analysis specially adapted for diagnostic purposes
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6801—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient specially adapted to be attached to or worn on the body surface
- A61B5/6802—Sensor mounted on worn items
- A61B5/6803—Head-worn items, e.g. helmets, masks, headphones or goggles
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/68—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient
- A61B5/6887—Arrangements of detecting, measuring or recording means, e.g. sensors, in relation to patient mounted on external non-worn devices, e.g. non-medical devices
- A61B5/6898—Portable consumer electronic devices, e.g. music players, telephones, tablet computers
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/72—Signal processing specially adapted for physiological signals or for diagnostic purposes
- A61B5/7235—Details of waveform analysis
- A61B5/7264—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems
- A61B5/7267—Classification of physiological signals or data, e.g. using neural networks, statistical classifiers, expert systems or fuzzy systems involving training the classification device
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F30/00—Computer-aided design [CAD]
- G06F30/30—Circuit design
- G06F30/39—Circuit design at the physical level
- G06F30/398—Design verification or optimisation, e.g. using design rule check [DRC], layout versus schematics [LVS] or finite element methods [FEM]
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/26—Recognition of special voice characteristics, e.g. for use in lie detectors; Recognition of animal voices
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/27—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
- G10L25/30—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04B—TRANSMISSION
- H04B1/00—Details of transmission systems, not covered by a single one of groups H04B3/00 - H04B13/00; Details of transmission systems not characterised by the medium used for transmission
- H04B1/38—Transceivers, i.e. devices in which transmitter and receiver form a structural unit and in which at least one part is used for functions of transmitting and receiving
- H04B1/40—Circuits
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
- G10L25/66—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for extracting parameters related to health condition
Definitions
- the disclosure relates to a voice signal analysis method, a technique for chip design, and a chip design device.
- the disclosure provides a voice signal analysis method, a voice signal analysis device, a technique for chip design, and a chip design device capable of reducing power consumption and signal errors, facilitating accuracy, and reducing chip size.
- An embodiment of the disclosure provides a voice signal analysis method.
- the method includes: in a first updating gradient, training a resolution recovery model in an audio processing chip by using multiple first voice training data meeting a same grouping condition in multiple mission sets; in a second updating gradient, training the resolution recovery model by interleavingly using second voice training data meeting different grouping conditions in the mission sets; iteratively executing the first updating gradient and the second updating gradient to set an initial model parameter of the resolution recovery model; and recovering a high-resolution snore signal from a low-resolution snore signal by using the resolution recovery model.
- a resolution of the low-resolution snore signal is lower than a resolution of the high-resolution snore signal.
- a voice signal analysis device including an audio processing chip and a storage circuit.
- a resolution recovery model is built is built in the audio processing chip.
- the storage circuit is coupled to the audio processing chip and configured to store multiple mission sets.
- the audio processing chip is configured to: in a first updating gradient, train the resolution recovery model by using multiple first voice training data meeting a same grouping condition in the mission sets; in a second updating gradient, train the resolution recovery model by interleavingly using multiple second voice training data meeting different grouping conditions in the mission sets; iteratively execute the first updating gradient and the second updating gradient to set an initial model parameter of the resolution recovery model; and recover a high-resolution snore signal from a low-resolution snore signal by using the resolution recovery model.
- a resolution of the low-resolution snore signal is lower than a resolution of the high-resolution snore signal.
- Another embodiment of the disclosure provides a chip design method that is provided for a chip layout prediction model in a chip design device.
- the chip design method includes: executing model rebuilding according to multiple constraints; and iteratively looking for a chip layout design corresponding to an audio processing chip through multiple decision paths to manufacture the audio processing chip.
- the chip layout design meets a minimum layout area of the audio processing chip.
- a chip design device including a storage circuit and a processor.
- the storage circuit is configured to store a chip layout prediction model.
- the processor is coupled to the storage circuit.
- the processor is configured to run the chip layout prediction model to: execute model rebuilding according to multiple constraints; and iteratively look for a chip layout design corresponding to an audio processing chip through multiple decision paths to manufacture the audio processing chip.
- the chip layout design meets a minimum layout area of the audio processing chip.
- FIG. 1 A is a schematic diagram illustrating a voice signal analysis device according to an embodiment of the disclosure.
- FIG. 1 B is a schematic diagram illustrating a flow of setting an initial model parameter of a resolution recovery model according to an embodiment of the disclosure.
- FIG. 1 C is a schematic diagram illustrating a computation framework of the resolution recovery model according to an embodiment of the disclosure.
- FIG. 2 is a schematic diagram illustrating a voice signal analysis device according to an embodiment of the disclosure.
- FIG. 3 is a schematic diagram illustrating a flow of recovering a high-resolution snore signal from a low-resolution snore signal according to an embodiment of the disclosure.
- FIG. 4 is a schematic diagram illustrating a chip design device according to an embodiment of the disclosure.
- FIG. 5 is a schematic diagram illustrating reducing a chip size according to an embodiment of the disclosure.
- FIG. 6 is a schematic diagram illustrating a conventional process of looking for a minimum layout area of a chip by using a heuristic algorithm.
- FIG. 7 is a schematic diagram illustrating a flow of a relaxation rebuild branch method according to an embodiment of the disclosure.
- FIG. 8 is a flowchart illustrating a voice signal analysis method according to an embodiment of the disclosure.
- FIG. 9 is a flowchart of a chip design method according to an embodiment of the disclosure.
- a low-resolution snore may be added with a random initialization model parameter and entered into a first layer of a neural network (i.e., to learn micro-features including weight calculation, deviation calculation, and excitation function calculation) and a second layer of neural network (i.e., to learn macro-features including weight calculation, deviation calculation, and excitation function calculation) to generate a high-resolution snore.
- a neural network i.e., to learn micro-features including weight calculation, deviation calculation, and excitation function calculation
- a second layer of neural network i.e., to learn macro-features including weight calculation, deviation calculation, and excitation function calculation
- a dual-gradient learning generalization method is adopted to replace random initialization, thereby reducing power consumption and signal errors and facilitating accuracy.
- FIG. 1 A is a schematic diagram illustrating a voice signal analysis device according to an embodiment of the disclosure.
- a voice signal analysis device 10 may include various electronic devices with a voice signal processing function, such as a wired earphone, a wireless earphone, a smart phone, a tablet computer, a laptop computer, or a digital voice recorder.
- the types of the voice signal analysis device 10 are not limited to the above.
- the voice signal analysis device 10 includes an audio processing chip 11 and a storage circuit 12 .
- the audio processing chip 11 is configured to process a voice signal.
- the audio processing chip 11 may include a central processing unit (CPU), or a programmable microprocessor of common usage or specific usage, a digital signal processor (DSP), a programmable controller, an application specific integrated circuit (ASIC), a programmable logic device (PLD), other similar devices, or a combination thereof.
- the audio processing chip 11 may also handle a portion or the entirety of the operation of the audio signal analysis device 10 .
- the storage circuit 12 is coupled to the audio processing chip 11 .
- the storage circuit 12 is configured to store data.
- the storage circuit 12 may include a volatile storage circuit and a non-volatile storage circuit.
- the volatile storage circuit is configured to store data in a volatile manner.
- the volatile storage circuit may include a random access memory (RAM) or a similar volatile storage medium.
- the non-volatile storage circuit is configured to store data in a non-volatile manner.
- the non-volatile storage circuit may include a read-only memory (ROM), a solid-state drive (SSD), and/or a conventional hard disk drive (HDD), and/or a similar non-volatile storage medium.
- the voice signal analysis device 10 may further include a power management circuit for managing the power of the audio signal analysis device 10 , various sensors, and various input/output (I/O) interfaces.
- a power management circuit for managing the power of the audio signal analysis device 10
- various sensors for detecting obstacles in the environment
- various input/output (I/O) interfaces for detecting obstacles in the environment.
- the disclosure is not particularly limited by the types of the sensors and the I/O interfaces included in the voice signal analysis device 10 .
- the audio processing chip 11 may include a resolution recovery model 13 .
- the resolution recovery model 13 may be built in the audio processing chip 11 .
- the resolution recovery model 13 may serve to recover a voice sampled signal (also referred to as “second voice sampled signal”) with a resolution (also referred to as “second resolution”) from a voice sampled signal (also referred to as “first voice sampled signal”) with another resolution (also referred to as “first resolution”).
- the resolution of the second voice sampled signal is higher than the resolution of the first voice sampled signal.
- the resolution recovery model 13 may recover the second voice sampled signal with a higher resolution from the first voice sampled signal.
- the resolution of the second voice sampled signal may be close to or even same as the resolution (also referred to as “original resolution”) of the original voice signal.
- the first voice sampled signal includes a low-resolution snore signal and/or the second voice sampled signal includes a high-resolution snore signal. The resolution of the low-resolution snore signal is lower than the resolution of the high-resolution snore signal.
- the resolution recovery model 13 may include one or more neural networks.
- Various neural networks such as a convolutional neural network (CNN) and/or a feed forward deep neural network (FFDNN), and/or deep learning models may be used to realize the neural network, and such neural work may be modified as needed.
- CNN convolutional neural network
- FFDNN feed forward deep neural network
- deep learning models may be used to realize the neural network, and such neural work may be modified as needed.
- the storage circuit 12 may be configured to store multiple mission sets 14 ( 1 ) to 14 ( n ).
- Each of the mission sets 14 ( 1 ) to 14 ( n ) may include at least one voice training data to train the resolution recovery model 13 .
- the voice training data may include sampled data of voice signals (also referred to as “voice sampled data”).
- the disclosure is not particularly limited by the total number of the mission sets 14 ( 1 ) to 14 ( n ).
- the voice training data in the mission sets 14 ( 1 ) to 14 ( n ) may be grouped according to specific grouping conditions (also referred to filter conditions).
- the grouping conditions may include conditions, such as age, gender, and nationality, for distinguishing voice sensing targets of different types.
- the grouping conditions are not limited to the above. Voice training data meeting the same grouping condition may be stored in the same mission set among the mission sets 14 ( 1 ) to 14 ( n ).
- the mission set 14 ( 1 ) may serve to store voice training data of voice sensing targets whose ages range from 15 to 18 years old
- the mission set 14 ( 2 ) may serve to store voice training data of voice sensing targets whose ages range from 19 to 22 years old
- the mission set 14 ( 3 ) may serve to store voice training data of male voice sensing targets
- the mission set 14 ( 4 ) may serve to store voice training data of female voice sensing targets.
- the grouping conditions may be modified based on practical needs and the disclosure is not particularly limited by the grouping conditions.
- different mission sets in the mission sets 14 ( 1 ) to 14 ( n ) respectively store voice training data of different grouping conditions.
- the audio processing chip 11 may iteratively execute operations of different types of updating gradients according to the mission sets 14 ( 1 ) to 14 ( n ) to set an initial parameter (also referred to as “initial model parameter”) of the resolution recovery model 13 .
- FIG. 1 B is a schematic diagram illustrating a flow of setting an initial model parameter of a resolution recovery model according to an embodiment of the disclosure.
- the audio processing chip 11 may set an initialization parameter of the resolution recovery model 13 .
- iteration starts.
- the audio processing chip 11 may randomly select missions for training from the mission sets 14 ( 1 ) to 14 ( n ) of FIG. 1 A .
- the audio processing chip 11 may execute an operation of an updating gradient (also referred to “first updating gradient” (learning individual missions)).
- the audio processing chip 11 may train the resolution recovery model 13 by using multiple voice training data (also referred to as “first voice training data”) meeting the same grouping condition among the mission sets 14 ( 1 ) to 14 ( n ) to learn audio features of individual missions.
- first voice training data also referred to as “first voice training data”
- all or at least some of the voice training data in the mission set 14 ( i ) may be sequentially input to the resolution recovery model 13 . Accordingly, the audio recovery model 13 continuously learns the voice features of the individual missions in the learning mission set 14 ( i ).
- all or at least some of the voice training data in the mission set 14 ( j ) may be sequentially input to the resolution recovery model 13 . Accordingly, the audio recovery model 13 continuously learns the voice features of the individual missions in the learning mission set 14 ( j ).
- the audio processing chip 11 may execute weight calculation, deviation calculation, and excitation function calculation related to the resolution recovery model 13 according to the audio features of the individual missions learned by the resolution recovery model 13 . Then, the audio processing chip 11 may calculate cross-entropy and update the parameter of the resolution recovery model 13 according to a result of calculation.
- FIG. 1 C is a schematic diagram illustrating a computation framework of the resolution recovery model according to an embodiment of the disclosure.
- an input signal 110 may include voice training data and be input to the resolution recovery model 13 .
- the input signal 110 may include multiple input values x 1 to x 1 to x N T.
- the resolution recovery model 13 may perform iterative computing on the voice training data 110 by using computation models 120 ( 1 ) to 120 (L) and generate an output signal 130 .
- the output signal 130 include multiple output values y 1 to y M .
- N and M are positive integers, and M is greater than N. In other words, the total number of the output values y 1 to y M is greater than the total number of the input values x 1 to x N .
- the resolution recovery model 13 may recover the output signal 130 with a higher resolution from the input signal 110 with a lower resolution.
- the computation parameter adopted by the resolution recovery mode 13 may be constantly updated to look for an optimal parameter.
- the resolution recovery model 13 may include the multiple computation models 120 ( 1 ) to 120 (L).
- the number of layers of the computation models 120 ( 1 ) to 120 (L) is L, and L may be an arbitrary positive integer greater than 1.
- a computation model 120 ( k ) in the computation models 120 ( 1 ) to 120 (L) calculates the excitation function according to a weight value W(k) and a deviation value b(k), and k is between 1 and L.
- Formula (1.1) may represent the excitation function.
- W(k) represents the weight value adopted by the computation model 120 ( k )
- b(k) represents the deviation value adopted by the computation model 120 ( k )
- x i represents the input value at a computation node in the computation model 120 ( k ).
- CE represents cross-entropy.
- CE reflects a difference between the output signal 130 generated by the resolution recovery model 13 and validation data.
- the audio processing chip 11 may constantly update (e.g., adjust) the computation parameter (e.g., the weight value and the deviation) adopted by the resolution recovery model 13 according to the calculation result of Formula (1.3) until an optimal parameter is found.
- Step 105 the audio processing chip 11 may execute an operation of another updating gradient (also referred to as “second updating gradient” (learning lumped individual missions)).
- the audio processing chip 11 may interleavingly train the resolution recovery model 13 by using multiple voice training data (also referred to as “second voice training data”) meeting different grouping conditions among the mission sets 14 ( 1 ) to 14 ( n ).
- a voice training data in the mission set 14 ( p ) may be firstly input to the resolution recovery model 13 to train the resolution recovery model 13 .
- a voice training data in the mission set 14 ( q ) may be input to the resolution recovery model 13 to train the resolution recovery model 13 , where p is not equal to q.
- a greater number of voice training data belonging to different mission sets may be used interleavingly to train the resolution recovery model 13 , thereby facilitating as much as possible the accuracy of recovering the voice signals of different types of voice sensing targets by using the resolution recovery model 13 .
- the audio processing chip 11 may execute weight calculation, deviation calculation, and excitation function calculation related to the resolution recovery model 13 according to the audio features of the lumped individual missions (i.e., missions across mission sets) learned by the resolution recovery model 13 . Then, the audio processing chip 11 may calculate cross-entropy and update the lumped parameter of the resolution recovery model 13 according to a result of calculation. Details relating to the operation may be referred to the description about Step 104 and FIG. 1 C , and therefore will not be repeated in the following.
- the audio processing chip 11 may iteratively execute Steps 104 and 105 to set an initial parameter (i.e., initial model parameter) of the resolution recovery model 13 .
- an initial parameter i.e., initial model parameter
- the audio processing chip 11 may determine that the iteration ends in Step 106 and generate the model parameter to be used (i.e., the initial model parameter) in Step 107 .
- the resolution recovery model 13 may recover the voice sampled signal based on the initial model parameter.
- the process of setting the initial parameter (i.e., the initial model parameter) of the resolution recovery model 13 may also be referred to as a dual-gradient learning generalization method.
- the missions are used randomly to train the neural network model, and the initialization parameter of the model may be set randomly. Therefore, even after training, the neural network model is still unable to effectively facilitate the accuracy of recovering the voice signal.
- the resolution recovery model 13 by obligatorily adopting the customized, stage-based model training mechanism, after the training is completed, the resolution recovery model 13 is able to accurately recover a high-resolution snore signal (i.e., the second voice sampled signal) from a low-resolution snore signal (i.e., the first voice sampled signal) based on a favorable initial model parameter, and the recovery error can be significantly reduced.
- the low-resolution snore signal (i.e., the first voice sampled signal) is obtained by sampling the original voice signal using a sampling mode with a lower power consumption. Accordingly, the power (e.g., device power consumption) consumed by the voice signal analysis device 10 for sampling the voice signal can be effectively reduced.
- the resolution recovery model 13 is able to output a high-resolution snore signal (i.e., the second voice sampled signal) close to or even nearly the same as the original voice signal for subsequent analysis. Accordingly, in the entire process of sampling (i.e., sensing) and recovering the voice signal, the power consumption of the voice signal analysis device 10 is reduced significantly, whereas the accuracy of analyzing the voice signal is still maintained or even increased.
- some of the parameters (e.g., weight value) adopted by the neural network in the resolution recovery model 13 may be constantly updated to reduce the prediction error of the resolution recovery model 13 and/or facilitate the prediction accuracy of the resolution recovery model 13 .
- the audio processing chip 11 may evaluate the prediction error of the resolution recovery model 13 in each iteration according to the excitation function and the cross-entropy commonly used in neural networks, deep learning, and machine learning as well as the parameter updating (in “learning individual missions”) or lumped parameter updating (in “learning lumped individual missions”), until the iterative computation ends. Relevant operation details have been described above and therefore will not be repeated in the following.
- FIG. 2 is a schematic diagram illustrating a voice signal analysis device according to an embodiment of the disclosure.
- a voice signal analysis device 20 may include various electronic devices with voice signal reception and processing functions, such as a wired earphone, a wireless earphone, a smart phone, a tablet computer, a laptop computer, or a digital voice recorder.
- the types of the voice signal analysis device 20 are not limited to the above.
- the voice signal analysis device 20 may include an audio processing chip 21 , a storage circuit 22 , a sound receiving device 23 , and a resolution recovery model 24 .
- the audio processing chip 21 , the storage circuit 22 , and the resolution recovery model 24 are respectively the same as or similar to the audio processing chip 11 , the storage circuit 12 , and the resolution recovery model 13 .
- the sound receiving device 23 is coupled to the audio processing chip 21 .
- the sound receiving device 23 is configured to receive a voice signal (i.e., the original voice signal).
- the sound receiving device 23 may include an audio receiving device such as a microphone.
- the original voice signal may include a voice signal made by a target person, such as an ambient sound.
- the original voice signal may reflect the sound of breathing when the target person is asleep or at another moment.
- the audio processing chip 21 may receive the original voice signal through the sound receiving device 23 . Then, the audio processing chip 21 may sample the original voice signal based on a default sampling frequency (also referred to as “first sampling frequency”) to obtain the low-resolution snore signal. Then, the audio processing chip 21 may recover the high-resolution snore signal from the low-resolution snore signal through the resolution recovery model 24 .
- the low-resolution snore signal corresponds to the first sampling frequency
- the high-resolution snore signal corresponds to another sampling frequency (also referred to as “second sampling frequency”).
- the second sampling frequency is higher than the first sampling frequency.
- the first sampling frequency is positively correlated with the power consumption at the time when the audio processing chip 21 samples (i.e., senses) the original voice signal. That is, the higher the first sampling frequency, the greater the power consumption when the audio processing chip 21 samples the original voice signal. Also, the lower the first sampling frequency, the lower the power consumption when the audio processing chip 21 samples the original voice signal. Therefore, by lowering the first sampling frequency, the power consumed when the voice signal analysis device 20 executes signal sampling can be effectively reduced.
- FIG. 3 is a schematic diagram illustrating a flow of recovering the high-resolution snore signal from the low-resolution snore signal according to an embodiment of the disclosure.
- the low-resolution snore signal is obtained.
- the audio processing chip 21 may set the initial parameter (i.e., the initial model parameter) of the resolution recovery model 13 by using the dual-gradient learning generalization process. Then, the audio processing chip 21 may recover the high-resolution snore signal from the low-resolution snore signal through a resolution recovery model 31 .
- the process in which the resolution recovery model 13 processes the low-resolution snore signal may include increasing the resolution of the low-resolution snore signal.
- the resolution recovery model 31 may include the resolution recovery model 13 of FIG. 1 C or the resolution recovery model 24 of FIG. 2 .
- the input signal 110 may include the low-resolution snore signal
- the output signal 130 may include the high-resolution snore signal.
- the audio processing chip 21 may obtain the high-resolution snore signal according to the output of the resolution recovery model 31 .
- the high-resolution snore signal may be used in various applications.
- the resolution recovery mode 31 may include a first layer 311 of a neural network and a second layer 312 of the neural network.
- the first layer 311 of the neural network layer and the second layer 312 of the neural network layer may sequentially process the low-resolution snore signal according to the initial model parameter set in Step 302 .
- the first layer 311 of the neural network layer may be used to execute micro-feature learning on the low-resolution snore signal
- the second layer 312 of the neural network layer may be used to execute macro-feature learning on the low-resolution snore signal.
- the resolution recovery model 31 may output the high-resolution snore signal according to the result of micro-feature learning and the result of macro-feature learning.
- the computation processes of the micro-feature learning and the macro-feature learning also include weight calculation, deviation calculation, and excitation function calculation, etc. Relevant operation details have been described above and will not be repeated in the following.
- the resolution recovery model 31 may further include a greater number of layers of neural network. Such number may be determined based on practical needs.
- the audio processing chip 21 may evaluate a respiratory condition of the target person according to a signal feature of the high-resolution snore signal.
- the audio processing chip 21 may compare the signal feature, such as waveform, of the second voice sampled signal to a signal feature template in a database, and evaluate the respiratory condition of the target person according to a comparison result. For example, when finding that the signal feature of the second voice sampled signal matches a specific signal template in the database, the audio processing chip 21 may generate evaluation information to reflect that the respiratory tract of the target person exhibits a specific symptom.
- the evaluation information may reflect whether the target person suffers from a sleep apnea syndrome, whether the respiratory tract of the target person is obstructed, and/or the obstructed position of the respiratory tract of the target person, as well as information related to the respiratory condition of the target person.
- the audio processing chip 21 may generate evaluation information to reflect that the respiratory tract of the target person is in healthy condition.
- the resolution recovery model 31 may output the high-resolution snore signal close to or nearly the same as the original sound signal according to the low-resolution snore signal. Then, the signal feature of the high-resolution snore signal is analyzed to evaluate the respiratory condition of the target person, which is nearly equivalent to carrying out an analysis using the original voice signal of the target person. Accordingly, a balance is struck between lowering device power consumption and increasing detection accuracy.
- the steps for reducing a chip size includes defining a target function, minimizing a total area, constraints (including device size, device pins, component arrangement), a random initial parameter, a heuristic algorithm, and a second-minimum total area may be rendered. However, it remains an issue to further design and reduce the chip size.
- a relaxation rebuild branch method replaces the heuristic algorithm to obtain a minimum total area solution.
- FIG. 4 is a schematic diagram illustrating a chip design device according to an embodiment of the disclosure.
- a chip design device 40 may include various electronic devices with a data processing function, such as a smart phone, a tablet computer, a notebook computer, a desktop computer, or a server.
- the types of the chip design device 40 are not limited thereto.
- the chip design device 40 may include a processor 41 and a storage circuit 42 .
- the processor 41 may handle the entirety or a portion of the operation of the chip design device 40 .
- the processor 41 may include a CPU, or a programmable microprocessor of common or specific usage, a DSP, a programmable controller, an ASIC, a PLD, other similar devices, or a combination thereof.
- the storage circuit 42 is coupled to the processor 41 .
- the storage circuit 42 is configured to store data.
- the storage circuit 42 may include a volatile storage circuit and a non-volatile storage circuit.
- the volatile storage circuit is configured to store data in a volatile manner.
- the volatile storage circuit may include a RAM or a similar volatile storage medium.
- the non-volatile storage circuit is configured to store data in a non-volatile manner.
- the non-volatile storage circuit may include a ROM, an SSD, a HDD, or a similar non-volatile storage medium.
- the storage circuit 42 may be configured to store a chip layout prediction model 43 .
- the chip layout prediction model 43 may carry out a design by using a customized algorithm.
- the customized algorithm may include the relaxation rebuild branch method.
- the chip layout prediction model 43 is applicable at the chip design stage, and is able to reduce as much as possible the size of the audio processing chip 11 of FIG. 1 A or the audio processing chip 21 of FIG. 2 according to basic chip design requirements (e.g., the device size required in the chip, the device pin, and the device component arrangement, etc.).
- FIG. 5 is a schematic diagram illustrating reducing a chip size according to an embodiment of the disclosure.
- a possible layout region of the chip may include a region 51 .
- the width, height, area, and total line length of the region 51 may be 35.9 mm, 37.7 mm, 1353.43 mm 2 , and 3622.9 mm, respectively.
- FIG. 6 is a schematic diagram illustrating a conventional process of looking for a minimum layout area of a chip by using a heuristic algorithm.
- the target function is defined.
- the minimum layout area of the chip may be found by solving the target function.
- it is attempted to look for the minimum layout area of the chip by using the target function.
- constraints are introduced to set up the basic chip design requirements (e.g., the device size required in the chip, the device pin, and the device component arrangement, etc.).
- Step 604 a random initial parameter is set.
- the minimum layout area of the chip is found by using the heuristic algorithm.
- the second-minimum chip total area instead of the minimum chip total area, may be rendered.
- the minimum layout region of the chip may be reduced to a region 52 from the region 51 .
- the width, height, area, and total line length of the region 52 may be 34.4 mm, 35.8 mm, 1231.52 mm 2 , and 3149.2 mm, respectively.
- the invention is not limited thereto.
- the processor 41 may run the chip layout prediction model 43 adopting the relaxation rebuild branch method and introduce the basic chip design requirements (e.g., the device size required in the chip, the device pin, and the device component arrangement, etc.).
- the chip layout prediction model 43 may execute model rebuilding according to multiple constraints and iteratively look for a chip layout design corresponding to the target chip (e.g., the audio processing chip 11 of FIG. 1 A or the audio processing chip 21 of FIG. 2 ) through multiple decision paths.
- the chip layout design may be used to manufacture a target chip.
- the chip layout design may meet the (substantially) minimum layout area of the target chip.
- FIG. 7 is a schematic diagram illustrating a flow of a relaxation rebuild branch method according to an embodiment of the disclosure.
- the processor 41 may activate the relaxation rebuild branch method (also referred to as “constraint relaxation method”).
- the processor 41 may obtain a first-type constraint (also referred to as “hard constraint”) and a second-type constraint (also referred to as “soft constraint”) corresponding to the relaxation rebuild branch method.
- the processor 41 may execute model rebuilding according to the hard constraint and the soft constraint. Specifically, the hard constraint is required to be met during the process of model rebuilding, and the soft constraint is not necessarily met in the process of model rebuilding.
- the processor 41 may rebuild a mathematical model adopted in default by the chip layout prediction model 43 according to the hard constraint and the soft constraint. For example, the processor 41 may modify the mathematical model adopted in default by the chip layout prediction model 43 into a new mathematical model. Compared with the mathematical model adopted in default by the chip layout prediction model 43 , it is easier to find the optimal solution of the target function by using the new mathematical model.
- the optimal resolution may include the solution for the minimum total area of the target chip (e.g., the length and the width of the target chip).
- the target function may be represented as Formula (2.1) as follows: ⁇ ( H+W )+ ⁇ e ( R e ⁇ L e +U e ⁇ D e ) (2.1)
- H represents the length of a carrier
- W represents the width of the carrier
- R e represents a right boundary position of an e th netlist
- L e represents a left boundary position of the e th netlist
- U c represents an upper boundary position of the e th netlist
- D e represents a lower boundary position of the e th netlist.
- ⁇ (H+W) may represent the sum of the length and the width of the carrier
- ⁇ (R e ⁇ L e +U e ⁇ D e ) may represent the line length of each netlist.
- x i represents the X-axis position of a central point of a device i
- y i represents the Y-axis position of the central point of the device i
- z ij and z′ ij represent relative positions between the devices i and j
- r i represents whether the device i is rotated 90 degrees
- r′ i represents whether the device i is rotated 180 degrees
- r′′ i represents whether the device i is rotated 270 degrees.
- the processor 41 may execute the branch method through multiple decision paths corresponding to Steps 704 to 707 to iteratively look for the solution for the minimum area of the chip layout design corresponding to the target chip.
- the multiple decision paths corresponding to Steps 704 to 707 each correspond to an alternative solution-finding rule for finding the solution for the target function.
- the processor 41 may obtain the solution for the minimum total area.
- the processor 41 may replace some decision variables in the chip layout prediction model 43 , such as integrating or splitting at least some unknown decision variables (i.e., replacing decision variables). Then, the processor 41 may narrow down feasible solutions or the solution space of the mathematical model adopted by the chip layout prediction model 43 in attempt to urge the solution of the target function (i.e., narrowing down feasible solutions/reducing the solution space).
- the processor 41 may modify some inequalities in the chip layout prediction model 43 into equalities with the same or similar meanings and look for solutions (i.e., generating valid inequalities) by using a relaxation variable. Then, the processor 41 may loosen the limitation of being “0” or “1” in some variables into being between “0” and “1” in attempt to urge the solution of the target function (i.e., obtaining an enhanced linear relaxation solution).
- the processor 41 may optimize some or all of the variables in the chip layout prediction model 43 , such as choosing some or all of the variables to look for a solution or adjust variable values, in attempt to urge the solution of the target function (i.e., optimizing some or all of the variables).
- the processor 41 may combine some independent constraints in the chip layout prediction model 43 with the target function to generate sub-issues (i.e., independent constraints). Then, the processor 41 may decompose the mathematical model adopted in the chip layout prediction model 43 into multiple sub-issues and derive the solution of the original issue (i.e., the target function) according to the solutions of the sub-issues in attempt to urge the solution of the target function (i.e., looking for the solutions of the sub-issues). Depending on practical needs, there may be a greater or fewer number of the decision paths. In addition, the contents of the decision paths may also be adjusted based on practical needs. The disclosure is not particularly limited in this regard.
- the processor 41 may combine and optimize the decision paths corresponding to Steps 704 to 707 by using the branch method. For example, the processor 41 may execute iterative search according to the branch method, and look for the solution through different branch variables and/or sub-issues in the decision paths.
- the optimal solution of the target function that is eventually obtained is the solution for the minimum total area.
- the solution for the minimum total area may reflect the minimum layout area of the target chip (e.g., the length and the width of the target chip) of the target chip derived from the chip layout prediction model 43 .
- the relaxation rebuild branch method shown in FIG. 7 may replace the conventional heuristic algorithm in Step 605 of FIG. 6 to thereby further reduce the chip layout area.
- the processor 41 may provide the chip layout design to the chip manufacturing department or the chip manufacturer of the next stage to set a plan for and manufacture the target chip.
- the type of device whose device size needs to be reduced as much as possible such as a wireless earphone
- the design flexibility of the wireless earphone and the comfort when a user wears the wireless earphone are greatly facilitated.
- FIG. 8 is a flowchart illustrating a voice signal analysis method according to an embodiment of the disclosure.
- Step S 801 in the first updating gradient, the resolution recovery model in the audio processing chip is trained by using multiple first voice training data meeting the same grouping condition in the mission sets.
- Step S 802 in the second updating gradient, the resolution recovery model is trained by interleavingly using multiple second voice training data meeting different grouping conditions in the mission sets.
- Step S 803 the first updating gradient and the second updating gradient are iteratively executed to set the initial model parameter of the resolution recovery model.
- Step S 804 the high-resolution snore signal is restored from the low-resolution snore signal by using the resolution recovery model. The resolution of the low-resolution snore signal is lower than the resolution of the high-resolution snore signal.
- FIG. 9 is a flowchart of a chip design method according to an embodiment of the disclosure.
- Step S 901 model rebuilding is executed according to multiple constraints.
- Step 902 the chip layout design corresponding to the audio processing chip is found iteratively through multiple decision paths to manufacture the audio processing chip. In addition, the chip layout design meets the minimum layout area of the audio processing chip.
- FIGS. 8 and 9 may be implemented as programming codes or circuits.
- the disclosure is not particularly limited in this regard.
- the methods in FIGS. 8 and 9 may be used together with or independently from the exemplary embodiment.
- the disclosure is not particularly limited in this regard.
Landscapes
- Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Signal Processing (AREA)
- Artificial Intelligence (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Multimedia (AREA)
- Evolutionary Computation (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Veterinary Medicine (AREA)
- Medical Informatics (AREA)
- Public Health (AREA)
- General Health & Medical Sciences (AREA)
- Animal Behavior & Ethology (AREA)
- Surgery (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Biomedical Technology (AREA)
- Heart & Thoracic Surgery (AREA)
- Computational Linguistics (AREA)
- Computer Hardware Design (AREA)
- Theoretical Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Engineering & Computer Science (AREA)
- Physiology (AREA)
- Fuzzy Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Mathematical Physics (AREA)
- Geometry (AREA)
- Psychiatry (AREA)
- General Physics & Mathematics (AREA)
- Quality & Reliability (AREA)
- Architecture (AREA)
- Circuit For Audible Band Transducer (AREA)
- Image Analysis (AREA)
- Input Circuits Of Receivers And Coupling Of Receivers And Audio Equipment (AREA)
- Stored Programmes (AREA)
Abstract
Description
σ(x)=max(0, x) (1.1)
x=W(k)×x i +b(k) (1.2)
CE=−Σ i c t i×log((f(s)n) (1.3)
α(H+W)+βΣe(R e −L e +U e −D e) (2.1)
Claims (9)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| TW111117761 | 2022-05-12 | ||
| TW111117761A TWI820704B (en) | 2022-05-12 | 2022-05-12 | Method and device for voice signal analyzation, method and device for chip design |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20230368810A1 US20230368810A1 (en) | 2023-11-16 |
| US12198714B2 true US12198714B2 (en) | 2025-01-14 |
Family
ID=88699397
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US17/850,906 Active 2043-04-12 US12198714B2 (en) | 2022-05-12 | 2022-06-27 | Voice signal analysis method and device and chip design method and device |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US12198714B2 (en) |
| CN (1) | CN117113019A (en) |
| TW (1) | TWI820704B (en) |
Families Citing this family (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US12249345B2 (en) * | 2022-08-26 | 2025-03-11 | Google Llc | Ephemeral learning and/or federated learning of audio-based machine learning model(s) from stream(s) of audio data generated via radio station(s) |
Citations (20)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| TW201121495A (en) | 2009-12-16 | 2011-07-01 | Ind Tech Res Inst | Method and system for activity monitoring |
| US8332793B2 (en) | 2006-05-18 | 2012-12-11 | Otrsotech, Llc | Methods and systems for placement and routing |
| US20150073306A1 (en) | 2012-03-29 | 2015-03-12 | The University Of Queensland | Method and apparatus for processing patient sounds |
| US20170105679A1 (en) | 2015-05-20 | 2017-04-20 | Huinno Co., Ltd. | Earphone comprising bio-signal measurement means, and bio-signal monitoring system comprising same |
| CN106971042A (en) | 2017-03-31 | 2017-07-21 | 福州大学 | A kind of method that legalizes for being used to mix the design of Height Standard element circuit |
| CN107438398A (en) | 2015-01-06 | 2017-12-05 | 大卫·伯顿 | Mobile Wearable Monitoring System |
| US20180235537A1 (en) | 2017-02-21 | 2018-08-23 | Zoll Medical Corporation | Assigning zone-based rankings and actions |
| US20180322386A1 (en) * | 2017-05-05 | 2018-11-08 | Intel Corporation | Fine-grain compute communication execution for deep learning frameworks |
| US20190350532A1 (en) | 2014-07-30 | 2019-11-21 | Valencell, Inc. | Physiological monitoring devices and methods using optical sensors |
| JP2020506491A (en) | 2017-01-08 | 2020-02-27 | アルファイクス コーポレイションAlphaics Corporation | System and chip design system and method with artificial intelligence and reinforcement learning |
| US20200118000A1 (en) * | 2018-10-10 | 2020-04-16 | NEC Laboratories Europe GmbH | Method and system for distributed deep learning |
| US20200293888A1 (en) * | 2019-03-13 | 2020-09-17 | Cognizant Technology Solutions U.S. Corporation | System and Method For Implementing Modular Universal Reparameterization For Deep Multi-Task Learning Across Diverse Domains |
| US10783405B2 (en) | 2018-10-31 | 2020-09-22 | Salesforce.Com, Inc. | Refinement of machine learning engines for automatically generating component-based user interfaces |
| US10848855B2 (en) | 2018-08-17 | 2020-11-24 | Htc Corporation | Method, electronic device and recording medium for compensating in-ear audio signal |
| CN112185352A (en) | 2020-08-31 | 2021-01-05 | 华为技术有限公司 | Voice recognition method and device and electronic equipment |
| US20210124865A1 (en) | 2016-03-28 | 2021-04-29 | Motivo, Inc. | Design and optimization of physical cell placement for integrated circuits |
| WO2021189990A1 (en) | 2020-03-25 | 2021-09-30 | 李彦涛 | Fitting method and apparatus for hearing earphone |
| CN113689867A (en) | 2021-08-18 | 2021-11-23 | 北京百度网讯科技有限公司 | Training method and device of voice conversion model, electronic equipment and medium |
| TWI749547B (en) | 2020-05-08 | 2021-12-11 | 元智大學 | Speech enhancement system based on deep learning |
| US20220218273A1 (en) * | 2021-01-13 | 2022-07-14 | Anna Barnacka | System and Method for Noninvasive Sleep Monitoring and Reporting |
-
2022
- 2022-05-12 TW TW111117761A patent/TWI820704B/en active
- 2022-05-18 CN CN202210543152.0A patent/CN117113019A/en active Pending
- 2022-06-27 US US17/850,906 patent/US12198714B2/en active Active
Patent Citations (20)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US8332793B2 (en) | 2006-05-18 | 2012-12-11 | Otrsotech, Llc | Methods and systems for placement and routing |
| TW201121495A (en) | 2009-12-16 | 2011-07-01 | Ind Tech Res Inst | Method and system for activity monitoring |
| US20150073306A1 (en) | 2012-03-29 | 2015-03-12 | The University Of Queensland | Method and apparatus for processing patient sounds |
| US20190350532A1 (en) | 2014-07-30 | 2019-11-21 | Valencell, Inc. | Physiological monitoring devices and methods using optical sensors |
| CN107438398A (en) | 2015-01-06 | 2017-12-05 | 大卫·伯顿 | Mobile Wearable Monitoring System |
| US20170105679A1 (en) | 2015-05-20 | 2017-04-20 | Huinno Co., Ltd. | Earphone comprising bio-signal measurement means, and bio-signal monitoring system comprising same |
| US20210124865A1 (en) | 2016-03-28 | 2021-04-29 | Motivo, Inc. | Design and optimization of physical cell placement for integrated circuits |
| JP2020506491A (en) | 2017-01-08 | 2020-02-27 | アルファイクス コーポレイションAlphaics Corporation | System and chip design system and method with artificial intelligence and reinforcement learning |
| US20180235537A1 (en) | 2017-02-21 | 2018-08-23 | Zoll Medical Corporation | Assigning zone-based rankings and actions |
| CN106971042A (en) | 2017-03-31 | 2017-07-21 | 福州大学 | A kind of method that legalizes for being used to mix the design of Height Standard element circuit |
| US20180322386A1 (en) * | 2017-05-05 | 2018-11-08 | Intel Corporation | Fine-grain compute communication execution for deep learning frameworks |
| US10848855B2 (en) | 2018-08-17 | 2020-11-24 | Htc Corporation | Method, electronic device and recording medium for compensating in-ear audio signal |
| US20200118000A1 (en) * | 2018-10-10 | 2020-04-16 | NEC Laboratories Europe GmbH | Method and system for distributed deep learning |
| US10783405B2 (en) | 2018-10-31 | 2020-09-22 | Salesforce.Com, Inc. | Refinement of machine learning engines for automatically generating component-based user interfaces |
| US20200293888A1 (en) * | 2019-03-13 | 2020-09-17 | Cognizant Technology Solutions U.S. Corporation | System and Method For Implementing Modular Universal Reparameterization For Deep Multi-Task Learning Across Diverse Domains |
| WO2021189990A1 (en) | 2020-03-25 | 2021-09-30 | 李彦涛 | Fitting method and apparatus for hearing earphone |
| TWI749547B (en) | 2020-05-08 | 2021-12-11 | 元智大學 | Speech enhancement system based on deep learning |
| CN112185352A (en) | 2020-08-31 | 2021-01-05 | 华为技术有限公司 | Voice recognition method and device and electronic equipment |
| US20220218273A1 (en) * | 2021-01-13 | 2022-07-14 | Anna Barnacka | System and Method for Noninvasive Sleep Monitoring and Reporting |
| CN113689867A (en) | 2021-08-18 | 2021-11-23 | 北京百度网讯科技有限公司 | Training method and device of voice conversion model, electronic equipment and medium |
Non-Patent Citations (2)
| Title |
|---|
| "Office Action of Taiwan Counterpart Application", issued on Feb. 22, 2023, p. 1-p. 5. |
| "Office Action of Taiwan Counterpart Application", issued on Nov. 21, 2023, p. 1-p. 5. |
Also Published As
| Publication number | Publication date |
|---|---|
| TWI820704B (en) | 2023-11-01 |
| TW202345140A (en) | 2023-11-16 |
| US20230368810A1 (en) | 2023-11-16 |
| CN117113019A (en) | 2023-11-24 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| CN111126574B (en) | Method, device and storage medium for training machine learning model based on endoscopic image | |
| US20240346314A1 (en) | End-to-end data format selection for hardware implementation of deep neural network | |
| US10755026B1 (en) | Circuit design including design rule violation correction utilizing patches based on deep reinforcement learning | |
| Brooks et al. | Efficient construction of reversible jump Markov chain Monte Carlo proposal distributions | |
| Gao et al. | Camp: Co-attention memory networks for diagnosis prediction in healthcare | |
| Gupta et al. | Monotonic calibrated interpolated look-up tables | |
| US11275936B2 (en) | Systems and methods for classification of scholastic works | |
| US20230334742A1 (en) | Apparatus, system, and method for generating a video avatar | |
| Selosse et al. | Textual data summarization using the self-organized co-clustering model | |
| US12198714B2 (en) | Voice signal analysis method and device and chip design method and device | |
| CN117349494A (en) | Graph classification method, system, medium and equipment for space graph convolution neural network | |
| US12412073B2 (en) | Supporting database constraints in synthetic data generation based on generative adversarial networks | |
| Yue et al. | Sparse boosting for high‐dimensional survival data with varying coefficients | |
| CN111401569A (en) | Hyper-parameter optimization method and device and electronic equipment | |
| Zhang | Low-dimensional approximation searching strategy for transfer entropy from non-uniform embedding | |
| Wang et al. | Towards actionable risk stratification: A bilinear approach | |
| Guyon | A practical guide to model selection | |
| Nehler et al. | Missing Data Handling via EM and Multiple Imputation in Network Analysis Using Glasso and Atan Regularization | |
| TWI859854B (en) | Method and device for chip design | |
| Zhan et al. | Reliability-enhanced data cleaning in biomedical machine learning using inductive conformal prediction | |
| Lisas et al. | IQGO: Iterative Quantum Gate Optimiser for Quantum Data Embedding 1 | |
| Lipkovich et al. | Statistical data mining of clinical data | |
| Prosperi et al. | Bagged random causal networks for interventional queries on observational biomedical datasets | |
| Wang | Predictive Learning from Real-World Medical Data: Overcoming Quality Challenges | |
| US12346302B1 (en) | Apparatus and method for updating a user data structure |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: INDUSTRIAL TECHNOLOGY RESEARCH INSTITUTE, TAIWAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TAI, LIANG-HSUAN;CHEN, HONG-YU;WU, YEN-TING;AND OTHERS;REEL/FRAME:060416/0622 Effective date: 20220615 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |