[go: up one dir, main page]

WO2017215558A1 - Voiceprint recognition method and device - Google Patents

Voiceprint recognition method and device Download PDF

Info

Publication number
WO2017215558A1
WO2017215558A1 PCT/CN2017/087911 CN2017087911W WO2017215558A1 WO 2017215558 A1 WO2017215558 A1 WO 2017215558A1 CN 2017087911 W CN2017087911 W CN 2017087911W WO 2017215558 A1 WO2017215558 A1 WO 2017215558A1
Authority
WO
WIPO (PCT)
Prior art keywords
character
voice information
verification
voice
voiceprint
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/CN2017/087911
Other languages
French (fr)
Chinese (zh)
Inventor
李为
钱柄桦
金星明
李科
吴富章
吴永坚
黄飞跃
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Publication of WO2017215558A1 publication Critical patent/WO2017215558A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/06Decision making techniques; Pattern matching strategies
    • G10L17/08Use of distortion metrics or a particular distance between probe pattern and reference templates
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/02Preprocessing operations, e.g. segment selection; Pattern representation or modelling, e.g. based on linear discriminant analysis [LDA] or principal components; Feature selection or extraction
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/04Training, enrolment or model building

Definitions

  • the present invention relates to the field of voice recognition technology, and in particular, to a voiceprint recognition method and apparatus.
  • Voiceprint recognition is a method of biometric information recognition, including user registration and user identification.
  • the registration phase maps speech through a series of processes to a user model.
  • the similarity is matched with the model, and then the identity of the unknown voice is consistent with the identity of the registered voice.
  • the existing voiceprint modeling methods usually model from the text-independent level to realize the description of the speaker's identity characteristics, but the text-independent modeling method is difficult to meet when the user reads different content. Claim.
  • the embodiment of the invention provides a voiceprint recognition method and device, which can effectively improve the accuracy of voiceprint recognition.
  • a first aspect of the embodiments of the present invention provides a voiceprint recognition method, where the method includes:
  • the user verification is verified as the registered user corresponding to the registered voice information.
  • the calculating, before verifying, the similarity between the voice segment corresponding to each character in the voice information and the voiceprint feature of the voice segment corresponding to the corresponding character in the preset registration voice information include:
  • the voiceprint features of the voice segments corresponding to the respective characters are extracted.
  • the similarity between the voiceprint feature of the voice segment corresponding to the corresponding character in the preset registration voice information is verified by the calculation:
  • the method before the obtaining the verification voice information generated by the verification user to read the first character string, the method further includes:
  • the feature vector corresponding to each character in the registered voice information is obtained according to the voiceprint feature corresponding to each character in the registered voice information, combined with the common background model corresponding to the preset corresponding character.
  • the voiceprint feature corresponding to the voice segment corresponding to the respective characters is combined with the common background model corresponding to the preset corresponding character to obtain a feature corresponding to each character in the verification voice information.
  • Vectors include:
  • the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and obtain the The feature vector corresponding to each character in the voice information is verified.
  • the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as training sample data, and the maximum a posteriori probability algorithm is used to correspond to the preset corresponding character.
  • the mean supervector of the background model is adjusted, so that the feature vector corresponding to each character in the verification voice information is obtained:
  • the maximum a posteriori probability algorithm is used to adjust the mean supervector of the universal background model corresponding to the preset corresponding character, and combined with the preset supervector subspace matrix, each of the verified voice information is obtained.
  • the feature vector corresponding to the character is obtained.
  • the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to correspond to the preset corresponding character.
  • Adjusting the mean supervector of the universal background model, and combining the preset supervector subspace matrix, obtaining the feature vector corresponding to each character in the verification voice information includes:
  • the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted
  • the general background model corresponding to the corresponding character has the largest posterior probability:
  • M m+T ⁇ , where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector
  • is a feature vector corresponding to the corresponding character in the verification speech information.
  • the preset super-vector sub-space matrix is determined according to a correlation between weights of respective Gaussian modules in the universal background model.
  • the calculating a similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information includes:
  • the performing voice recognition on the verification voice information to obtain the voice segments respectively included in the verification voice information and corresponding to the plurality of characters in the first character string includes:
  • the method before the determining, by the verification user, the registered user corresponding to the registration voice information, the method further includes:
  • the verification user determines the registered user corresponding to the registration voice information.
  • the method further includes: before the obtaining the verification voice information generated by the verification user to read the first character string, the method further includes:
  • the first character string is randomly generated, and the first character string is displayed.
  • a second aspect of the embodiments of the present invention provides a voiceprint recognition apparatus, where the apparatus includes:
  • a voice acquiring module configured to obtain verification voice information generated by the user to read the first character string
  • the similarity judging module is configured to calculate a similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset voice information in the voice information corresponding to each character in the verification voice information;
  • the user identification module is configured to verify the verification user as the registered user corresponding to the registered voice information when the similarity reaches a preset threshold.
  • the voiceprint recognition apparatus further includes:
  • a voice segment identification module configured to perform voice recognition on the verification voice information, and obtain a voice segment respectively included in the verification voice information corresponding to multiple characters in the first character string;
  • the voiceprint feature extraction module is configured to extract a voiceprint feature of the voice segment corresponding to each character in the verification voice information.
  • the voiceprint recognition apparatus further includes:
  • a feature model training module configured to perform, according to the voiceprint feature of the voice segment corresponding to each character, and the common background model corresponding to the corresponding corresponding character, to obtain a feature vector corresponding to each character in the verification voice information
  • the similarity judging module is configured to calculate a similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information, as the verification corresponding to each character in the voice information The similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset registered voice information.
  • the voice acquiring module is further configured to obtain registration voice information generated by a registered user reading a second character string, where the second character string has at least one identical character with the first character string. ;
  • the feature model training module is further configured to: according to each character in the registered voice information
  • the voiceprint feature is combined with a common background model corresponding to the preset corresponding character to obtain a feature vector corresponding to each character in the registered voice information.
  • the feature model training module is configured to:
  • the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and obtain the The feature vector corresponding to each character in the voice information is verified.
  • the feature model training module is configured to:
  • the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character. Combined with the preset super-vector subspace matrix, the feature vector corresponding to each character in the voice information is obtained.
  • the feature model training module is specifically configured to:
  • the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted corresponding
  • the general background model corresponding to the character has the largest posterior probability:
  • M m+T ⁇ , where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector
  • is a feature vector corresponding to the corresponding character in the verification speech information.
  • the preset super-vector sub-space matrix is determined according to a correlation between respective dimension vectors in the mean super-vector of the Gaussian mixture model.
  • the similarity determination module is configured to:
  • the voice segment identification module includes:
  • a valid segment identification unit configured to identify a valid speech segment and an invalid speech segment in the verification speech information
  • the voice recognition unit is configured to perform voice recognition on the valid voice segment to obtain a voice segment corresponding to multiple characters in the first character string.
  • the voiceprint recognition apparatus further includes:
  • a character order determining module configured to determine whether a sorting of the voice segments of the plurality of characters in the verification voice information is consistent with a ranking of the corresponding characters in the first character string
  • the user identification module is further configured to: when the similarity reaches a preset threshold, and verify the order of the voice segments of the plurality of characters in the voice information and the corresponding characters in the first string In the case that the sorting is consistent, the verification user is determined as the registered user corresponding to the registered voice information.
  • the voiceprint recognition apparatus further includes:
  • a string display module configured to randomly generate the first string, and display the first string.
  • a third aspect of the embodiments of the present invention further provides a voiceprint recognition apparatus, including:
  • a processor for invoking the computer executable program code to perform the following operations:
  • the user verification is verified as the registered user corresponding to the registered voice information.
  • the processor before the obtaining the verification voice information generated by the verification user to read the first character string, the processor further calls the computer executable program code to perform the following operations:
  • the voiceprint features of the voice segments corresponding to the respective characters are extracted.
  • the processor invokes the computer executable program code to perform the following operations to calculate a sound segment of a voice segment corresponding to a corresponding character in the preset registered voice information in the voice information of the verification voice information. Similarity of the pattern features:
  • the processor before the obtaining the verification voice information generated by the verification user to read the first character string, the processor further calls the computer executable program code to perform the following operations:
  • the feature vector corresponding to each character in the registered voice information is obtained according to the voiceprint feature corresponding to each character in the registered voice information, combined with the common background model corresponding to the preset corresponding character.
  • the processor invokes the computer executable program code to perform the following operations to combine a common background model corresponding to the preset corresponding character according to a voiceprint feature of the voice segment corresponding to the respective character Training, obtaining a feature vector corresponding to each character in the verification voice information:
  • the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and obtain the The feature vector corresponding to each character in the voice information is verified.
  • the processor invokes the computer executable program code to perform the following operations to use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as training sample data, using the maximum The probability algorithm adjusts the mean supervector of the universal background model corresponding to the preset corresponding character, thereby obtaining a feature vector corresponding to each character in the verified voice information:
  • the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character. And combining the preset super vector subspace matrix to obtain a feature vector corresponding to each character in the verification voice information.
  • the processor invokes the computer executable program code to perform the following operations to use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as Training the sample data, using the maximum posterior probability algorithm to adjust the mean supervector of the universal background model corresponding to the preset corresponding character, and combining the preset supervector subspace matrix to obtain the verification voice
  • the feature vector corresponding to each character in the message :
  • the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted
  • the general background model corresponding to the corresponding character has the largest posterior probability:
  • M m+T ⁇ , where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector
  • is a feature vector corresponding to the corresponding character in the verification speech information.
  • the preset super-vector sub-space matrix is determined according to a correlation between weights of respective Gaussian modules in the universal background model.
  • the processor calls the computer executable program code to perform the following operations to calculate a feature vector corresponding to a feature vector corresponding to each character in the verification voice information and a corresponding character in the preset registration voice information.
  • the similarity scores include:
  • the processor calls the computer executable program code to perform the following operations to perform voice recognition on the verification voice information to obtain the verification voice information included in the first string and the first character string respectively
  • the voice segments corresponding to multiple characters include:
  • the processor before the determining the determined user as the registered user corresponding to the registered voice information, the processor further invokes the computer executable program code to perform the following operations:
  • the similarity reaches a preset threshold, and the plurality of characters in the voice information are verified If the order of the voice segments is consistent with the ordering of the corresponding characters in the first character string, the verification user is determined as the registered user corresponding to the registered voice information.
  • the processor before the obtaining the verification voice information generated by the verification user to read the first character string, the processor further calls the computer executable program code to perform the following operations:
  • the first character string is randomly generated, and the first character string is displayed through the user interface.
  • the fourth aspect of the embodiments of the present invention further provides a storage medium, where the computer program stores a computer program, where the computer program is used to perform any implementation manner according to the first aspect of the embodiments of the present invention.
  • the voiceprint feature of the voice segment corresponding to each character in the verification voice information of the user is obtained, and the UTM training corresponding to the preset corresponding character is used to obtain the feature vector corresponding to each character in the voice message, and the voice information is verified.
  • the feature vector corresponding to each character in the registration is compared with the feature vector of the corresponding character in the registered voice information, thereby determining the user identity of the verified user, and the user feature vector used for comparison corresponds to the specific character, fully considering the user reading aloud
  • the voiceprint features of different characters can effectively improve the accuracy of voiceprint recognition.
  • FIG. 1 is a schematic diagram showing the stages of a voiceprint recognition method in an embodiment of the present invention
  • FIG. 2 is a schematic flow chart of a voiceprint recognition method according to an embodiment of the present invention.
  • FIG. 3 is a schematic diagram showing the principle of identifying a voice segment corresponding to a plurality of characters from voice information in the embodiment of the present invention
  • FIG. 4 is a schematic diagram of a principle for acquiring feature vectors corresponding to respective characters from voice information according to an embodiment of the present invention
  • FIG. 5 is a schematic diagram of a voiceprint registration process of a registered user in an embodiment of the present invention.
  • FIG. 6 is a schematic flow chart of a voiceprint recognition method in another embodiment of the present invention.
  • FIG. 7 is a schematic structural diagram of a voiceprint recognition apparatus according to an embodiment of the present invention.
  • FIG. 8 is a schematic structural diagram of a voice segment identification module in an embodiment of the present invention.
  • FIG. 9 is a schematic structural diagram of another voiceprint recognition apparatus in an embodiment of the present invention.
  • Embodiments of the present invention provide a voiceprint recognition method and apparatus.
  • the voiceprint recognition method and apparatus can be applied to all scenes or devices that need to identify an unknown user.
  • the characters in the string used for voiceprint recognition may be Arabic numerals, English letters or other language characters.
  • the characters in the embodiments of the present invention are exemplified by taking Arabic numerals as an example.
  • the voiceprint recognition method in the embodiment of the present invention can be divided into two stages, as shown in FIG. 1:
  • the registered user can read a registration string (ie, the second character string appearing later), and the voiceprint recognition device collects the registered voice information when the registered user reads the registration string, and then registers the voice.
  • UBM Universal Background Model
  • the voiceprint recognition device can respectively save the feature vectors corresponding to the plurality of characters in the registered voice information that the different registered users read in the voiceprint registration stage in the model library of the voiceprint recognition device.
  • the registration string is a numeric string 0185851, and includes four numbers “0”, “1”, “5”, “8”, and the voiceprint recognition device performs sound according to the voice segment corresponding to each character in the registered voice information.
  • the pattern feature extraction and the voiceprint model training obtain the voiceprint features of the voice segments corresponding to “0”, “1”, “5”, and “8”, and then obtain the registered voice information in combination with the UBM training corresponding to the preset corresponding characters.
  • Individual characters Corresponding feature vectors include a feature vector corresponding to the number “0”, a feature vector corresponding to the number “1”, a feature vector corresponding to the number “5”, and a feature vector corresponding to the number “8”.
  • the user who authenticates the user reads a verification string (ie, the first character string that appears later, and the second character string has at least one of the same characters as the first character string).
  • the pattern recognition device collects the verification voice information when the verification user reads the verification character string, and then performs voice recognition on the verification voice information to obtain corresponding to the plurality of characters in the verification string respectively included in the verification voice information.
  • the voice segment, and then the voiceprint feature extraction and the voiceprint model training for the voice segment corresponding to each character, including the voiceprint feature corresponding to the voice segment corresponding to the respective character, and the UBM training corresponding to the corresponding corresponding character is used to obtain the verification voice.
  • the feature vector corresponding to each character in the information and finally calculating the similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information, if the similarity score reaches the preset verification a threshold, the verification user is determined as a note corresponding to the registration voice information Users.
  • the voiceprint recognition device performs voiceprint feature extraction and voiceprint model training according to the voice segment corresponding to each character in the verification voice information generated when the user reads aloud, and obtains “0” and “ The GMM corresponding to 1", "5", and “8", and then combined with the preset UBM corresponding to the corresponding character, can calculate the feature vector of the verified voice information of the verified user, including the feature vector corresponding to the number "0", and the number a feature vector corresponding to "1", a feature vector corresponding to the number "5", and a feature vector corresponding to the number "8", thereby respectively calculating "0", "1", "5", "8” in the verification voice information a similarity score between the corresponding feature vectors and the feature vectors corresponding to "0", “1", "5", and “8” in the registered voice information, and if the similarity score reaches a preset verification threshold, And determining, by the verification user, a registered user
  • the voiceprint registration phase of the registered user and the identity phase of the authenticated user may be implemented in the same device or device, or may be implemented in different devices or devices, for example, the voiceprint registration phase of the registered user is
  • the first device is implemented, and the first device sends the feature vector corresponding to the multiple characters in the registered voice information to the second device, so that the identity recognition phase of the authenticated user can be implemented in the second device.
  • FIG. 2 is a schematic flow chart of a method for identifying a voiceprint according to an embodiment of the present invention. As shown in the figure, the flow of the voiceprint recognition method in the embodiment may include:
  • the user who authenticates the user needs to verify the identity of the user through the voiceprint recognition device.
  • the first character string is a character string used for verifying the user's identity verification, and may be randomly generated, or may be a preset fixed string, for example, a second string corresponding to the pre-generated registration voice information. Part of the same string.
  • the first character string may include m characters, wherein n characters are different from each other, m and n are positive integers, and m ⁇ n.
  • the first string is "12358948", a total of 8 characters, including 7 different characters "1", “2", “3", “4", "5", "8", " 9".
  • the voiceprint recognition device may generate and display the first character string for the verification user to read aloud according to the displayed first character string.
  • S202 Perform speech recognition on the verification voice information to obtain a voice segment respectively included in the verification voice information and corresponding to multiple characters in the first character string.
  • the voiceprint recognition device can divide the verification voice information into voice segments corresponding to multiple characters through voice recognition and voice intensity filtering, and optionally remove the invalid voice segments, and do not participate in subsequent Process.
  • the voiceprint recognition device may extract MFCC (Mel Frequency Cepstrum Coefficient) or PLP (Perceptual Linear Predictive Coefficient) in the voice segment corresponding to each character as the voice corresponding to each character.
  • MFCC Mel Frequency Cepstrum Coefficient
  • PLP Perceptual Linear Predictive Coefficient
  • the universal background model UBM in the embodiment of the present invention is a mixed Gaussian model in which a plurality of speech segments of a plurality of speakers are mixed and trained, and the distribution of the corresponding digital speech in the feature space is represented, and The data comes from a large number of speakers, so it does not represent a specific type of speaker, and has an identity-independent nature, which can be regarded as a general background model.
  • a speech sample with more than 1000 speakers and a duration of more than 20 hours can be used, and each character is out.
  • the current frequency is relatively balanced and training is obtained by UBM.
  • the mathematical expression of UBM is:
  • P(x) represents the probability distribution of UBM
  • C represents the total of C Gaussian modules in UBM
  • a i represents the weight of the i-th Gaussian module
  • ⁇ i represents the mean of the i-th Gaussian module
  • ⁇ i represents The variance of the i-th Gaussian module
  • N(x) represents a Gaussian distribution
  • x represents the voiceprint characteristics of the input sample.
  • the voiceprint recognition device may use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, and adopt a maximum a posteriori probability algorithm (MAP) to the common background model corresponding to the preset corresponding character.
  • MAP maximum a posteriori probability algorithm
  • the parameter is adjusted, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the posterior check is made by continuously adjusting the parameters of the common background model corresponding to the corresponding corresponding character.
  • the probability P(x) is the largest, so that the feature vector corresponding to the corresponding character in the verification speech information can be determined according to the parameter that makes the posterior probability P(x) the largest.
  • the mean supervector of the UBM model is defined as:
  • the voiceprint recognition device can use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, and adopt the maximum posterior probability algorithm (MAP) to the common background corresponding to the preset corresponding character.
  • MAP maximum posterior probability algorithm
  • the mean value supervector of the model is adjusted, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the posterior probability P(x) is adjusted by continuously adjusting the mean supervector.
  • the maximum so that the mean supervector that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the verification speech information.
  • the profiling of the mean supervector is limited to a subspace by probabilistic principal component analysis (PPCA), and the voiceprint recognition device can verify the voiceprint features of the speech segments corresponding to each character in the speech information.
  • PPCA probabilistic principal component analysis
  • the voiceprint recognition device can verify the voiceprint features of the speech segments corresponding to each character in the speech information.
  • the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the corresponding corresponding character
  • the preset super-vector subspace matrix is used to obtain the corresponding features of each character in the verified voice information. vector.
  • the mean supervector of the common background model corresponding to the preset corresponding character may be adjusted by using the following formula, so that the posterior probability of the universal background model corresponding to the adjusted corresponding character is the largest:
  • M m+T ⁇
  • M represents the mean supervector of the universal background model of the adjusted character
  • m represents the mean supervector of the universal background model of the corresponding character before adjustment
  • T is the preset supervector subspace matrix
  • is the feature vector corresponding to the corresponding character in the verification voice information, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the adjustment can be realized by continuously adjusting ⁇ .
  • the mean supervector in (1) maximizes the posterior probability P(x), so that ⁇ that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the verification speech information.
  • the super-vector subspace matrix T is determined according to the correlation between each dimension vector in the mean supervector of the Gaussian mixture model.
  • the voiceprint recognition device can obtain the registered voice information of the registered user in the voiceprint registration stage, and obtain the voiceprint feature extraction and the voiceprint model training similar to the embodiment, so that the characters in the registered voice information can be obtained.
  • the registration voice information may be that the voiceprint recognition device acquires the registration voice information generated by the registered user reading the second character string, and the second character string has at least one of the same characters as the first character string, that is, the The second character string corresponding to the registration voice information is at least partially identical to the first character string.
  • the voiceprint recognition device may further acquire the feature vector corresponding to the corresponding character in the registered voice information from the outside, that is, after the registered user inputs the registered voice information through other devices, the other device or the server passes the voiceprint.
  • Feature extraction and voiceprint model training obtain feature vectors corresponding to voice segments of each character in the registered voice information, and the voiceprint recognition device acquires the phase in the registered voice information by using other devices or servers The feature vector corresponding to the character is used to compare the feature vector corresponding to each character in the verification voice information in the identification phase of the verification user.
  • the similarity score is that the voiceprint recognition device compares the feature vector corresponding to each character in the voice information with the feature vector corresponding to the corresponding character in the preset registered voice information, and then measures two features of the same character. The degree of similarity between vectors.
  • a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information may be calculated, and the cosine distance value is used as the
  • the similarity score that is, the similarity score between the corresponding feature vector in the verification voice information and the feature vector in the registered voice information is calculated by the following formula:
  • the subscript i indicates a character common to the i-th verification voice information and the registration voice information
  • ⁇ i (tar) indicates a corresponding feature vector of the character in the verification voice information
  • ⁇ i (test) indicates that the character is in the registered voice.
  • the registered user may be authenticated according to the similarity between the feature vector of a certain character of the user and the corresponding character of each registered user.
  • the feature vector of the corresponding character and the feature vector of the character of the verification voice have the highest similarity score and the similarity reaches the preset verification threshold, and the registered user is used as the identification result of the verification user.
  • the same character appears in the verification voice information more than once, for example, if 0, 1, 5, and 8 appear in the verification voice information as shown in FIG. 2, respectively.
  • the average of the similarity scores of the feature vectors processed by the two-character 0 corresponding to the speech segment and the feature vector of the character 0 in the preset registration speech information is used as the feature vector and preset of the character 0 in the verification speech information.
  • the similarity score of the feature vector of character 0 in the registered voice information and so on.
  • the UTM training corresponding to the preset corresponding character is used to obtain the feature vector corresponding to each character in the verification voice information, and the verification is performed.
  • the feature vector corresponding to each character in the voice information is compared with the feature vector of the corresponding character in the registered voice information, thereby determining the user identity of the verified user.
  • the user feature vector used for comparison corresponds to the specific character, and the user is fully considered.
  • the voiceprint features of different characters are read aloud, so that the accuracy of voiceprint recognition can be effectively improved.
  • FIG. 5 is a schematic diagram of a voiceprint registration process of a registered user in the embodiment of the present invention. As shown in the figure, the voiceprint registration process in this embodiment may include:
  • the registered user is a user who determines a legal identity.
  • the second character string is a string used to collect the voiceprint feature vector of the registered user, and may be randomly generated or a fixed string.
  • the second character string may also include m characters, wherein there are n characters different from each other, m and n are positive integers, and m ⁇ n.
  • the voiceprint recognition device may generate and display the second character string for the registered user to read aloud according to the displayed second character string.
  • the voiceprint recognition device can divide the verification voice information into voice segments corresponding to multiple characters through voice recognition and voice intensity filtering, and optionally remove the invalid voice segments, and do not participate in subsequent processing.
  • the voiceprint recognition device may extract MFCC (Mel Frequency Cepstrum Coefficient) or PLP (Perceptual Linear Predictive Coefficient) in the voice segment corresponding to each character as the voice corresponding to each character.
  • MFCC Mel Frequency Cepstrum Coefficient
  • PLP Perceptual Linear Predictive Coefficient
  • UBM can refer to the previous embodiment.
  • This step of the voiceprint registration process is similar to S204 of the voiceprint recognition process.
  • the voiceprint recognition device can use the voiceprint feature of the voice segment corresponding to each character in the registered voice information as the training sample data, and adopt the maximum posterior probability algorithm (Maximum A).
  • Posteriori, MAP adjusts the parameters of the common background model corresponding to the corresponding corresponding characters, that is, after the voiceprint feature of the voice segment corresponding to each character in the registered voice information is substituted into the formula (1) as an input sample,
  • the parameters of the common background model corresponding to the preset corresponding characters are such that the posterior probability P(x) is the largest, so that the feature vector corresponding to the corresponding character in the registered voice information can be determined according to the parameter that makes the posterior probability P(x) the largest.
  • the voiceprint recognition device can use the voiceprint feature of the voice segment corresponding to each character in the registered voice information as the training sample data, using the maximum posterior
  • the Maximum A Posteriori (MAP) adjusts the mean supervector of the universal background model corresponding to the preset corresponding characters, that is, the voiceprint feature of the voice segment corresponding to each character in the registered voice information is used as an input sample substitution type ( 1) After, by continuously adjusting the mean supervector, the posterior probability P(x) is maximized, so that the mean supervector that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the registered speech information.
  • MAP Maximum A Posteriori
  • the mean supervector of the universal background model corresponding to the preset corresponding character may be adjusted by using the following formula, so that the posterior probability of the universal background model corresponding to the adjusted corresponding character is the largest:
  • M m+T ⁇
  • M represents the mean supervector of the universal background model of the adjusted character
  • m represents the mean supervector of the universal background model of the corresponding character before adjustment
  • T is the preset supervector subspace matrix
  • is the feature vector corresponding to the corresponding character in the registered voice information, that is, after the voiceprint feature of the voice segment corresponding to each character in the registered voice information is substituted into the formula (1) as an input sample, the adjustment can be realized by continuously adjusting ⁇ .
  • the mean supervector in (1) maximizes the posterior probability P(x), so that ⁇ that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the registered speech information.
  • FIG. 6 is a schematic flow chart of a voiceprint recognition method according to another embodiment of the present invention, as shown in the figure.
  • the voiceprint recognition method in the embodiment may include the following processes:
  • the verification voice may be divided according to the sound intensity, and the voice segment with less sound intensity is regarded as an invalid voice segment (for example, including a silent segment and impulse noise).
  • a voice segment respectively corresponding to a plurality of characters in the first character string can be obtained by voice recognition.
  • a different first character string may be randomly generated each time, and in this step, the voice segment of the plurality of characters in the voice information is determined. Whether the sorting is consistent with the sorting of the corresponding characters in the first string, if not, it can be judged that the voiceprint recognition fails, and if the sorting of the corresponding characters in the first string is consistent, the subsequent process is performed.
  • the voiceprint recognition device may extract MFCC (Mel Frequency Cepstrum Coefficient) or PLP (Perceptual Linear Predictive Coefficient) in the voice segment corresponding to each character as the voice corresponding to each character.
  • MFCC Mel Frequency Cepstrum Coefficient
  • PLP Perceptual Linear Predictive Coefficient
  • the voiceprint feature of the voice segment corresponding to each character in the voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the corresponding corresponding character, so that the estimation is verified.
  • a feature vector corresponding to each character in the voice message is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the corresponding corresponding character, so that the estimation is verified.
  • the voiceprint recognition device can verify the voiceprint feature of the voice segment corresponding to each character in the voice information as a training sample.
  • Data using Maximum A Posteriori (MAP) to adjust the mean supervector of the common background model corresponding to the corresponding corresponding character, that is, the voiceprint feature of the voice segment corresponding to each character in the verification voice information
  • MAP Maximum A Posteriori
  • the voiceprint recognition apparatus may adjust the mean super vector of the common background model corresponding to the preset corresponding character by using the following formula:
  • the posterior probability of the universal background model corresponding to the adjusted corresponding characters is the largest:
  • M m+T ⁇
  • M represents the mean supervector of the universal background model of the adjusted character
  • m represents the mean supervector of the universal background model of the corresponding character before adjustment
  • T is the preset supervector subspace matrix
  • is the feature vector corresponding to the corresponding character in the verification voice information, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the adjustment can be realized by continuously adjusting ⁇ .
  • the mean supervector in (1) maximizes the posterior probability P(x), so that ⁇ that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the verification speech information.
  • the voiceprint recognition apparatus may calculate a cosine distance value between the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, and use the cosine distance value as
  • the similarity score that is, the similarity score between the corresponding feature vector in the verification voice information and the feature vector in the registered voice information is calculated by the following formula:
  • the subscript i indicates a character common to the i-th verification voice information and the registration voice information
  • ⁇ i (tar) indicates a corresponding feature vector of the character in the verification voice information
  • ⁇ i (test) indicates that the character is in the registered voice.
  • the registered user may be authenticated according to the similarity between the feature vector of a certain character of the user and the corresponding character of each registered user.
  • the feature vector of the corresponding character and the feature vector of the character of the verification voice have the highest similarity score and the similarity reaches the preset verification threshold, and the registered user is used as the identification result of the verification user.
  • the user identity of the user can be further verified. accuracy.
  • FIG. 7 is a schematic structural diagram of a voiceprint identifying apparatus according to an embodiment of the present invention. As shown in the figure, the voiceprint identifying apparatus in this embodiment may include:
  • the voice obtaining module 710 is configured to obtain verification voice information generated by the user to read the first character string.
  • the user who authenticates the user needs to verify the identity of the user through the voiceprint recognition device.
  • the first character string is a character string used for verifying the user's identity verification, and may be randomly generated, or may be a preset fixed string, for example, a second string corresponding to the pre-generated registration voice information. Part of the same string.
  • the first character string may include m characters, wherein n characters are different from each other, m and n are positive integers, and m ⁇ n.
  • the first string is "12358948", a total of 8 characters, including 7 different characters "1", “2", “3", “4", "5", "8", " 9".
  • the voice segment identification module 720 is configured to perform voice recognition on the verification voice information to obtain voice segments respectively corresponding to the plurality of characters in the first character string included in the verification voice information.
  • the voice segment identification module 720 can divide the verification voice information into voice segments corresponding to multiple characters through voice recognition and voice intensity filtering, and optionally remove the invalid voice segments without participating. Subsequent processing.
  • the voice segment identification module may further include:
  • the valid segment identification unit 721 is configured to identify a valid speech segment and an invalid speech segment in the verification speech information.
  • the effective segment recognizing unit 721 can divide the verification speech according to the sound intensity, and treat the speech segment with a small sound intensity as an invalid speech segment (for example, including a silent segment and impulse noise).
  • a voice recognition unit 722 configured to perform voice recognition on the valid voice segment to obtain a separate voice A speech segment corresponding to a plurality of characters in the first character string.
  • the voiceprint feature extraction module 730 is configured to extract a voiceprint feature of the voice segment corresponding to each character in the verification voice information.
  • the voiceprint feature extraction module 730 can extract MFCC (Mel Frequency Cepstrum Coefficient) or PLP (Perceptual Linear Predictive Coefficient) in the voice segment corresponding to each character, as corresponding to each character.
  • MFCC Mel Frequency Cepstrum Coefficient
  • PLP Perceptual Linear Predictive Coefficient
  • the feature model training module 740 is configured to train, according to the voiceprint feature of the voice segment corresponding to the respective characters, a feature vector corresponding to each character in the verification voice information according to the common background model corresponding to the preset corresponding character.
  • the feature model training module 740 may use the voiceprint feature of the voice segment corresponding to each character in the voice information as the training sample data, and adopt a maximum posterior probability algorithm (MAP) to the common background model corresponding to the preset corresponding character.
  • MAP maximum posterior probability algorithm
  • the parameters are adjusted, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the parameters of the common background model corresponding to the corresponding corresponding characters are continuously adjusted, so that The probability P(x) is the largest, so that the feature model training module 740 can determine the feature vector corresponding to the corresponding character in the verification voice information according to the parameter that maximizes the posterior probability P(x).
  • the feature model training module 740 can use the voiceprint feature of the voice segment corresponding to each character in the voice information as the training sample data, and use the Maximum A Posteriori (MAP) algorithm to correspond to the preset corresponding character.
  • MAP Maximum A Posteriori
  • the mean supervector of the background model is adjusted In the whole, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1), the posterior probability P(x) is maximized by continuously adjusting the mean supervector, and the feature model training module 740 may use the mean supervector that maximizes the posterior probability P(x) as the feature vector corresponding to the corresponding character in the verification speech information.
  • the feature model training module 740 may use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, and adopt the maximum posterior probability algorithm to the common background model corresponding to the preset corresponding character.
  • the mean super vector is adjusted, and the preset super vector subspace matrix is combined to obtain a feature vector corresponding to each character in the verified speech information.
  • the feature model training module 740 may adjust the mean super vector of the common background model corresponding to the preset corresponding character by using the following formula, so that the posterior probability of the common background model corresponding to the adjusted corresponding character is the largest:
  • M m+T ⁇
  • M represents the mean supervector of the universal background model of the adjusted character
  • m represents the mean supervector of the universal background model of the corresponding character before adjustment
  • T is the preset supervector subspace matrix
  • is the feature vector corresponding to the corresponding character in the verification voice information, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the adjustment can be realized by continuously adjusting ⁇ .
  • the mean supervector in (1) maximizes the posterior probability P(x), so that ⁇ that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the verification speech information.
  • the super-vector subspace matrix T is determined according to the correlation between each dimension vector in the mean supervector of the Gaussian mixture model.
  • the similarity determining module 750 is configured to calculate a similarity score of the feature vector corresponding to the corresponding character in the preset registered voice information in the verification voice information.
  • the voiceprint recognition device can obtain the registered voice information of the registered user in the voiceprint registration stage, and can obtain the registered voice information through the voice segment identification module 720, the voiceprint feature extraction module 730, and the feature model training module 740.
  • the registration voice information may be that the voiceprint recognition device acquires the registration voice information generated by the registered user reading the second character string, and the second character string has at least one of the same characters as the first character string, that is, the The second character string corresponding to the registration voice information is at least partially identical to the first character string.
  • the voiceprint recognition device may further acquire a feature vector corresponding to the corresponding character in the registered voice information, that is, after the registered user inputs the registered voice information through other devices, the other device or the server passes the voiceprint feature.
  • the extraction and voiceprint model training obtains a feature vector corresponding to the voice segment of each character in the registered voice information, and the voiceprint recognition device obtains the feature vector corresponding to the corresponding character in the registered voice information from other devices or servers, thereby verifying the user
  • the identification stage similarity judgment module 750 is configured to compare the feature vectors corresponding to the respective characters in the verification voice information.
  • the similarity score is that the voiceprint recognition device compares the feature vector corresponding to each character in the voice information with the feature vector corresponding to the corresponding character in the preset registered voice information, and then measures two features of the same character. The degree of similarity between vectors.
  • the similarity determination module 750 may calculate a cosine distance value between the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, and the cosine distance value is The distance value is used as the similarity score, that is, the similarity score between the corresponding feature vector in the verification voice information and the feature vector in the registered voice information is calculated by the following formula:
  • the subscript i indicates a character common to the i-th verification voice information and the registration voice information
  • ⁇ i (tar) indicates a corresponding feature vector of the character in the verification voice information
  • ⁇ i (test) indicates that the character is in the registered voice.
  • the corresponding feature vector in the message if the same character appears in the verification voice information more than once, for example, if 0, 1, 5, and 8 appear in the verification voice information as shown in FIG. 2, respectively, then The average of the similarity scores of the feature vectors processed by the two-character 0 corresponding to the speech segment and the feature vector of the character 0 in the preset registration speech information is used as the feature vector and preset of the character 0 in the verification speech information.
  • the similarity score of the feature vector of character 0 in the registered voice information and so on.
  • a user identification module 760 configured to: if the similarity score reaches a preset verification threshold, The verification user determines the registered user corresponding to the registered voice information.
  • the user identification module 760 may take the average value of the similarity scores of the respective characters calculated by the similarity determination module 750, if the average value of the similarity scores of the respective characters reaches the corresponding value.
  • the preset verification threshold determines the authenticated user as the registered user corresponding to the registered voice information. If there are multiple registered users, such as the registered users A, B, and C shown in FIG. 1, the user identification module 760 may perform the similarity between the feature vector of a certain character of the user and the feature vector of the corresponding character of each registered user. If the feature vector of the corresponding character of a registered user and the feature vector of the character of the verification voice have the highest similarity score and the similarity reaches the preset verification threshold, the registered user is used as the identification result of the verification user.
  • the voice acquiring module 710 is further configured to obtain registration voice information generated by the registered user reading the second character string, where the second character string has at least one of the same as the first character string. character of;
  • the voice segment identification module 720 is further configured to perform voice recognition on the registration voice information to obtain a voice segment respectively included in the registration voice information and corresponding to multiple characters in the second character string;
  • the voiceprint feature extraction module 730 is further configured to extract a voiceprint feature of a voice segment corresponding to each character in the registered voice information;
  • the feature model training module 740 is further configured to: according to the voiceprint feature of the voice segment corresponding to each character in the registered voice information, and the common background model corresponding to the preset corresponding character, the corresponding characters in the registered voice information are trained. Feature vector.
  • the voiceprint recognition apparatus may further include:
  • the character order determining module 770 is configured to determine whether the order of the voice segments of the plurality of characters in the verification voice information is consistent with the order of the corresponding characters in the first character string.
  • a different first character string may be randomly generated each time, and in this step, the voice segment of the plurality of characters in the voice information is determined. Whether the sorting is consistent with the sorting of the corresponding characters in the first string, if not, it may be determined that the voiceprint recognition fails, and if the sorting of the corresponding characters in the first character string is consistent, the voiceprint feature extraction module 730 may be notified. Or feature model training module 740 performs feature extraction and voiceprint training for the verification voice information.
  • the voiceprint recognition apparatus may further include:
  • the string display module 700 is configured to randomly generate the first string and display it.
  • the UTM training corresponding to the preset corresponding character is used to obtain the feature vector corresponding to each character in the verification voice information, and the verification is performed.
  • the feature vector corresponding to each character in the voice information is compared with the feature vector of the corresponding character in the registered voice information, thereby determining the user identity of the verified user.
  • the user feature vector used for comparison corresponds to the specific character, and the user is fully considered.
  • the voiceprint features of different characters are read aloud, so that the accuracy of voiceprint recognition can be effectively improved.
  • FIG. 9 is a schematic structural diagram of another voiceprint recognition apparatus according to an embodiment of the present invention.
  • the voiceprint recognition apparatus 1000 may include at least one processor 1001 (eg, a CPU), a user interface 1003, a memory 1005, and at least one communication bus 1002.
  • the communication bus 1002 is used to implement connection communication between these components.
  • the user interface 1003 may include a display, a keyboard, a microphone, and the like.
  • the user interface 1003 may further include a standard wired interface and a wireless interface.
  • the memory 1005 may be a high speed RAM memory or a non-volatile memory such as at least one disk memory.
  • an operating system, a user interface module, and computer executable program code eg, a voiceprint recognition program
  • the processor 1001 can be used to call the computer executable program code stored in the memory 1005, and specifically perform the following steps:
  • the user verification is verified as the registered user corresponding to the registered voice information.
  • the processor 1001 also calls the computer executable program code to perform the following operations:
  • the voiceprint features of the voice segments corresponding to the respective characters are extracted.
  • the processor 1001 invokes the computer executable program code to perform the following operations to calculate a voice segment corresponding to a corresponding one of the preset voice information in the voice message corresponding to each character in the verification voice information. Similarity of voiceprint features:
  • the processor 1001 before the obtaining the verification voice information generated by the verification user to read the first character string, the processor 1001 further invokes the computer executable program code to perform the following operations: acquiring Registering a user to read the registered voice information generated by the second character string, the second character string having at least one of the same characters as the first character string; performing voice recognition on the registered voice information to obtain the registered voice information a voice segment respectively corresponding to the plurality of characters in the second character string; extracting a voiceprint feature of the voice segment corresponding to each character in the registered voice information; and corresponding to each character in the registered voice information
  • the voiceprint feature of the voice segment is trained in conjunction with a common background model corresponding to the corresponding corresponding character to obtain a feature vector corresponding to each character in the registered voice information.
  • the processor 1001 invokes the computer executable program code to perform the following operations, according to the voiceprint feature of the voice segment corresponding to the respective characters, and the common background corresponding to the preset corresponding character Model training, obtaining a feature vector corresponding to each character in the verification voice information: using a voiceprint feature of a voice segment corresponding to each character in the verification voice information as training sample data, and using a maximum posterior probability algorithm on the preset
  • the mean supervector of the universal background model corresponding to the corresponding character is adjusted to obtain a feature vector corresponding to each character in the verified voice information.
  • the processor 1001 invokes the computer executable program code to perform the following operations to use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as training sample data, using the maximum The posterior probability algorithm adjusts the mean supervector of the universal background model corresponding to the preset corresponding character, thereby obtaining a feature vector corresponding to each character in the verification voice information: corresponding to each character in the verification voice information
  • the voiceprint feature of the voice segment is used as the training sample data
  • the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and combined with the preset supervector subspace matrix. Obtaining a feature vector corresponding to each character in the verification voice information.
  • the processor 1001 invokes the computer executable program code to perform the following operations to use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data,
  • the maximum posterior probability algorithm adjusts the mean supervector of the universal background model corresponding to the preset corresponding character, and combines the preset supervector subspace matrix to obtain corresponding characters of the verified voice information.
  • Feature vector using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, and adjusting the mean super vector of the common background model corresponding to the preset corresponding character by using the following formula:
  • Mean supervector of the background model, T is the preset supervector subspace matrix, and ⁇ is the verification speech information Eigenvectors corresponding to the respective characters.
  • the preset super-vector sub-space matrix is determined according to a correlation between weights of respective Gaussian modules in the universal background model.
  • the processor 1001 invokes the computer executable program code to perform the following operations to calculate that a feature vector corresponding to each character in the verification voice information corresponds to a corresponding character in the preset registration voice information.
  • the similarity score of the feature vector calculating a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information, and The cosine distance value is determined as the similarity score.
  • the processor 1001 invokes the computer executable program code to perform the following operations to perform voice recognition on the verification voice information to obtain the verification voice information package.
  • a speech segment respectively corresponding to the plurality of characters in the first character string identifying a valid speech segment and an invalid speech segment in the verification speech information; performing speech recognition on the valid speech segment respectively A speech segment corresponding to a plurality of characters in the first character string.
  • the processor 1001 before the verifying the user is determined as the registered user corresponding to the registered voice information, the processor 1001 further invokes the computer executable program code to perform the following operations: determining the Verifying whether the ordering of the voice segments of the plurality of characters in the voice information is consistent with the ordering of the corresponding characters in the first character string; and when the similarity reaches a preset threshold value, and the verification voice information is In the case that the ordering of the speech segments of the plurality of characters is consistent with the ordering of the corresponding characters in the first character string, the verification user is determined as the registered user corresponding to the registration voice information.
  • the processor 1001 before the obtaining the verification voice information generated by the verification user to read the first character string, the processor 1001 further invokes the computer executable program code to perform the following operations: random Generating the first character string and displaying the first character string.
  • the storage medium may be a magnetic disk, an optical disk, a read-only memory (ROM), or a random access memory (RAM).

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Game Theory and Decision Science (AREA)
  • Business, Economics & Management (AREA)
  • Telephonic Communication Services (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Telephone Function (AREA)

Abstract

A voiceprint recognition method and device. The method comprises: obtaining authentication voice information generated by reading a first character string by an authenticated user (S201); performing voice recognition on the authentication voice information to obtain video segments included in the authentication voice information and respectively corresponding to a plurality of characters in the first character string (S202); extracting a voiceprint feature of the video segment corresponding to each character (S203); obtaining, according to the voiceprint feature of the video segment corresponding to each character, a feature vector corresponding to each character in the authentication voice information on the basis of preset general background model training corresponding to a corresponding character (S204); and calculating a similarity score between the feature vector corresponding to each character in the authentication voice information and the feature vector corresponding to a corresponding character in preset registration voice information, and if the similarity score reaches a preset authentication threshold, determining the authenticated user to be a registered user corresponding to the registration voice information (S205). The voiceprint recognition accuracy can be effectively improved.

Description

一种声纹识别方法和装置Voiceprint recognition method and device

本发明要求2016年6月12日递交的发明名称为“一种声纹识别方法和装置”的申请号201610416650.3的在先申请优先权,上述在先申请的内容以引入的方式并入本文本中。The present invention claims priority to the priority of the application No. 201610416650.3, entitled "A Voiceprint Identification Method and Apparatus", filed on June 12, 2016, the contents of which are incorporated herein by reference. .

技术领域Technical field

本发明涉及声音识别技术领域,尤其涉及一种声纹识别方法和装置。The present invention relates to the field of voice recognition technology, and in particular, to a voiceprint recognition method and apparatus.

背景技术Background technique

声纹识别作为一种生物信息识别的方法,包括用户注册和用户身份识别两个阶段。注册阶段将语音通过一系列处理映射为用户模型。在识别阶段对于一段身份未知的语音,与模型进行相似度的匹配,进而对未知语音的身份与注册语音的身份是否一致进行判断。现有的声纹建模方法通常是从文本无关的层面进行建模以实现对说话人身份特征的描述,但是文本无关的建模方式在用户朗读不同内容时,识别准确率较低,难以满足要求。Voiceprint recognition is a method of biometric information recognition, including user registration and user identification. The registration phase maps speech through a series of processes to a user model. In the identification stage, for a voice with unknown identity, the similarity is matched with the model, and then the identity of the unknown voice is consistent with the identity of the registered voice. The existing voiceprint modeling methods usually model from the text-independent level to realize the description of the speaker's identity characteristics, but the text-independent modeling method is difficult to meet when the user reads different content. Claim.

发明内容Summary of the invention

有鉴于此,本发明实施例提供一种声纹识别方法和装置,可有效提高声纹识别准确率。In view of this, the embodiment of the invention provides a voiceprint recognition method and device, which can effectively improve the accuracy of voiceprint recognition.

为了解决上述技术问题,本发明实施例第一方面提供了一种声纹识别方法,所述方法包括:In order to solve the above technical problem, a first aspect of the embodiments of the present invention provides a voiceprint recognition method, where the method includes:

获取验证用户朗读第一字符串所产生的验证语音信息;Obtaining verification voice information generated by the verification user reading the first character string;

计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度;Calculating a similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset voice information in the voice information of the verification voice information;

当所述相似度达到预设的门限值时,将验证用户验证为所述注册语音信息对应的注册用户。When the similarity reaches a preset threshold, the user verification is verified as the registered user corresponding to the registered voice information.

在一种实现方式中,所述计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度之前还 包括:In an implementation manner, the calculating, before verifying, the similarity between the voice segment corresponding to each character in the voice information and the voiceprint feature of the voice segment corresponding to the corresponding character in the preset registration voice information include:

对所述验证语音信息进行语音识别,得到所述验证语音信息中包含的分别与所述第一字符串中的各个字符对应的语音片段;Performing voice recognition on the verification voice information, and obtaining a voice segment respectively corresponding to each character in the first character string included in the verification voice information;

提取各个字符对应的语音片段的声纹特征。The voiceprint features of the voice segments corresponding to the respective characters are extracted.

在一种实现方式中,所述计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度包括:In an implementation manner, the similarity between the voiceprint feature of the voice segment corresponding to the corresponding character in the preset registration voice information is verified by the calculation:

根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量;And performing, according to the voiceprint feature of the voice segment corresponding to each character, in combination with a common background model corresponding to the preset corresponding character, obtaining a feature vector corresponding to each character in the verification voice information;

计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量的相似度分数,作为验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度。Calculating a similarity score of the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, as a voice segment corresponding to each character in the verification voice information and a preset registration The similarity of the voiceprint features of the voice segments corresponding to the corresponding characters in the voice information.

在一种实现方式中,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述方法还包括:In an implementation manner, before the obtaining the verification voice information generated by the verification user to read the first character string, the method further includes:

获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符;Obtaining registration voice information generated by the registered user reading the second character string, the second character string having at least one of the same characters as the first character string;

根据所述注册语音信息中各个字符对应的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述注册语音信息中各个字符对应的特征向量。The feature vector corresponding to each character in the registered voice information is obtained according to the voiceprint feature corresponding to each character in the registered voice information, combined with the common background model corresponding to the preset corresponding character.

在一种实现方式中,所述根据所述各个字符对应的语音片段的声纹特征,结合所述预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量包括:In an implementation manner, the voiceprint feature corresponding to the voice segment corresponding to the respective characters is combined with the common background model corresponding to the preset corresponding character to obtain a feature corresponding to each character in the verification voice information. Vectors include:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,得到所述验证语音信息中各个字符对应的特征向量。Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and obtain the The feature vector corresponding to each character in the voice information is verified.

在一种实现方式中,所述将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,从而得到所述验证语音信息中各个字符对应的特征向量包括:In an implementation manner, the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as training sample data, and the maximum a posteriori probability algorithm is used to correspond to the preset corresponding character. The mean supervector of the background model is adjusted, so that the feature vector corresponding to each character in the verification voice information is obtained:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样 本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合预设的超向量子空间矩阵,得到所述验证语音信息中各个字符对应的特征向量。Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as a training sample In the data, the maximum a posteriori probability algorithm is used to adjust the mean supervector of the universal background model corresponding to the preset corresponding character, and combined with the preset supervector subspace matrix, each of the verified voice information is obtained. The feature vector corresponding to the character.

在一种实现方式中,所述将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合所述预设的超向量子空间矩阵,得到所述验证语音信息中各个字符对应的特征向量包括:In an implementation manner, the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to correspond to the preset corresponding character. Adjusting the mean supervector of the universal background model, and combining the preset supervector subspace matrix, obtaining the feature vector corresponding to each character in the verification voice information includes:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述训练样本数据,采用下式对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted The general background model corresponding to the corresponding character has the largest posterior probability:

M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为所述预设的超向量子空间矩阵,ω为所述验证语音信息中相应字符对应的特征向量。M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector The spatial matrix, ω is a feature vector corresponding to the corresponding character in the verification speech information.

在一种实现方式中,所述预设的超向量子空间矩阵为根据所述通用背景模型中各个高斯模块的权重之间的相关性确定得到的。In an implementation manner, the preset super-vector sub-space matrix is determined according to a correlation between weights of respective Gaussian modules in the universal background model.

在一种实现方式中,所述计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数包括:In an implementation manner, the calculating a similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information includes:

计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值确定为所述相似度分数。Calculating a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information, and determining the cosine distance value as the similarity score.

在一种实现方式中,所述对所述验证语音信息进行语音识别得到所述验证语音信息中包含的分别与所述第一字符串中的多个字符对应的语音片段包括:In an implementation manner, the performing voice recognition on the verification voice information to obtain the voice segments respectively included in the verification voice information and corresponding to the plurality of characters in the first character string includes:

识别所述验证语音信息中的有效语音片段和无效语音片段;Identifying valid voice segments and invalid voice segments in the verification voice information;

对所述有效语音片段进行语音识别得到分别与所述第一字符串中的多个字符对应的语音片段。Performing speech recognition on the valid speech segment to obtain a speech segment respectively corresponding to a plurality of characters in the first character string.

在一种实现方式中,所述将所述验证用户确定为所述注册语音信息对应的所述注册用户之前,所述方法还包括:In an implementation manner, before the determining, by the verification user, the registered user corresponding to the registration voice information, the method further includes:

确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序是否一致;以及 Determining whether the ordering of the voice segments of the plurality of characters in the verification voice information is consistent with the ordering of the corresponding characters in the first character string;

在所述相似度达到预设的门限值,并且所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序一致的情况下,将所述验证用户确定为所述注册语音信息对应的所述注册用户。And if the similarity reaches a preset threshold value, and the order of the voice segments of the plurality of characters in the verification voice information is consistent with the ordering of the corresponding characters in the first character string, The verification user determines the registered user corresponding to the registration voice information.

在一种实现方式中,其特征在于,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述方法还包括:In an implementation, the method further includes: before the obtaining the verification voice information generated by the verification user to read the first character string, the method further includes:

随机生成所述第一字符串,并显示所述第一字符串。The first character string is randomly generated, and the first character string is displayed.

相应的,本发明实施例第二方面提供了一种声纹识别装置,所述装置包括:Correspondingly, a second aspect of the embodiments of the present invention provides a voiceprint recognition apparatus, where the apparatus includes:

语音获取模块,用于获取验证用户朗读第一字符串所产生的验证语音信息;a voice acquiring module, configured to obtain verification voice information generated by the user to read the first character string;

相似度判断模块,用于计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度;The similarity judging module is configured to calculate a similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset voice information in the voice information corresponding to each character in the verification voice information;

用户识别模块,用于在所述相似度达到预设的门限值时,将验证用户验证为所述注册语音信息对应的注册用户。The user identification module is configured to verify the verification user as the registered user corresponding to the registered voice information when the similarity reaches a preset threshold.

在一种实现方式中,所述声纹识别装置还包括:In an implementation manner, the voiceprint recognition apparatus further includes:

语音片段识别模块,用于对所述验证语音信息进行语音识别,得到所述验证语音信息中包含的分别与所述第一字符串中的多个字符对应的语音片段;a voice segment identification module, configured to perform voice recognition on the verification voice information, and obtain a voice segment respectively included in the verification voice information corresponding to multiple characters in the first character string;

声纹特征提取模块,用于提取验证语音信息中各个字符对应的语音片段的声纹特征。The voiceprint feature extraction module is configured to extract a voiceprint feature of the voice segment corresponding to each character in the verification voice information.

在一种实现方式中,所述声纹识别装置还包括:In an implementation manner, the voiceprint recognition apparatus further includes:

特征模型训练模块,用于根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量;a feature model training module, configured to perform, according to the voiceprint feature of the voice segment corresponding to each character, and the common background model corresponding to the corresponding corresponding character, to obtain a feature vector corresponding to each character in the verification voice information;

所述相似度判断模块,用于计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数,作为验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度。The similarity judging module is configured to calculate a similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information, as the verification corresponding to each character in the voice information The similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset registered voice information.

在一种实现方式中,所述语音获取模块,还用于获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符;In an implementation manner, the voice acquiring module is further configured to obtain registration voice information generated by a registered user reading a second character string, where the second character string has at least one identical character with the first character string. ;

所述特征模型训练模块,还用于根据所述注册语音信息中各个字符对应的 声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述注册语音信息中各个字符对应的特征向量。The feature model training module is further configured to: according to each character in the registered voice information The voiceprint feature is combined with a common background model corresponding to the preset corresponding character to obtain a feature vector corresponding to each character in the registered voice information.

在一种实现方式中,所述特征模型训练模块用于:In one implementation, the feature model training module is configured to:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,得到所述验证语音信息中各个字符对应的特征向量。Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and obtain the The feature vector corresponding to each character in the voice information is verified.

在一种实现方式中,所述特征模型训练模块用于:In one implementation, the feature model training module is configured to:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合预设的超向量子空间矩阵,得到验证语音信息中各个字符对应的特征向量。The voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character. Combined with the preset super-vector subspace matrix, the feature vector corresponding to each character in the voice information is obtained.

在一种实现方式中,所述特征模型训练模块具体用于:In an implementation manner, the feature model training module is specifically configured to:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用下式对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:The voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted corresponding The general background model corresponding to the character has the largest posterior probability:

M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为所述预设的超向量子空间矩阵,ω为所述验证语音信息中相应字符对应的特征向量。M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector The spatial matrix, ω is a feature vector corresponding to the corresponding character in the verification speech information.

在一种实现方式中,所述预设的超向量子空间矩阵为根据所述高斯混合模型的均值超向量中各个维度向量之间的相关性确定得到的。In an implementation manner, the preset super-vector sub-space matrix is determined according to a correlation between respective dimension vectors in the mean super-vector of the Gaussian mixture model.

在一种实现方式中,所述相似度判断模块用于:In an implementation manner, the similarity determination module is configured to:

计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值确定为所述相似度分数。Calculating a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information, and determining the cosine distance value as the similarity score .

在一种实现方式中,所述语音片段识别模块包括:In an implementation manner, the voice segment identification module includes:

有效片段识别单元,用于识别所述验证语音信息中的有效语音片段和无效语音片段;a valid segment identification unit, configured to identify a valid speech segment and an invalid speech segment in the verification speech information;

语音识别单元,用于对所述有效语音片段进行语音识别得到分别与所述第一字符串中的多个字符对应的语音片段。 The voice recognition unit is configured to perform voice recognition on the valid voice segment to obtain a voice segment corresponding to multiple characters in the first character string.

在一种实现方式中,所述声纹识别装置还包括:In an implementation manner, the voiceprint recognition apparatus further includes:

字符排序确定模块,用于确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序是否一致;a character order determining module, configured to determine whether a sorting of the voice segments of the plurality of characters in the verification voice information is consistent with a ranking of the corresponding characters in the first character string;

所述用户识别模块,还用于在所述相似度达到预设的门限值,并且所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序一致的情况下,将所述验证用户确定为所述注册语音信息对应的所述注册用户。The user identification module is further configured to: when the similarity reaches a preset threshold, and verify the order of the voice segments of the plurality of characters in the voice information and the corresponding characters in the first string In the case that the sorting is consistent, the verification user is determined as the registered user corresponding to the registered voice information.

在一种实现方式中,所述声纹识别装置还包括:In an implementation manner, the voiceprint recognition apparatus further includes:

字符串显示模块,用于随机生成所述第一字符串,并显示所述第一字符串。a string display module, configured to randomly generate the first string, and display the first string.

本发明实施例第三方面还提供了一种声纹识别装置,包括:A third aspect of the embodiments of the present invention further provides a voiceprint recognition apparatus, including:

用户接口,用于获取语音信息;a user interface for obtaining voice information;

存储器,存储计算机可执行程序代码;以及a memory storing computer executable program code;

处理器,用于调用所述计算机可执行程序代码以执行以下操作:a processor for invoking the computer executable program code to perform the following operations:

通过所述用户接口获取验证用户朗读第一字符串所产生的验证语音信息;Acquiring, by the user interface, verification voice information generated by the user to read the first character string;

计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度;Calculating a similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset voice information in the voice information of the verification voice information;

当所述相似度达到预设的门限值时,将验证用户验证为所述注册语音信息对应的注册用户。When the similarity reaches a preset threshold, the user verification is verified as the registered user corresponding to the registered voice information.

在一种实现方式中,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述处理器还调用所述计算机可执行程序代码执行以下操作:In an implementation manner, before the obtaining the verification voice information generated by the verification user to read the first character string, the processor further calls the computer executable program code to perform the following operations:

对所述验证语音信息进行语音识别,得到所述验证语音信息中包含的分别与所述第一字符串中的各个字符对应的语音片段;Performing voice recognition on the verification voice information, and obtaining a voice segment respectively corresponding to each character in the first character string included in the verification voice information;

提取各个字符对应的语音片段的声纹特征。The voiceprint features of the voice segments corresponding to the respective characters are extracted.

在一种实现方式中,所述处理器调用所述计算机可执行程序代码执行以下操作以计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度:In an implementation manner, the processor invokes the computer executable program code to perform the following operations to calculate a sound segment of a voice segment corresponding to a corresponding character in the preset registered voice information in the voice information of the verification voice information. Similarity of the pattern features:

根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量; And performing, according to the voiceprint feature of the voice segment corresponding to each character, in combination with a common background model corresponding to the preset corresponding character, obtaining a feature vector corresponding to each character in the verification voice information;

计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量的相似度分数,作为验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度。Calculating a similarity score of the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, as a voice segment corresponding to each character in the verification voice information and a preset registration The similarity of the voiceprint features of the voice segments corresponding to the corresponding characters in the voice information.

在一种实现方式中,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,之前,所述处理器还调用所述计算机可执行程序代码执行以下操作:In an implementation manner, before the obtaining the verification voice information generated by the verification user to read the first character string, the processor further calls the computer executable program code to perform the following operations:

获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符;Obtaining registration voice information generated by the registered user reading the second character string, the second character string having at least one of the same characters as the first character string;

根据所述注册语音信息中各个字符对应的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述注册语音信息中各个字符对应的特征向量。The feature vector corresponding to each character in the registered voice information is obtained according to the voiceprint feature corresponding to each character in the registered voice information, combined with the common background model corresponding to the preset corresponding character.

在一种实现方式中,所述处理器调用所述计算机可执行程序代码执行以下操作以根据所述各个字符对应的语音片段的声纹特征,结合所述预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量:In one implementation, the processor invokes the computer executable program code to perform the following operations to combine a common background model corresponding to the preset corresponding character according to a voiceprint feature of the voice segment corresponding to the respective character Training, obtaining a feature vector corresponding to each character in the verification voice information:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,得到所述验证语音信息中各个字符对应的特征向量。Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and obtain the The feature vector corresponding to each character in the voice information is verified.

在一种实现方式中,所述处理器调用所述计算机可执行程序代码执行以下操作以将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,从而得到所述验证语音信息中各个字符对应的特征向量:In one implementation, the processor invokes the computer executable program code to perform the following operations to use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as training sample data, using the maximum The probability algorithm adjusts the mean supervector of the universal background model corresponding to the preset corresponding character, thereby obtaining a feature vector corresponding to each character in the verified voice information:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合预设的超向量子空间矩阵,得到所述验证语音信息中各个字符对应的特征向量。The voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character. And combining the preset super vector subspace matrix to obtain a feature vector corresponding to each character in the verification voice information.

在一种实现方式中,所述处理器调用所述计算机可执行程序代码执行以下操作以将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述 训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合所述预设的超向量子空间矩阵,得到所述验证语音信息中各个字符对应的特征向量:In one implementation, the processor invokes the computer executable program code to perform the following operations to use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as Training the sample data, using the maximum posterior probability algorithm to adjust the mean supervector of the universal background model corresponding to the preset corresponding character, and combining the preset supervector subspace matrix to obtain the verification voice The feature vector corresponding to each character in the message:

将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述训练样本数据,采用下式对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted The general background model corresponding to the corresponding character has the largest posterior probability:

M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为所述预设的超向量子空间矩阵,ω为所述验证语音信息中相应字符对应的特征向量。M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector The spatial matrix, ω is a feature vector corresponding to the corresponding character in the verification speech information.

在一种实现方式中,所述预设的超向量子空间矩阵为根据所述通用背景模型中各个高斯模块的权重之间的相关性确定得到的。In an implementation manner, the preset super-vector sub-space matrix is determined according to a correlation between weights of respective Gaussian modules in the universal background model.

在一种实现方式中,所述处理器调用所述计算机可执行程序代码执行以下操作以计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数包括:In an implementation manner, the processor calls the computer executable program code to perform the following operations to calculate a feature vector corresponding to a feature vector corresponding to each character in the verification voice information and a corresponding character in the preset registration voice information. The similarity scores include:

计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值确定为所述相似度分数。Calculating a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information, and determining the cosine distance value as the similarity score.

在一种实现方式中,所述处理器调用所述计算机可执行程序代码执行以下操作以对所述验证语音信息进行语音识别得到所述验证语音信息中包含的分别与所述第一字符串中的多个字符对应的语音片段包括:In an implementation manner, the processor calls the computer executable program code to perform the following operations to perform voice recognition on the verification voice information to obtain the verification voice information included in the first string and the first character string respectively The voice segments corresponding to multiple characters include:

识别所述验证语音信息中的有效语音片段和无效语音片段;Identifying valid voice segments and invalid voice segments in the verification voice information;

对所述有效语音片段进行语音识别得到分别与所述第一字符串中的多个字符对应的语音片段。Performing speech recognition on the valid speech segment to obtain a speech segment respectively corresponding to a plurality of characters in the first character string.

在一种实现方式中,所述将所述验证用户确定为所述注册语音信息对应的所述注册用户之前,所述处理器还调用所述计算机可执行程序代码执行以下操作:In an implementation manner, before the determining the determined user as the registered user corresponding to the registered voice information, the processor further invokes the computer executable program code to perform the following operations:

确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序是否一致;以及Determining whether the ordering of the voice segments of the plurality of characters in the verification voice information is consistent with the ordering of the corresponding characters in the first character string;

在所述相似度达到预设的门限值,并且所述验证语音信息中的多个字符的 语音片段的排序与所述第一字符串中的相应字符的排序一致的情况下,将所述验证用户确定为所述注册语音信息对应的所述注册用户。The similarity reaches a preset threshold, and the plurality of characters in the voice information are verified If the order of the voice segments is consistent with the ordering of the corresponding characters in the first character string, the verification user is determined as the registered user corresponding to the registered voice information.

在一种实现方式中,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述处理器还调用所述计算机可执行程序代码执行以下操作:In an implementation manner, before the obtaining the verification voice information generated by the verification user to read the first character string, the processor further calls the computer executable program code to perform the following operations:

随机生成所述第一字符串,并通过所述用户接口显示所述第一字符串。The first character string is randomly generated, and the first character string is displayed through the user interface.

相应的,本发明实施例第四方面还提供了一种存储介质,所述存储介质中存储有计算机程序,所述计算机程序用以执行如本发明实施例第一方面的任一种实现方式所述的声纹识别方法。Correspondingly, the fourth aspect of the embodiments of the present invention further provides a storage medium, where the computer program stores a computer program, where the computer program is used to perform any implementation manner according to the first aspect of the embodiments of the present invention. The voiceprint recognition method described.

本实施例通过获取验证用户的验证语音信息中各个字符对应的语音片段的声纹特征,结合预设的相应字符的UBM训练得到验证语音信息中各个字符对应的特征向量,并通过将验证语音信息中各个字符对应的特征向量与注册语音信息中相应字符的特征向量进行相似度比较,从而确定验证用户的的用户身份,该方式用以比较的用户特征向量与具体字符对应,充分考虑到用户朗读不同字符时的声纹特征,从而可有效提高声纹识别准确率。In this embodiment, the voiceprint feature of the voice segment corresponding to each character in the verification voice information of the user is obtained, and the UTM training corresponding to the preset corresponding character is used to obtain the feature vector corresponding to each character in the voice message, and the voice information is verified. The feature vector corresponding to each character in the registration is compared with the feature vector of the corresponding character in the registered voice information, thereby determining the user identity of the verified user, and the user feature vector used for comparison corresponds to the specific character, fully considering the user reading aloud The voiceprint features of different characters can effectively improve the accuracy of voiceprint recognition.

附图说明DRAWINGS

为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the embodiments or the description of the prior art will be briefly described below. Obviously, the drawings in the following description are only It is a certain embodiment of the present invention, and other drawings can be obtained from those skilled in the art without any creative work.

图1是本发明实施例中的声纹识别方法的阶段概述示意图;1 is a schematic diagram showing the stages of a voiceprint recognition method in an embodiment of the present invention;

图2是本发明实施例中的一种声纹识别方法的流程示意图;2 is a schematic flow chart of a voiceprint recognition method according to an embodiment of the present invention;

图3是本发明实施例中从语音信息中识别得到多个字符对应的语音片段的原理示意图;3 is a schematic diagram showing the principle of identifying a voice segment corresponding to a plurality of characters from voice information in the embodiment of the present invention;

图4是本发明实施例中从语音信息中获取各个字符对应的特征向量的原理示意图;4 is a schematic diagram of a principle for acquiring feature vectors corresponding to respective characters from voice information according to an embodiment of the present invention;

图5是本发明实施例中注册用户的声纹注册流程示意图;5 is a schematic diagram of a voiceprint registration process of a registered user in an embodiment of the present invention;

图6是本发明另一实施例中的声纹识别方法的流程示意图; 6 is a schematic flow chart of a voiceprint recognition method in another embodiment of the present invention;

图7是本发明实施例中的一种声纹识别装置的结构示意图;FIG. 7 is a schematic structural diagram of a voiceprint recognition apparatus according to an embodiment of the present invention; FIG.

图8是本发明实施例中的语音片段识别模块的结构示意图。FIG. 8 is a schematic structural diagram of a voice segment identification module in an embodiment of the present invention.

图9是本发明实施例中的另一种声纹识别装置的结构示意图。FIG. 9 is a schematic structural diagram of another voiceprint recognition apparatus in an embodiment of the present invention.

具体实施方式detailed description

下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,而不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The technical solutions in the embodiments of the present invention are clearly and completely described in the following with reference to the accompanying drawings in the embodiments of the present invention. It is obvious that the described embodiments are only a part of the embodiments of the present invention, but not all embodiments. All other embodiments obtained by those skilled in the art based on the embodiments of the present invention without creative efforts are within the scope of the present invention.

本发明实施例提供了一种声纹识别方法及装置。该声纹识别方法及装置可应用于所有需要识别未知用户身份的场景或设备中。用于进行声纹识别的字符串中的字符可以是阿拉伯数字、英文字母或其他语言字符等。为了简化描述,本发明实施例中的字符以阿拉伯数字为例进行举例说明。Embodiments of the present invention provide a voiceprint recognition method and apparatus. The voiceprint recognition method and apparatus can be applied to all scenes or devices that need to identify an unknown user. The characters in the string used for voiceprint recognition may be Arabic numerals, English letters or other language characters. In order to simplify the description, the characters in the embodiments of the present invention are exemplified by taking Arabic numerals as an example.

本发明实施例中的声纹识别方法可以分为两个阶段,如图1所示:The voiceprint recognition method in the embodiment of the present invention can be divided into two stages, as shown in FIG. 1:

1)注册用户的声纹注册阶段1) Registered user's voiceprint registration stage

在声纹注册阶段,注册用户可以朗读一个注册字符串(即后文出现的第二字符串),声纹识别装置采集该注册用户在朗读该注册字符串时的注册语音信息,然后对注册语音信息进行声音识别得到所述注册语音信息中包含的分别与所述注册字符串中的多个字符对应的语音片段,进而对各个字符对应的语音片段进行声纹特征提取和声纹模型训练,包括根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型(Universal Background Model,UBM,即GMM-UBM)训练得到注册语音信息中各个字符对应的特征向量,然后声纹识别装置可以分别为不同的注册用户将其在声纹注册阶段朗读的注册语音信息中的多个字符对应的特征向量保存在声纹识别装置的模型库中。In the voiceprint registration phase, the registered user can read a registration string (ie, the second character string appearing later), and the voiceprint recognition device collects the registered voice information when the registered user reads the registration string, and then registers the voice. Performing voice recognition on the information to obtain voice segments respectively corresponding to the plurality of characters in the registration string, and performing voiceprint feature extraction and voiceprint model training on the voice segments corresponding to each character, including And obtaining, according to the voiceprint feature of the voice segment corresponding to each character, a feature vector corresponding to each character in the registered voice information by using a Universal Background Model (UBM, ie, GMM-UBM) corresponding to the preset corresponding character, Then, the voiceprint recognition device can respectively save the feature vectors corresponding to the plurality of characters in the registered voice information that the different registered users read in the voiceprint registration stage in the model library of the voiceprint recognition device.

比如,注册字符串是数字字符串0185851,包含了四种数字“0”、“1”、“5”、“8”,则声纹识别装置根据注册语音信息中各个字符对应的语音片段进行声纹特征提取和声纹模型训练,得到“0”、“1”、“5”、“8”对应的语音片段的声纹特征,进而结合预设的相应字符对应的UBM训练得到注册语音信息中各个字符 对应的特征向量,包括与数字“0”对应的特征向量、与数字“1”对应的特征向量、与数字“5”对应的特征向量以及与数字“8”对应的特征向量。For example, the registration string is a numeric string 0185851, and includes four numbers “0”, “1”, “5”, “8”, and the voiceprint recognition device performs sound according to the voice segment corresponding to each character in the registered voice information. The pattern feature extraction and the voiceprint model training obtain the voiceprint features of the voice segments corresponding to “0”, “1”, “5”, and “8”, and then obtain the registered voice information in combination with the UBM training corresponding to the preset corresponding characters. Individual characters Corresponding feature vectors include a feature vector corresponding to the number “0”, a feature vector corresponding to the number “1”, a feature vector corresponding to the number “5”, and a feature vector corresponding to the number “8”.

2)验证用户的身份识别阶段2) Verify the user's identification phase

在身份识别阶段,验证用户即未知身份的用户朗读一个验证字符串(即后文出现的第一字符串,所述第二字符串与所述第一字符串拥有至少一个相同的字符),声纹识别装置采集该验证用户在朗读该验证字符串时的验证语音信息,然后对验证语音信息进行声音识别得到所述验证语音信息中包含的分别与所述验证字符串中的多个字符对应的语音片段,进而对各个字符对应的语音片段进行声纹特征提取和声纹模型训练,包括根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的UBM训练得到验证语音信息中各个字符对应的特征向量,最后计算验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数,若所述相似度分数达到预设验证门限,则将所述验证用户确定为所述注册语音信息对应的注册用户。In the identification phase, the user who authenticates the user, that is, the unknown identity, reads a verification string (ie, the first character string that appears later, and the second character string has at least one of the same characters as the first character string). The pattern recognition device collects the verification voice information when the verification user reads the verification character string, and then performs voice recognition on the verification voice information to obtain corresponding to the plurality of characters in the verification string respectively included in the verification voice information. The voice segment, and then the voiceprint feature extraction and the voiceprint model training for the voice segment corresponding to each character, including the voiceprint feature corresponding to the voice segment corresponding to the respective character, and the UBM training corresponding to the corresponding corresponding character is used to obtain the verification voice. The feature vector corresponding to each character in the information, and finally calculating the similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information, if the similarity score reaches the preset verification a threshold, the verification user is determined as a note corresponding to the registration voice information Users.

比如,验证字符串为数字字符串85851510,则声纹识别装置根据验证用户朗读时产生的验证语音信息中各个字符对应的语音片段进行声纹特征提取和声纹模型训练,得到“0”、“1”、“5”、“8”对应的GMM,进而结合预设的相应字符对应的UBM可以计算得到验证用户的验证语音信息的特征向量,包括与数字“0”对应的特征向量、与数字“1”对应的特征向量、与数字“5”对应的特征向量以及与数字“8”对应的特征向量,进而分别计算验证语音信息中的“0”、“1”、“5”、“8”对应的特征向量分别与注册语音信息中的“0”、“1”、“5”、“8”对应的特征向量之间的相似度分数,若所述相似度分数达到预设验证门限,则将所述验证用户确定为所述注册语音信息对应的注册用户。For example, if the verification string is a numeric string 85851510, the voiceprint recognition device performs voiceprint feature extraction and voiceprint model training according to the voice segment corresponding to each character in the verification voice information generated when the user reads aloud, and obtains “0” and “ The GMM corresponding to 1", "5", and "8", and then combined with the preset UBM corresponding to the corresponding character, can calculate the feature vector of the verified voice information of the verified user, including the feature vector corresponding to the number "0", and the number a feature vector corresponding to "1", a feature vector corresponding to the number "5", and a feature vector corresponding to the number "8", thereby respectively calculating "0", "1", "5", "8" in the verification voice information a similarity score between the corresponding feature vectors and the feature vectors corresponding to "0", "1", "5", and "8" in the registered voice information, and if the similarity score reaches a preset verification threshold, And determining, by the verification user, a registered user corresponding to the registered voice information.

需要指出的是,上述注册用户的声纹注册阶段和验证用户的身份识别阶段可以在同一设备或装置中实现,也可以分别在不同的设备或装置中实现,例如注册用户的声纹注册阶段在第一设备中实施,进而第一设备将注册语音信息中的多个字符对应的特征向量发送给第二设备,从而可以在第二设备中实施验证用户的身份识别阶段。 It should be noted that the voiceprint registration phase of the registered user and the identity phase of the authenticated user may be implemented in the same device or device, or may be implemented in different devices or devices, for example, the voiceprint registration phase of the registered user is The first device is implemented, and the first device sends the feature vector corresponding to the multiple characters in the registered voice information to the second device, so that the identity recognition phase of the authenticated user can be implemented in the second device.

下面通过具体实施例分别对上述两个过程进行详细阐述。The above two processes are respectively explained in detail through specific embodiments.

图2是本发明实施例中的一种声纹识别方法的流程示意图,如图所示本实施例中的声纹识别方法流程可以包括:2 is a schematic flow chart of a method for identifying a voiceprint according to an embodiment of the present invention. As shown in the figure, the flow of the voiceprint recognition method in the embodiment may include:

S201,获取验证用户朗读第一字符串所产生的验证语音信息。S201. Acquire verification voice information generated by verifying that the user reads the first character string.

所述验证用户即未知身份的用户,需要通过声纹识别装置验证其用户身份。所述第一字符串是用于验证用户进行身份验证的字符串,可以是随机生成的,也可以是预设固定的一个字符串,例如与预先生成的注册语音信息对应的第二字符串至少部分相同的一个字符串。具体的,所述第一字符串可以包含m个字符,其中有n个互不相同的字符,m,n均为正整数,且m≥n。The user who authenticates the user, that is, the unknown identity, needs to verify the identity of the user through the voiceprint recognition device. The first character string is a character string used for verifying the user's identity verification, and may be randomly generated, or may be a preset fixed string, for example, a second string corresponding to the pre-generated registration voice information. Part of the same string. Specifically, the first character string may include m characters, wherein n characters are different from each other, m and n are positive integers, and m≥n.

比如,第一字符串是“12358948”,共8个字符,包括了7种互不相同的字符“1”、“2”、“3”、“4”、“5”、“8”、“9”。For example, the first string is "12358948", a total of 8 characters, including 7 different characters "1", "2", "3", "4", "5", "8", " 9".

在可选实施例中,声纹识别装置可以生成并显示所述第一字符串,让验证用户根据显示的所述第一字符串进行朗读。In an alternative embodiment, the voiceprint recognition device may generate and display the first character string for the verification user to read aloud according to the displayed first character string.

S202,对所述验证语音信息进行语音识别,以得到所述验证语音信息中包含的分别与所述第一字符串中的多个字符对应的语音片段。S202. Perform speech recognition on the verification voice information to obtain a voice segment respectively included in the verification voice information and corresponding to multiple characters in the first character string.

如图3所示,声纹识别装置可以通过语音识别以及声音强度过滤,将所述验证语音信息划分得到多个字符对应的语音片段,可选的还可以将无效语音片段剔除掉,不参与后续的处理过程。As shown in FIG. 3, the voiceprint recognition device can divide the verification voice information into voice segments corresponding to multiple characters through voice recognition and voice intensity filtering, and optionally remove the invalid voice segments, and do not participate in subsequent Process.

S203,提取各个字符对应的语音片段的声纹特征。S203. Extract a voiceprint feature of the voice segment corresponding to each character.

具体的,声纹识别装置可以提取各个字符对应的语音片段中的MFCC(Mel Frequency Cepstrum Coefficient,梅尔倒谱系数)或PLP(Perceptual Linear Predictive,感知线性预测系数),作为各个字符所对应的语音片段的声纹特征。Specifically, the voiceprint recognition device may extract MFCC (Mel Frequency Cepstrum Coefficient) or PLP (Perceptual Linear Predictive Coefficient) in the voice segment corresponding to each character as the voice corresponding to each character. The voiceprint features of the clip.

S204,根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到验证语音信息中各个字符对应的特征向量;S204, according to the voiceprint feature of the voice segment corresponding to each character, combined with the common background model corresponding to the preset corresponding character, to obtain a feature vector corresponding to each character in the verification voice information;

本发明实施例中的所述通用背景模型UBM,是一种将大量说话人的特定数字的语音段混合训练而成的混合高斯模型,表征对应数字的语音在特征空间中的分布,又由于训练数据来源于大量的说话人,因此它不表征某一类具体的说话人,具有身份无关的特性,可看作是一种通用背景模型。示意性的,可以采用说话人数大于1000人、时长超过20小时的语音样本,并且各个字符的出 现频率相对均衡,训练得到UBM。UBM的数学表达式为:The universal background model UBM in the embodiment of the present invention is a mixed Gaussian model in which a plurality of speech segments of a plurality of speakers are mixed and trained, and the distribution of the corresponding digital speech in the feature space is represented, and The data comes from a large number of speakers, so it does not represent a specific type of speaker, and has an identity-independent nature, which can be regarded as a general background model. Schematically, a speech sample with more than 1000 speakers and a duration of more than 20 hours can be used, and each character is out. The current frequency is relatively balanced and training is obtained by UBM. The mathematical expression of UBM is:

P(x)=∑t=1...CaiN(x|μi,∑i)…………式(1)P(x)=∑ t=1...C a i N(x|μ i ,∑ i )............(1)

其中,P(x)代表UBM的概率分布,C代表UBM中共有C个高斯模块进行加和,ai代表第i个高斯模块的权重,μi代表第i个高斯模块的均值,∑i代表第i个高斯模块的方差,N(x)代表高斯分布,x代表输入的样本的声纹特征。Where P(x) represents the probability distribution of UBM, C represents the total of C Gaussian modules in UBM, a i represents the weight of the i-th Gaussian module, μ i represents the mean of the i-th Gaussian module, ∑ i represents The variance of the i-th Gaussian module, N(x) represents a Gaussian distribution, and x represents the voiceprint characteristics of the input sample.

声纹识别装置可以将验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法(Maximum A Posteriori,MAP)对预设的相应字符对应的通用背景模型的参数进行调整,即在将验证语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整预设的相应字符对应的通用背景模型的参数,使得后验概率P(x)最大,从而可以根据使得后验概率P(x)最大的参数确定验证语音信息中相应字符对应的特征向量。The voiceprint recognition device may use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, and adopt a maximum a posteriori probability algorithm (MAP) to the common background model corresponding to the preset corresponding character. The parameter is adjusted, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the posterior check is made by continuously adjusting the parameters of the common background model corresponding to the corresponding corresponding character. The probability P(x) is the largest, so that the feature vector corresponding to the corresponding character in the verification speech information can be determined according to the parameter that makes the posterior probability P(x) the largest.

由于大量的实验和论文验证了UBM模型中每个高斯模块的均值可以用于区分说话人的身份信息,定义UBM模型的均值超向量为:Since a large number of experiments and papers have verified that the mean of each Gaussian module in the UBM model can be used to distinguish the identity information of the speaker, the mean supervector of the UBM model is defined as:

Figure PCTCN2017087911-appb-000001
Figure PCTCN2017087911-appb-000001

从而,声纹识别装置可以将验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法(Maximum A Posteriori,MAP)对预设的相应字符对应的通用背景模型的均值超向量进行调整,即在将验证语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整均值超向量,使得后验概率P(x)最大,从而可以将使得后验概率P(x)最大的均值超向量作为验证语音信息中相应字符对应的特征向量。Therefore, the voiceprint recognition device can use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, and adopt the maximum posterior probability algorithm (MAP) to the common background corresponding to the preset corresponding character. The mean value supervector of the model is adjusted, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the posterior probability P(x) is adjusted by continuously adjusting the mean supervector. The maximum, so that the mean supervector that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the verification speech information.

在另一可选实施例中,为了降低超向量的高维度带来的收敛速度慢的问题, 通过基于概率的主成分分析方法(PPCA,probabilistic principal component analysis)将均值超向量的变化范围限制在一个子空间中,声纹识别装置可以将验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法对预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合预设的超向量子空间矩阵从而得到验证语音信息中各个字符对应的特征向量。具体实现中,可以采用下式对预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:In another alternative embodiment, in order to reduce the slow convergence speed caused by the high dimension of the super vector, The profiling of the mean supervector is limited to a subspace by probabilistic principal component analysis (PPCA), and the voiceprint recognition device can verify the voiceprint features of the speech segments corresponding to each character in the speech information. As the training sample data, the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the corresponding corresponding character, and the preset super-vector subspace matrix is used to obtain the corresponding features of each character in the verified voice information. vector. In a specific implementation, the mean supervector of the common background model corresponding to the preset corresponding character may be adjusted by using the following formula, so that the posterior probability of the universal background model corresponding to the adjusted corresponding character is the largest:

M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为预设的超向量子空间矩阵,ω即为验证语音信息中相应字符对应的特征向量,即在将验证语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整ω可以实现调整式(1)中的均值超向量,使得后验概率P(x)最大,从而可以将使得后验概率P(x)最大的ω作为验证语音信息中相应字符对应的特征向量。所述超向量子空间矩阵T为根据所述高斯混合模型的均值超向量中各个维度向量之间的相关性确定得到的。M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector subspace matrix , ω is the feature vector corresponding to the corresponding character in the verification voice information, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the adjustment can be realized by continuously adjusting ω. The mean supervector in (1) maximizes the posterior probability P(x), so that ω that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the verification speech information. The super-vector subspace matrix T is determined according to the correlation between each dimension vector in the mean supervector of the Gaussian mixture model.

S205,计算验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数,若所述相似度分数达到预设验证门限,则将所述验证用户确定为所述注册语音信息对应的注册用户。S205. Calculate a similarity score of the feature vector corresponding to the corresponding character in the preset registered voice information in the verification voice information, and if the similarity score reaches the preset verification threshold, the verification user is The registered user corresponding to the registered voice information is determined.

具体的,声纹识别装置可以在声纹注册阶段获取到注册用户的注册语音信息,并通过与本实施例相类似的声纹特征提取和声纹模型训练,可以得到注册语音信息中各个字符的语音片段对应的特征向量。所述注册语音信息,可以是声纹识别装置获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符,即所述注册语音信息对应的第二字符串与所述第一字符串至少部分相同。进而在可选实施例中,声纹识别装置还可以从外部获取所述注册语音信息中相应字符对应的特征向量,即注册用户通过其他设备录入了注册语音信息后,其他设备或者服务器通过声纹特征提取和声纹模型训练得到注册语音信息中各个字符的语音片段对应的特征向量,声纹识别装置通过从其他设备或者服务器获取到所述注册语音信息中相 应字符对应的特征向量,从而在验证用户的身份识别阶段用以与验证语音信息中各个字符对应的特征向量进行比较。Specifically, the voiceprint recognition device can obtain the registered voice information of the registered user in the voiceprint registration stage, and obtain the voiceprint feature extraction and the voiceprint model training similar to the embodiment, so that the characters in the registered voice information can be obtained. The feature vector corresponding to the speech segment. The registration voice information may be that the voiceprint recognition device acquires the registration voice information generated by the registered user reading the second character string, and the second character string has at least one of the same characters as the first character string, that is, the The second character string corresponding to the registration voice information is at least partially identical to the first character string. In an alternative embodiment, the voiceprint recognition device may further acquire the feature vector corresponding to the corresponding character in the registered voice information from the outside, that is, after the registered user inputs the registered voice information through other devices, the other device or the server passes the voiceprint. Feature extraction and voiceprint model training obtain feature vectors corresponding to voice segments of each character in the registered voice information, and the voiceprint recognition device acquires the phase in the registered voice information by using other devices or servers The feature vector corresponding to the character is used to compare the feature vector corresponding to each character in the verification voice information in the identification phase of the verification user.

具体实现中,所述相似度分数是声纹识别装置将验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量进行比较后,衡量相同字符的两个特征向量之间的相似程度的分值。在可选实施例中,可以计算验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值作为所述相似度分数,即通过下式计算某个字符分别在验证语音信息中对应的特征向量和注册语音信息中的特征向量之间的相似度分数:In a specific implementation, the similarity score is that the voiceprint recognition device compares the feature vector corresponding to each character in the voice information with the feature vector corresponding to the corresponding character in the preset registered voice information, and then measures two features of the same character. The degree of similarity between vectors. In an optional embodiment, a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information may be calculated, and the cosine distance value is used as the The similarity score, that is, the similarity score between the corresponding feature vector in the verification voice information and the feature vector in the registered voice information is calculated by the following formula:

Figure PCTCN2017087911-appb-000002
Figure PCTCN2017087911-appb-000002

其中,下标i表示第i个验证语音信息和注册语音信息中共有的字符,ωi(tar)表示该字符在验证语音信息中对应的特征向量,ωi(test)表示该字符在注册语音信息中对应的特征向量。若验证语音信息和注册语音信息中包含多个相同的字符,则可以根据上式计算得到的各个字符的相似度分数取均值,若各个字符的相似度分数均值达到对应的预设验证门限,则将所述验证用户确定为所述注册语音信息对应的注册用户。若存在多位注册用户,例如图1所示的注册用户A、B以及C,可以根据验证用户某个字符的特征向量与各个注册用户的相应字符的特征向量的相似度,当某个注册用户的相应字符的特征向量与验证语音的该字符的特征向量的相似度分数最高且相似度达到预设验证门限,则将该注册用户作为验证用户的身份识别结果。Wherein, the subscript i indicates a character common to the i-th verification voice information and the registration voice information, ω i (tar) indicates a corresponding feature vector of the character in the verification voice information, and ω i (test) indicates that the character is in the registered voice. The corresponding feature vector in the message. If the verification voice information and the registration voice information include multiple identical characters, the similarity scores of the characters calculated according to the above formula may be averaged. If the similarity scores of the respective characters reach the corresponding preset verification threshold, then The verification user is determined to be a registered user corresponding to the registration voice information. If there are multiple registered users, such as the registered users A, B, and C shown in FIG. 1, the registered user may be authenticated according to the similarity between the feature vector of a certain character of the user and the corresponding character of each registered user. The feature vector of the corresponding character and the feature vector of the character of the verification voice have the highest similarity score and the similarity reaches the preset verification threshold, and the registered user is used as the identification result of the verification user.

在可选实施例中,若所述验证语音信息中存在同一字符出现不止一次,例如出现如图2所示的验证语音信息中0、1、5以及8分别都出现了2次,那么可以按照两次字符0对应的语音片段处理得到的特征向量分别与预设的注册语音信息中字符0的特征向量的相似度分数的平均值,作为本次验证语音信息中字符0的特征向量与预设的注册语音信息中字符0的特征向量的相似度分数,以此类推。In an optional embodiment, if the same character appears in the verification voice information more than once, for example, if 0, 1, 5, and 8 appear in the verification voice information as shown in FIG. 2, respectively, then The average of the similarity scores of the feature vectors processed by the two-character 0 corresponding to the speech segment and the feature vector of the character 0 in the preset registration speech information is used as the feature vector and preset of the character 0 in the verification speech information. The similarity score of the feature vector of character 0 in the registered voice information, and so on.

需要指出的是,衡量两个特征向量之间的相似度的方式还有很多,以上仅是本发明提供的一种实施方式,本领域技术人员在本发明公开的方案的基础上 可以无需创造性劳动地获得更多的计算验证语音信息和注册语音信息中共有的字符的特征向量的相似度分数的方式,本发明无需穷举。It should be noted that there are many ways to measure the similarity between two feature vectors. The above is only one embodiment provided by the present invention, and those skilled in the art based on the solution disclosed by the present invention. It is possible to obtain more ways of calculating the similarity scores of the feature vectors of the characters shared in the voice information and the registered voice information without creative labor, and the present invention need not be exhaustive.

从而,本实施例通过获取验证用户的验证语音信息中各个字符对应的语音片段的声纹特征,结合预设的相应字符的UBM训练得到验证语音信息中各个字符对应的特征向量,并通过将验证语音信息中各个字符对应的特征向量与注册语音信息中相应字符的特征向量进行相似度比较,从而确定验证用户的用户身份,该方式用以比较的用户特征向量与具体字符对应,充分考虑到用户朗读不同字符时的声纹特征,从而可有效提高声纹识别准确率。Therefore, in this embodiment, by obtaining the voiceprint feature of the voice segment corresponding to each character in the verification voice information of the verification user, the UTM training corresponding to the preset corresponding character is used to obtain the feature vector corresponding to each character in the verification voice information, and the verification is performed. The feature vector corresponding to each character in the voice information is compared with the feature vector of the corresponding character in the registered voice information, thereby determining the user identity of the verified user. The user feature vector used for comparison corresponds to the specific character, and the user is fully considered. The voiceprint features of different characters are read aloud, so that the accuracy of voiceprint recognition can be effectively improved.

图5是本发明实施例中注册用户的声纹注册流程示意图,如图所示本实施例中的声纹注册流程可以包括:5 is a schematic diagram of a voiceprint registration process of a registered user in the embodiment of the present invention. As shown in the figure, the voiceprint registration process in this embodiment may include:

S501,获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符。S501. Acquire a registration voice information generated by a registered user reading a second character string, where the second character string has at least one identical character with the first character string.

所述注册用户即确定合法身份的用户,所述第二字符串是用于采集注册用户声纹特征向量的字符串,可以是随机生成的,也可以是预设固定的一个字符串。具体的,所述第二字符串也可以包含m个字符,其中有n个互不相同的字符,m,n均为正整数,且m≥n。The registered user is a user who determines a legal identity. The second character string is a string used to collect the voiceprint feature vector of the registered user, and may be randomly generated or a fixed string. Specifically, the second character string may also include m characters, wherein there are n characters different from each other, m and n are positive integers, and m≥n.

在可选实施例中,声纹识别装置可以生成并显示所述第二字符串,让注册用户根据显示的所述第二字符串进行朗读。In an alternative embodiment, the voiceprint recognition device may generate and display the second character string for the registered user to read aloud according to the displayed second character string.

S502,对所述注册语音信息进行语音识别得到所述注册语音信息中包含的分别与所述第二字符串中的多个字符对应的语音片段;S502. Perform speech recognition on the registration voice information to obtain a voice segment respectively included in the registration voice information and corresponding to multiple characters in the second character string.

声纹识别装置可以通过语音识别以及声音强度过滤,将所述验证语音信息划分得到多个字符对应的语音片段,可选的还可以将无效语音片段剔除掉,不参与后续的处理过程。The voiceprint recognition device can divide the verification voice information into voice segments corresponding to multiple characters through voice recognition and voice intensity filtering, and optionally remove the invalid voice segments, and do not participate in subsequent processing.

S503,提取注册语音信息中各个字符对应的语音片段的声纹特征。S503. Extract a voiceprint feature of the voice segment corresponding to each character in the registered voice information.

具体的,声纹识别装置可以提取各个字符对应的语音片段中的MFCC(Mel Frequency Cepstrum Coefficient,梅尔倒谱系数)或PLP(Perceptual Linear Predictive,感知线性预测系数),作为各个字符所对应的语音片段的声纹特征。Specifically, the voiceprint recognition device may extract MFCC (Mel Frequency Cepstrum Coefficient) or PLP (Perceptual Linear Predictive Coefficient) in the voice segment corresponding to each character as the voice corresponding to each character. The voiceprint features of the clip.

S504,根据注册语音信息中各个字符对应的语音片段的声纹特征,结合预 设的相应字符对应的通用背景模型训练得到注册语音信息中各个字符对应的特征向量。S504, according to the voiceprint feature of the voice segment corresponding to each character in the registered voice information, combined with the pre- The general background model corresponding to the corresponding character is trained to obtain the feature vector corresponding to each character in the registered voice information.

UBM的表达式可以参考前文实施例。声纹注册流程的该步骤与声纹识别流程的S204类似,声纹识别装置可以将注册语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法(Maximum A Posteriori,MAP)对预设的相应字符对应的通用背景模型的参数进行调整,即在将注册语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整预设的相应字符对应的通用背景模型的参数,使得后验概率P(x)最大,从而可以根据使得后验概率P(x)最大的参数确定注册语音信息中相应字符对应的特征向量。The expression of UBM can refer to the previous embodiment. This step of the voiceprint registration process is similar to S204 of the voiceprint recognition process. The voiceprint recognition device can use the voiceprint feature of the voice segment corresponding to each character in the registered voice information as the training sample data, and adopt the maximum posterior probability algorithm (Maximum A). Posteriori, MAP) adjusts the parameters of the common background model corresponding to the corresponding corresponding characters, that is, after the voiceprint feature of the voice segment corresponding to each character in the registered voice information is substituted into the formula (1) as an input sample, The parameters of the common background model corresponding to the preset corresponding characters are such that the posterior probability P(x) is the largest, so that the feature vector corresponding to the corresponding character in the registered voice information can be determined according to the parameter that makes the posterior probability P(x) the largest.

而由于UBM模型中每个高斯模块的均值可以用于区分说话人的身份信息,声纹识别装置可以将注册语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法(Maximum A Posteriori,MAP)对预设的相应字符对应的通用背景模型的均值超向量进行调整,即在将注册语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整均值超向量,使得后验概率P(x)最大,从而可以将使得后验概率P(x)最大的均值超向量作为注册语音信息中相应字符对应的特征向量。Since the mean value of each Gaussian module in the UBM model can be used to distinguish the identity information of the speaker, the voiceprint recognition device can use the voiceprint feature of the voice segment corresponding to each character in the registered voice information as the training sample data, using the maximum posterior The Maximum A Posteriori (MAP) adjusts the mean supervector of the universal background model corresponding to the preset corresponding characters, that is, the voiceprint feature of the voice segment corresponding to each character in the registered voice information is used as an input sample substitution type ( 1) After, by continuously adjusting the mean supervector, the posterior probability P(x) is maximized, so that the mean supervector that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the registered speech information.

在另一可选实施例中,可以采用下式对预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:In another optional embodiment, the mean supervector of the universal background model corresponding to the preset corresponding character may be adjusted by using the following formula, so that the posterior probability of the universal background model corresponding to the adjusted corresponding character is the largest:

M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为预设的超向量子空间矩阵,ω即为注册语音信息中相应字符对应的特征向量,即在将注册语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整ω可以实现调整式(1)中的均值超向量,使得后验概率P(x)最大,从而可以将使得后验概率P(x)最大的ω作为注册语音信息中相应字符对应的特征向量。M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector subspace matrix , ω is the feature vector corresponding to the corresponding character in the registered voice information, that is, after the voiceprint feature of the voice segment corresponding to each character in the registered voice information is substituted into the formula (1) as an input sample, the adjustment can be realized by continuously adjusting ω. The mean supervector in (1) maximizes the posterior probability P(x), so that ω that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the registered speech information.

图6是本发明另一实施例中的声纹识别方法的流程示意图,如图所示本实 施例中的声纹识别方法可以包括以下流程:6 is a schematic flow chart of a voiceprint recognition method according to another embodiment of the present invention, as shown in the figure. The voiceprint recognition method in the embodiment may include the following processes:

S601,随机生成第一字符串并进行显示。S601. Randomly generate a first character string and display it.

S602,获取验证用户朗读第一字符串所产生的验证语音信息。S602. Acquire verification voice information generated by verifying that the user reads the first character string.

S603,识别所述验证语音信息中的有效语音片段和无效语音片段。S603. Identify valid voice segments and invalid voice segments in the verification voice information.

具体的,可以根据声音强度对验证语音进行划分,将声音强度较小的语音片段视为无效语音片段(例如包括静音段和脉冲噪声)。Specifically, the verification voice may be divided according to the sound intensity, and the voice segment with less sound intensity is regarded as an invalid voice segment (for example, including a silent segment and impulse noise).

S604,对所述有效语音片段进行语音识别得到分别与所述第一字符串中的多个字符对应的语音片段。S604. Perform speech recognition on the valid speech segment to obtain a speech segment respectively corresponding to the plurality of characters in the first character string.

可以通过语音识别,得到分别与所述第一字符串中的多个字符对应的语音片段。A voice segment respectively corresponding to a plurality of characters in the first character string can be obtained by voice recognition.

S605,确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序一致。S605. Determine that the order of the voice segments of the plurality of characters in the verification voice information is consistent with the order of the corresponding characters in the first character string.

为了有效避免注册用户的语音信息被盗录或非法拷贝后用以进行声纹识别,可以每次随机生成不同的第一字符串,并在本步骤判断验证语音信息中的多个字符的语音片段的排序是否与第一字符串中的相应字符的排序一致,若不一致,则可以判断声纹识别失败,若与第一字符串中的相应字符的排序一致,则执行后续流程。In order to effectively prevent the voice information of the registered user from being pirated or illegally copied for voiceprint recognition, a different first character string may be randomly generated each time, and in this step, the voice segment of the plurality of characters in the voice information is determined. Whether the sorting is consistent with the sorting of the corresponding characters in the first string, if not, it can be judged that the voiceprint recognition fails, and if the sorting of the corresponding characters in the first string is consistent, the subsequent process is performed.

S606,提取各个字符对应的语音片段的声纹特征。S606. Extract voiceprint features of the voice segments corresponding to the respective characters.

具体的,声纹识别装置可以提取各个字符对应的语音片段中的MFCC(Mel Frequency Cepstrum Coefficient,梅尔倒谱系数)或PLP(Perceptual Linear Predictive,感知线性预测系数),作为各个字符所对应的语音片段的声纹特征。Specifically, the voiceprint recognition device may extract MFCC (Mel Frequency Cepstrum Coefficient) or PLP (Perceptual Linear Predictive Coefficient) in the voice segment corresponding to each character as the voice corresponding to each character. The voiceprint features of the clip.

S607,将验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法对预设的相应字符对应的通用背景模型的均值超向量进行调整,从而估计得到验证语音信息中各个字符对应的特征向量。S607, the voiceprint feature of the voice segment corresponding to each character in the voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the corresponding corresponding character, so that the estimation is verified. A feature vector corresponding to each character in the voice message.

由于大量的实验和论文验证了UBM模型中每个高斯模块的均值可以用于区分说话人的身份信息,声纹识别装置可以将验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法(Maximum A Posteriori,MAP)对预设的相应字符对应的通用背景模型的均值超向量进行调整,即在将验证语音信息中各个字符对应的语音片段的声纹特征作为输入样本 代入式(1)后,通过不断调整均值超向量,使得后验概率P(x)最大,从而可以将使得后验概率P(x)最大的均值超向量作为验证语音信息中相应字符对应的特征向量。Since a large number of experiments and papers have verified that the mean of each Gaussian module in the UBM model can be used to distinguish the identity information of the speaker, the voiceprint recognition device can verify the voiceprint feature of the voice segment corresponding to each character in the voice information as a training sample. Data, using Maximum A Posteriori (MAP) to adjust the mean supervector of the common background model corresponding to the corresponding corresponding character, that is, the voiceprint feature of the voice segment corresponding to each character in the verification voice information As an input sample After substituting into equation (1), by continuously adjusting the mean supervector, the posterior probability P(x) is maximized, so that the mean supervector with the largest posterior probability P(x) can be used as the corresponding feature of the corresponding character in the verification speech information. vector.

在另一可选实施例中,为了降低超向量的高维度带来的收敛速度慢的问题,声纹识别装置可以采用下式对预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:In another optional embodiment, in order to reduce the slow convergence speed caused by the high dimension of the super vector, the voiceprint recognition apparatus may adjust the mean super vector of the common background model corresponding to the preset corresponding character by using the following formula: The posterior probability of the universal background model corresponding to the adjusted corresponding characters is the largest:

M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为预设的超向量子空间矩阵,ω即为验证语音信息中相应字符对应的特征向量,即在将验证语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整ω可以实现调整式(1)中的均值超向量,使得后验概率P(x)最大,从而可以将使得后验概率P(x)最大的ω作为验证语音信息中相应字符对应的特征向量。M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector subspace matrix , ω is the feature vector corresponding to the corresponding character in the verification voice information, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the adjustment can be realized by continuously adjusting ω. The mean supervector in (1) maximizes the posterior probability P(x), so that ω that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the verification speech information.

S608,计算验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数,若相似度分数达到预设验证门限,则将验证用户确定为注册语音信息对应的注册用户。S608. Calculate a similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information. If the similarity score reaches a preset verification threshold, the verification user is determined to be the registered voice. The registered user corresponding to the information.

本实施例中,声纹识别装置可以计算验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值作为所述相似度分数,即通过下式计算某个字符分别在验证语音信息中对应的特征向量和注册语音信息中的特征向量之间的相似度分数:In this embodiment, the voiceprint recognition apparatus may calculate a cosine distance value between the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, and use the cosine distance value as The similarity score, that is, the similarity score between the corresponding feature vector in the verification voice information and the feature vector in the registered voice information is calculated by the following formula:

Figure PCTCN2017087911-appb-000003
Figure PCTCN2017087911-appb-000003

其中,下标i表示第i个验证语音信息和注册语音信息中共有的字符,ωi(tar)表示该字符在验证语音信息中对应的特征向量,ωi(test)表示该字符在注册语音信息中对应的特征向量。若验证语音信息和注册语音信息中包含多个相同的字符,则可以根据上式计算得到的各个字符的相似度分数取均值,若各个字符的相似度分数均值达到对应的预设验证门限,则将所述验证用户确定为所述注册语音信息对应的注册用户。若存在多位注册用户,例如图1所示的注册用户A、B以及C,可以根据验证用户某个字符的特征向量与各个注册用户的 相应字符的特征向量的相似度,当某个注册用户的相应字符的特征向量与验证语音的该字符的特征向量的相似度分数最高且相似度达到预设验证门限,则将该注册用户作为验证用户的身份识别结果。Wherein, the subscript i indicates a character common to the i-th verification voice information and the registration voice information, ω i (tar) indicates a corresponding feature vector of the character in the verification voice information, and ω i (test) indicates that the character is in the registered voice. The corresponding feature vector in the message. If the verification voice information and the registration voice information include multiple identical characters, the similarity scores of the characters calculated according to the above formula may be averaged. If the similarity scores of the respective characters reach the corresponding preset verification threshold, then The verification user is determined to be a registered user corresponding to the registration voice information. If there are multiple registered users, such as the registered users A, B, and C shown in FIG. 1, the registered user may be authenticated according to the similarity between the feature vector of a certain character of the user and the corresponding character of each registered user. The feature vector of the corresponding character and the feature vector of the character of the verification voice have the highest similarity score and the similarity reaches the preset verification threshold, and the registered user is used as the identification result of the verification user.

从而,本实施例通过将验证语音信息中各个字符对应的特征向量与注册语音信息中相应字符的特征向量进行相似度比较,并且结合了语音片段的时序判断,可以进一步确保验证用户的用户身份的准确性。Therefore, in this embodiment, by verifying the similarity between the feature vector corresponding to each character in the verification voice information and the feature vector of the corresponding character in the registered voice information, and combining the timing judgment of the voice segment, the user identity of the user can be further verified. accuracy.

图7是本发明实施例中的一种声纹识别装置的结构示意图,如图所示本实施例中的声纹识别装置可以包括:FIG. 7 is a schematic structural diagram of a voiceprint identifying apparatus according to an embodiment of the present invention. As shown in the figure, the voiceprint identifying apparatus in this embodiment may include:

语音获取模块710,用于获取验证用户朗读第一字符串所产生的验证语音信息。The voice obtaining module 710 is configured to obtain verification voice information generated by the user to read the first character string.

所述验证用户即未知身份的用户,需要通过声纹识别装置验证其用户身份。所述第一字符串是用于验证用户进行身份验证的字符串,可以是随机生成的,也可以是预设固定的一个字符串,例如与预先生成的注册语音信息对应的第二字符串至少部分相同的一个字符串。具体的,所述第一字符串可以包含m个字符,其中有n个互不相同的字符,m,n均为正整数,且m≥n。The user who authenticates the user, that is, the unknown identity, needs to verify the identity of the user through the voiceprint recognition device. The first character string is a character string used for verifying the user's identity verification, and may be randomly generated, or may be a preset fixed string, for example, a second string corresponding to the pre-generated registration voice information. Part of the same string. Specifically, the first character string may include m characters, wherein n characters are different from each other, m and n are positive integers, and m≥n.

比如,第一字符串是“12358948”,共8个字符,包括了7种互不相同的字符“1”、“2”、“3”、“4”、“5”、“8”、“9”。For example, the first string is "12358948", a total of 8 characters, including 7 different characters "1", "2", "3", "4", "5", "8", " 9".

语音片段识别模块720,用于对所述验证语音信息进行语音识别得到所述验证语音信息中包含的分别与所述第一字符串中的多个字符对应的语音片段。The voice segment identification module 720 is configured to perform voice recognition on the verification voice information to obtain voice segments respectively corresponding to the plurality of characters in the first character string included in the verification voice information.

如图3所示,语音片段识别模块720可以通过语音识别以及声音强度过滤,将所述验证语音信息划分得到多个字符对应的语音片段,可选的还可以将无效语音片段剔除掉,不参与后续的处理过程。As shown in FIG. 3, the voice segment identification module 720 can divide the verification voice information into voice segments corresponding to multiple characters through voice recognition and voice intensity filtering, and optionally remove the invalid voice segments without participating. Subsequent processing.

在可选实施例中,所述语音片段识别模块如图8所示进一步可以包括:In an optional embodiment, the voice segment identification module may further include:

有效片段识别单元721,用于识别所述验证语音信息中的有效语音片段和无效语音片段。The valid segment identification unit 721 is configured to identify a valid speech segment and an invalid speech segment in the verification speech information.

具体的,有效片段识别单元721可以根据声音强度对验证语音进行划分,将声音强度较小的语音片段视为无效语音片段(例如包括静音段和脉冲噪声)。Specifically, the effective segment recognizing unit 721 can divide the verification speech according to the sound intensity, and treat the speech segment with a small sound intensity as an invalid speech segment (for example, including a silent segment and impulse noise).

语音识别单元722,用于对所述有效语音片段进行语音识别得到分别与所 述第一字符串中的多个字符对应的语音片段。a voice recognition unit 722, configured to perform voice recognition on the valid voice segment to obtain a separate voice A speech segment corresponding to a plurality of characters in the first character string.

声纹特征提取模块730,用于提取验证语音信息中各个字符对应的语音片段的声纹特征。The voiceprint feature extraction module 730 is configured to extract a voiceprint feature of the voice segment corresponding to each character in the verification voice information.

具体的,声纹特征提取模块730可以提取各个字符对应的语音片段中的MFCC(Mel Frequency Cepstrum Coefficient,梅尔倒谱系数)或PLP(Perceptual Linear Predictive,感知线性预测系数),作为各个字符所对应的语音片段的声纹特征。Specifically, the voiceprint feature extraction module 730 can extract MFCC (Mel Frequency Cepstrum Coefficient) or PLP (Perceptual Linear Predictive Coefficient) in the voice segment corresponding to each character, as corresponding to each character. The voiceprint features of the speech segment.

特征模型训练模块740,用于根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练得到验证语音信息中各个字符对应的特征向量。The feature model training module 740 is configured to train, according to the voiceprint feature of the voice segment corresponding to the respective characters, a feature vector corresponding to each character in the verification voice information according to the common background model corresponding to the preset corresponding character.

特征模型训练模块740可以将验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法(Maximum A Posteriori,MAP)对预设的相应字符对应的通用背景模型的参数进行调整,即在将验证语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整预设的相应字符对应的通用背景模型的参数,使得后验概率P(x)最大,从而特征模型训练模块740可以根据使得后验概率P(x)最大的参数确定验证语音信息中相应字符对应的特征向量。The feature model training module 740 may use the voiceprint feature of the voice segment corresponding to each character in the voice information as the training sample data, and adopt a maximum posterior probability algorithm (MAP) to the common background model corresponding to the preset corresponding character. The parameters are adjusted, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the parameters of the common background model corresponding to the corresponding corresponding characters are continuously adjusted, so that The probability P(x) is the largest, so that the feature model training module 740 can determine the feature vector corresponding to the corresponding character in the verification voice information according to the parameter that maximizes the posterior probability P(x).

由于大量的实验和论文验证了UBM模型中每个高斯模块的均值可以用于区分说话人的身份信息,我们定义UBM模型的均值超向量为:Since a large number of experiments and papers have verified that the mean of each Gaussian module in the UBM model can be used to distinguish the identity information of the speaker, we define the mean supervector of the UBM model as:

Figure PCTCN2017087911-appb-000004
Figure PCTCN2017087911-appb-000004

从而,特征模型训练模块740可以将验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法(Maximum A Posteriori,MAP)对预设的相应字符对应的通用背景模型的均值超向量进行调 整,即在将验证语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整均值超向量,使得后验概率P(x)最大,特征模型训练模块740可以将使得后验概率P(x)最大的均值超向量作为验证语音信息中相应字符对应的特征向量。Therefore, the feature model training module 740 can use the voiceprint feature of the voice segment corresponding to each character in the voice information as the training sample data, and use the Maximum A Posteriori (MAP) algorithm to correspond to the preset corresponding character. The mean supervector of the background model is adjusted In the whole, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1), the posterior probability P(x) is maximized by continuously adjusting the mean supervector, and the feature model training module 740 may use the mean supervector that maximizes the posterior probability P(x) as the feature vector corresponding to the corresponding character in the verification speech information.

在另一可选实施例中,为了降低超向量的高维度带来的收敛速度慢的问题,我们通过基于概率的主成分分析方法(PPCA,probabilistic principal component analysis)将均值超向量的变化范围限制在一个子空间中,特征模型训练模块740可以将验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法对预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合预设的超向量子空间矩阵从而得到验证语音信息中各个字符对应的特征向量。具体实现中,特征模型训练模块740可以采用下式对预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:In another alternative embodiment, in order to reduce the slow convergence rate caused by the high dimension of the supervector, we limit the variation range of the mean supervector by probabilistic principal component analysis (PPCA). In a subspace, the feature model training module 740 may use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, and adopt the maximum posterior probability algorithm to the common background model corresponding to the preset corresponding character. The mean super vector is adjusted, and the preset super vector subspace matrix is combined to obtain a feature vector corresponding to each character in the verified speech information. In a specific implementation, the feature model training module 740 may adjust the mean super vector of the common background model corresponding to the preset corresponding character by using the following formula, so that the posterior probability of the common background model corresponding to the adjusted corresponding character is the largest:

M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为预设的超向量子空间矩阵,ω即为验证语音信息中相应字符对应的特征向量,即在将验证语音信息中各个字符对应的语音片段的声纹特征作为输入样本代入式(1)后,通过不断调整ω可以实现调整式(1)中的均值超向量,使得后验概率P(x)最大,从而可以将使得后验概率P(x)最大的ω作为验证语音信息中相应字符对应的特征向量。所述超向量子空间矩阵T为根据所述高斯混合模型的均值超向量中各个维度向量之间的相关性确定得到的。M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector subspace matrix , ω is the feature vector corresponding to the corresponding character in the verification voice information, that is, after the voiceprint feature of the voice segment corresponding to each character in the verification voice information is substituted into the formula (1) as an input sample, the adjustment can be realized by continuously adjusting ω. The mean supervector in (1) maximizes the posterior probability P(x), so that ω that maximizes the posterior probability P(x) can be used as the feature vector corresponding to the corresponding character in the verification speech information. The super-vector subspace matrix T is determined according to the correlation between each dimension vector in the mean supervector of the Gaussian mixture model.

相似度判断模块750,用于计算验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数。The similarity determining module 750 is configured to calculate a similarity score of the feature vector corresponding to the corresponding character in the preset registered voice information in the verification voice information.

具体的,声纹识别装置在可以在声纹注册阶段获取到注册用户的注册语音信息,并通过语音片段识别模块720、声纹特征提取模块730以及特征模型训练模块740,可以得到注册语音信息中各个字符的语音片段对应的特征向量。所述注册语音信息,可以是声纹识别装置获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符,即所述注册语音信息对应的第二字符串与所述第一字符串至少部分相同。进而 在可选实施例中,声纹识别装置还可以从外部获取所述注册语音信息中相应字符对应的特征向量,即注册用户通过其他设备录入了注册语音信息后,其他设备或者服务器通过声纹特征提取和声纹模型训练得到注册语音信息中各个字符的语音片段对应的特征向量,声纹识别装置通过从其他设备或者服务器获取到所述注册语音信息中相应字符对应的特征向量,从而在验证用户的身份识别阶段相似度判断模块750用以与验证语音信息中各个字符对应的特征向量进行比较。Specifically, the voiceprint recognition device can obtain the registered voice information of the registered user in the voiceprint registration stage, and can obtain the registered voice information through the voice segment identification module 720, the voiceprint feature extraction module 730, and the feature model training module 740. The feature vector corresponding to the speech segment of each character. The registration voice information may be that the voiceprint recognition device acquires the registration voice information generated by the registered user reading the second character string, and the second character string has at least one of the same characters as the first character string, that is, the The second character string corresponding to the registration voice information is at least partially identical to the first character string. and then In an optional embodiment, the voiceprint recognition device may further acquire a feature vector corresponding to the corresponding character in the registered voice information, that is, after the registered user inputs the registered voice information through other devices, the other device or the server passes the voiceprint feature. The extraction and voiceprint model training obtains a feature vector corresponding to the voice segment of each character in the registered voice information, and the voiceprint recognition device obtains the feature vector corresponding to the corresponding character in the registered voice information from other devices or servers, thereby verifying the user The identification stage similarity judgment module 750 is configured to compare the feature vectors corresponding to the respective characters in the verification voice information.

具体实现中,所述相似度分数是声纹识别装置将验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量进行比较后,衡量相同字符的两个特征向量之间的相似程度的分值。在可选实施例中,相似度判断模块750可以计算验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值作为所述相似度分数,即通过下式计算某个字符分别在验证语音信息中对应的特征向量和注册语音信息中的特征向量之间的相似度分数:In a specific implementation, the similarity score is that the voiceprint recognition device compares the feature vector corresponding to each character in the voice information with the feature vector corresponding to the corresponding character in the preset registered voice information, and then measures two features of the same character. The degree of similarity between vectors. In an optional embodiment, the similarity determination module 750 may calculate a cosine distance value between the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, and the cosine distance value is The distance value is used as the similarity score, that is, the similarity score between the corresponding feature vector in the verification voice information and the feature vector in the registered voice information is calculated by the following formula:

Figure PCTCN2017087911-appb-000005
Figure PCTCN2017087911-appb-000005

其中,下标i表示第i个验证语音信息和注册语音信息中共有的字符,ωi(tar)表示该字符在验证语音信息中对应的特征向量,ωi(test)表示该字符在注册语音信息中对应的特征向量。在可选实施例中,若所述验证语音信息中存在同一字符出现不止一次,例如出现如图2所示的验证语音信息中0、1、5以及8分别都出现了2次,那么可以将两次字符0对应的语音片段处理得到的特征向量分别与预设的注册语音信息中字符0的特征向量的相似度分数的平均值,作为本次验证语音信息中字符0的特征向量与预设的注册语音信息中字符0的特征向量的相似度分数,以此类推。Wherein, the subscript i indicates a character common to the i-th verification voice information and the registration voice information, ω i (tar) indicates a corresponding feature vector of the character in the verification voice information, and ω i (test) indicates that the character is in the registered voice. The corresponding feature vector in the message. In an optional embodiment, if the same character appears in the verification voice information more than once, for example, if 0, 1, 5, and 8 appear in the verification voice information as shown in FIG. 2, respectively, then The average of the similarity scores of the feature vectors processed by the two-character 0 corresponding to the speech segment and the feature vector of the character 0 in the preset registration speech information is used as the feature vector and preset of the character 0 in the verification speech information. The similarity score of the feature vector of character 0 in the registered voice information, and so on.

需要指出的是,衡量两个特征向量之间的相似度的方式还有很多,以上仅是本发明提供的一种实施方式,本领域技术人员在本发明公开的方案的基础上可以无需创造性劳动地获得更多的计算验证语音信息和注册语音信息中共有的字符的特征向量的相似度分数的方式,本发明无需穷举。It should be noted that there are many ways to measure the similarity between two feature vectors. The above is only one embodiment provided by the present invention, and those skilled in the art can perform no creative work on the basis of the solution disclosed by the present invention. In order to obtain more ways of calculating the similarity scores of the feature vectors of the characters shared in the verification voice information and the registration voice information, the present invention need not be exhaustive.

用户识别模块760,用于若所述相似度分数达到预设验证门限,则将所述 验证用户确定为所述注册语音信息对应的注册用户。a user identification module 760, configured to: if the similarity score reaches a preset verification threshold, The verification user determines the registered user corresponding to the registered voice information.

若验证语音信息和注册语音信息中包含多个相同的字符,则用户识别模块760可以根据相似度判断模块750计算得到的各个字符的相似度分数取均值,若各个字符的相似度分数均值达到对应的预设验证门限,则将所述验证用户确定为所述注册语音信息对应的注册用户。若存在多位注册用户,例如图1所示的注册用户A、B以及C,用户识别模块760可以根据验证用户某个字符的特征向量与各个注册用户的相应字符的特征向量的相似度,当某个注册用户的相应字符的特征向量与验证语音的该字符的特征向量的相似度分数最高且相似度达到预设验证门限,则将该注册用户作为验证用户的身份识别结果。If the verification voice information and the registration voice information include a plurality of identical characters, the user identification module 760 may take the average value of the similarity scores of the respective characters calculated by the similarity determination module 750, if the average value of the similarity scores of the respective characters reaches the corresponding value. The preset verification threshold determines the authenticated user as the registered user corresponding to the registered voice information. If there are multiple registered users, such as the registered users A, B, and C shown in FIG. 1, the user identification module 760 may perform the similarity between the feature vector of a certain character of the user and the feature vector of the corresponding character of each registered user. If the feature vector of the corresponding character of a registered user and the feature vector of the character of the verification voice have the highest similarity score and the similarity reaches the preset verification threshold, the registered user is used as the identification result of the verification user.

进而在可选实施例中,所述语音获取模块710,还用于获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符;In an optional embodiment, the voice acquiring module 710 is further configured to obtain registration voice information generated by the registered user reading the second character string, where the second character string has at least one of the same as the first character string. character of;

所述语音片段识别模块720,还用于对所述注册语音信息进行语音识别得到所述注册语音信息中包含的分别与所述第二字符串中的多个字符对应的语音片段;The voice segment identification module 720 is further configured to perform voice recognition on the registration voice information to obtain a voice segment respectively included in the registration voice information and corresponding to multiple characters in the second character string;

所述声纹特征提取模块730,还用于提取注册语音信息中各个字符对应的语音片段的声纹特征;The voiceprint feature extraction module 730 is further configured to extract a voiceprint feature of a voice segment corresponding to each character in the registered voice information;

所述特征模型训练模块740,还用于根据所述注册语音信息中各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练得到注册语音信息中各个字符对应的特征向量。The feature model training module 740 is further configured to: according to the voiceprint feature of the voice segment corresponding to each character in the registered voice information, and the common background model corresponding to the preset corresponding character, the corresponding characters in the registered voice information are trained. Feature vector.

在可选实施例中,声纹识别装置进一步还可以包括:In an alternative embodiment, the voiceprint recognition apparatus may further include:

字符排序确定模块770,用于确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序是否一致。The character order determining module 770 is configured to determine whether the order of the voice segments of the plurality of characters in the verification voice information is consistent with the order of the corresponding characters in the first character string.

为了有效避免注册用户的语音信息被盗录或非法拷贝后用以进行声纹识别,可以每次随机生成不同的第一字符串,并在本步骤判断验证语音信息中的多个字符的语音片段的排序是否与第一字符串中的相应字符的排序一致,若不一致,则可以判断声纹识别失败,若与第一字符串中的相应字符的排序一致,则可以通知声纹特征提取模块730或特征模型训练模块740执行针对该验证语音信息的特征提取和声纹训练。 In order to effectively prevent the voice information of the registered user from being pirated or illegally copied for voiceprint recognition, a different first character string may be randomly generated each time, and in this step, the voice segment of the plurality of characters in the voice information is determined. Whether the sorting is consistent with the sorting of the corresponding characters in the first string, if not, it may be determined that the voiceprint recognition fails, and if the sorting of the corresponding characters in the first character string is consistent, the voiceprint feature extraction module 730 may be notified. Or feature model training module 740 performs feature extraction and voiceprint training for the verification voice information.

在可选实施例中,声纹识别装置进一步还可以包括:In an alternative embodiment, the voiceprint recognition apparatus may further include:

字符串显示模块700,用于随机生成所述第一字符串并进行显示。The string display module 700 is configured to randomly generate the first string and display it.

从而,本实施例通过获取验证用户的验证语音信息中各个字符对应的语音片段的声纹特征,结合预设的相应字符的UBM训练得到验证语音信息中各个字符对应的特征向量,并通过将验证语音信息中各个字符对应的特征向量与注册语音信息中相应字符的特征向量进行相似度比较,从而确定验证用户的用户身份,该方式用以比较的用户特征向量与具体字符对应,充分考虑到用户朗读不同字符时的声纹特征,从而可有效提高声纹识别准确率。Therefore, in this embodiment, by obtaining the voiceprint feature of the voice segment corresponding to each character in the verification voice information of the verification user, the UTM training corresponding to the preset corresponding character is used to obtain the feature vector corresponding to each character in the verification voice information, and the verification is performed. The feature vector corresponding to each character in the voice information is compared with the feature vector of the corresponding character in the registered voice information, thereby determining the user identity of the verified user. The user feature vector used for comparison corresponds to the specific character, and the user is fully considered. The voiceprint features of different characters are read aloud, so that the accuracy of voiceprint recognition can be effectively improved.

在实际测试实例中,在1000人训练样本,29万次测试中(其中身份匹配的测试在1万次左右,不匹配测试约在28万次),能够实现千分之一错误率下79.8%的召回率,等错概率(EER,Equal Error Rate)为3.39%,相较于传统的文本无关建模方法,声纹识别性能提升超过40%以上。In the actual test case, in the 1000-person training sample, 290,000 tests (in which the identity matching test is about 10,000 times, the mismatch test is about 280,000 times), it is able to achieve 79.8% of the one-thousandth error rate. The recall rate, EER (Equal Error Rate) is 3.39%. Compared with the traditional text-independent modeling method, the voiceprint recognition performance is improved by more than 40%.

图9为本发明实施例提供的另一种声纹识别装置的结构示意图。如图9所示,所述声纹识别装置1000可以包括:至少一个处理器1001(例如CPU)、用户接口1003、存储器1005以及至少一个通信总线1002。其中,通信总线1002用于实现这些组件之间的连接通信。其中,用户接口1003可以包括显示屏(Display)、键盘(Keyboard)、麦克风等,可选地用户接口1003还可以包括标准的有线接口、无线接口。存储器1005可以是高速RAM存储器,也可以是非不稳定的存储器(non-volatile memory),例如至少一个磁盘存储器。如图9所示,作为一种计算机存储介质的存储器1005中可以包括操作系统、用户接口模块以及计算机可执行程序代码(例如声纹识别程序)。FIG. 9 is a schematic structural diagram of another voiceprint recognition apparatus according to an embodiment of the present invention. As shown in FIG. 9, the voiceprint recognition apparatus 1000 may include at least one processor 1001 (eg, a CPU), a user interface 1003, a memory 1005, and at least one communication bus 1002. Among them, the communication bus 1002 is used to implement connection communication between these components. The user interface 1003 may include a display, a keyboard, a microphone, and the like. Optionally, the user interface 1003 may further include a standard wired interface and a wireless interface. The memory 1005 may be a high speed RAM memory or a non-volatile memory such as at least one disk memory. As shown in FIG. 9, an operating system, a user interface module, and computer executable program code (eg, a voiceprint recognition program) may be included in the memory 1005 as a computer storage medium.

在图9所示的声纹识别装置1000中,处理器1001可以用于调用存储器1005中存储的计算机可执行程序代码,并具体执行以下步骤:In the voiceprint recognition apparatus 1000 shown in FIG. 9, the processor 1001 can be used to call the computer executable program code stored in the memory 1005, and specifically perform the following steps:

通过所述用户接口获取验证用户朗读第一字符串所产生的验证语音信息;Acquiring, by the user interface, verification voice information generated by the user to read the first character string;

计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度;Calculating a similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset voice information in the voice information of the verification voice information;

当所述相似度达到预设的门限值时,将验证用户验证为所述注册语音信息对应的注册用户。When the similarity reaches a preset threshold, the user verification is verified as the registered user corresponding to the registered voice information.

在一种实现方式中,所述获取所述验证用户朗读所述第一字符串所产生的 所述验证语音信息之前,所述处理器1001还调用所述计算机可执行程序代码执行以下操作:In an implementation manner, the obtaining, by the verification user, reading the first character string Before the verifying the voice information, the processor 1001 also calls the computer executable program code to perform the following operations:

对所述验证语音信息进行语音识别,得到所述验证语音信息中包含的分别与所述第一字符串中的各个字符对应的语音片段;Performing voice recognition on the verification voice information, and obtaining a voice segment respectively corresponding to each character in the first character string included in the verification voice information;

提取各个字符对应的语音片段的声纹特征。The voiceprint features of the voice segments corresponding to the respective characters are extracted.

在一种实现方式中,所述处理器1001调用所述计算机可执行程序代码执行以下操作以计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度:In an implementation manner, the processor 1001 invokes the computer executable program code to perform the following operations to calculate a voice segment corresponding to a corresponding one of the preset voice information in the voice message corresponding to each character in the verification voice information. Similarity of voiceprint features:

根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量;And performing, according to the voiceprint feature of the voice segment corresponding to each character, in combination with a common background model corresponding to the preset corresponding character, obtaining a feature vector corresponding to each character in the verification voice information;

计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量的相似度分数,作为验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度。Calculating a similarity score of the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, as a voice segment corresponding to each character in the verification voice information and a preset registration The similarity of the voiceprint features of the voice segments corresponding to the corresponding characters in the voice information.

在一种实现方式中,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述处理器1001还调用所述计算机可执行程序代码执行以下操作:获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符;对所述注册语音信息进行语音识别,得到所述注册语音信息中包含的分别与所述第二字符串中的多个字符对应的语音片段;提取所述注册语音信息中各个字符对应的语音片段的声纹特征;根据所述注册语音信息中各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述注册语音信息中各个字符对应的特征向量。In an implementation manner, before the obtaining the verification voice information generated by the verification user to read the first character string, the processor 1001 further invokes the computer executable program code to perform the following operations: acquiring Registering a user to read the registered voice information generated by the second character string, the second character string having at least one of the same characters as the first character string; performing voice recognition on the registered voice information to obtain the registered voice information a voice segment respectively corresponding to the plurality of characters in the second character string; extracting a voiceprint feature of the voice segment corresponding to each character in the registered voice information; and corresponding to each character in the registered voice information The voiceprint feature of the voice segment is trained in conjunction with a common background model corresponding to the corresponding corresponding character to obtain a feature vector corresponding to each character in the registered voice information.

在一种实现方式中,所述处理器1001调用所述计算机可执行程序代码执行以下操作以根据所述各个字符对应的语音片段的声纹特征,结合所述预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量:将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,得到所述验证语音信息中各个字符对应的特征向量。 In an implementation manner, the processor 1001 invokes the computer executable program code to perform the following operations, according to the voiceprint feature of the voice segment corresponding to the respective characters, and the common background corresponding to the preset corresponding character Model training, obtaining a feature vector corresponding to each character in the verification voice information: using a voiceprint feature of a voice segment corresponding to each character in the verification voice information as training sample data, and using a maximum posterior probability algorithm on the preset The mean supervector of the universal background model corresponding to the corresponding character is adjusted to obtain a feature vector corresponding to each character in the verified voice information.

在一种实现方式中,所述处理器1001调用所述计算机可执行程序代码执行以下操作以将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,从而得到所述验证语音信息中各个字符对应的特征向量:将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合预设的超向量子空间矩阵,得到所述验证语音信息中各个字符对应的特征向量。In one implementation, the processor 1001 invokes the computer executable program code to perform the following operations to use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as training sample data, using the maximum The posterior probability algorithm adjusts the mean supervector of the universal background model corresponding to the preset corresponding character, thereby obtaining a feature vector corresponding to each character in the verification voice information: corresponding to each character in the verification voice information The voiceprint feature of the voice segment is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and combined with the preset supervector subspace matrix. Obtaining a feature vector corresponding to each character in the verification voice information.

在一种实现方式中,所述处理器1001调用所述计算机可执行程序代码执行以下操作以将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合所述预设的超向量子空间矩阵,得到所述验证语音信息中各个字符对应的特征向量:将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述训练样本数据,采用下式对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为所述预设的超向量子空间矩阵,ω为所述验证语音信息中相应字符对应的特征向量。In an implementation manner, the processor 1001 invokes the computer executable program code to perform the following operations to use the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, The maximum posterior probability algorithm adjusts the mean supervector of the universal background model corresponding to the preset corresponding character, and combines the preset supervector subspace matrix to obtain corresponding characters of the verified voice information. Feature vector: using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, and adjusting the mean super vector of the common background model corresponding to the preset corresponding character by using the following formula: The posterior probability of the universal background model corresponding to the adjusted corresponding character is maximized: M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, and m represents the generality of the corresponding character before the adjustment. Mean supervector of the background model, T is the preset supervector subspace matrix, and ω is the verification speech information Eigenvectors corresponding to the respective characters.

在一种实现方式中,所述预设的超向量子空间矩阵为根据所述通用背景模型中各个高斯模块的权重之间的相关性确定得到的。In an implementation manner, the preset super-vector sub-space matrix is determined according to a correlation between weights of respective Gaussian modules in the universal background model.

在一种实现方式中,所述处理器1001调用所述计算机可执行程序代码执行以下操作以计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量的所述相似度分数:计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值确定为所述相似度分数。In an implementation manner, the processor 1001 invokes the computer executable program code to perform the following operations to calculate that a feature vector corresponding to each character in the verification voice information corresponds to a corresponding character in the preset registration voice information. The similarity score of the feature vector: calculating a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information, and The cosine distance value is determined as the similarity score.

在一种实现方式中,所述处理器1001调用所述计算机可执行程序代码执行以下操作以对所述验证语音信息进行语音识别得到所述验证语音信息中包 含的分别与所述第一字符串中的多个字符对应的语音片段:识别所述验证语音信息中的有效语音片段和无效语音片段;对所述有效语音片段进行语音识别得到分别与所述第一字符串中的多个字符对应的语音片段。In an implementation manner, the processor 1001 invokes the computer executable program code to perform the following operations to perform voice recognition on the verification voice information to obtain the verification voice information package. a speech segment respectively corresponding to the plurality of characters in the first character string: identifying a valid speech segment and an invalid speech segment in the verification speech information; performing speech recognition on the valid speech segment respectively A speech segment corresponding to a plurality of characters in the first character string.

在一种实现方式中,所述将所述验证用户确定为所述注册语音信息对应的所述注册用户之前,所述处理器1001还调用所述计算机可执行程序代码执行以下操作:确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序是否一致;以及在所述相似度达到预设的门限值,并且所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序一致的情况下,将所述验证用户确定为所述注册语音信息对应的所述注册用户。In an implementation manner, before the verifying the user is determined as the registered user corresponding to the registered voice information, the processor 1001 further invokes the computer executable program code to perform the following operations: determining the Verifying whether the ordering of the voice segments of the plurality of characters in the voice information is consistent with the ordering of the corresponding characters in the first character string; and when the similarity reaches a preset threshold value, and the verification voice information is In the case that the ordering of the speech segments of the plurality of characters is consistent with the ordering of the corresponding characters in the first character string, the verification user is determined as the registered user corresponding to the registration voice information.

在一种实现方式中,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述处理器1001还调用所述计算机可执行程序代码执行以下操作:随机生成所述第一字符串,并显示所述第一字符串。In an implementation manner, before the obtaining the verification voice information generated by the verification user to read the first character string, the processor 1001 further invokes the computer executable program code to perform the following operations: random Generating the first character string and displaying the first character string.

本领域普通技术人员可以理解实现上述实施例方法中的全部或部分流程,是可以通过计算机程序来指令相关的硬件来完成,所述的程序可存储于一计算机可读取存储介质中,该程序在执行时,可包括如上述各方法的实施例的流程。其中,所述的存储介质可为磁碟、光盘、只读存储记忆体(Read-Only Memory,ROM)或随机存储记忆体(Random Access Memory,RAM)等。One of ordinary skill in the art can understand that all or part of the process of implementing the foregoing embodiments can be completed by a computer program to instruct related hardware, and the program can be stored in a computer readable storage medium. When executed, the flow of an embodiment of the methods as described above may be included. The storage medium may be a magnetic disk, an optical disk, a read-only memory (ROM), or a random access memory (RAM).

以上所揭露的仅为本发明较佳实施例而已,当然不能以此来限定本发明之权利范围,因此依本发明权利要求所作的等同变化,仍属本发明所涵盖的范围。 The above is only the preferred embodiment of the present invention, and the scope of the present invention is not limited thereto, and thus equivalent changes made in the claims of the present invention are still within the scope of the present invention.

Claims (27)

一种声纹识别方法,其特征在于,所述方法包括:A voiceprint recognition method, characterized in that the method comprises: 获取验证用户朗读第一字符串所产生的验证语音信息;Obtaining verification voice information generated by the verification user reading the first character string; 计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度;Calculating a similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset voice information in the voice information of the verification voice information; 当所述相似度达到预设的门限值时,将验证用户验证为所述注册语音信息对应的注册用户。When the similarity reaches a preset threshold, the user verification is verified as the registered user corresponding to the registered voice information. 如权利要求1所述的声纹识别方法,其特征在于,所述计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度之前还包括:The voiceprint recognition method according to claim 1, wherein the calculating and verifying the similarity of the voiceprint features of the voice segments corresponding to the respective characters in the preset registered voice information in the voice information is verified. Previously included: 对所述验证语音信息进行语音识别,得到所述验证语音信息中包含的分别与所述第一字符串中的各个字符对应的语音片段;Performing voice recognition on the verification voice information, and obtaining a voice segment respectively corresponding to each character in the first character string included in the verification voice information; 提取各个字符对应的语音片段的声纹特征。The voiceprint features of the voice segments corresponding to the respective characters are extracted. 如权利要求1所述的声纹识别方法,其特征在于,所述计算验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度包括:The voiceprint recognition method according to claim 1, wherein the calculating and verifying the similarity of the voiceprint features of the voice segments corresponding to the respective characters in the preset registered voice information in the voice information is verified. include: 根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量;And performing, according to the voiceprint feature of the voice segment corresponding to each character, in combination with a common background model corresponding to the preset corresponding character, obtaining a feature vector corresponding to each character in the verification voice information; 计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量的相似度分数,作为验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度。Calculating a similarity score of the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, as a voice segment corresponding to each character in the verification voice information and a preset registration The similarity of the voiceprint features of the voice segments corresponding to the corresponding characters in the voice information. 如权利要求3所述的声纹识别方法,其特征在于,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述方法还包括:The voiceprint recognition method according to claim 3, wherein the method further comprises: before the obtaining the verification voice information generated by the verification user reading the first character string, the method further comprising: 获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与 所述第一字符串拥有至少一个相同的字符;Obtaining registration voice information generated by the registered user reading the second character string, the second string and The first string has at least one of the same characters; 根据所述注册语音信息中各个字符对应的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述注册语音信息中各个字符对应的特征向量。The feature vector corresponding to each character in the registered voice information is obtained according to the voiceprint feature corresponding to each character in the registered voice information, combined with the common background model corresponding to the preset corresponding character. 如权利要求3所述的声纹识别方法,其特征在于,所述根据所述各个字符对应的语音片段的声纹特征,结合所述预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量包括:The voiceprint recognition method according to claim 3, wherein the voiceprint feature according to the voice segment corresponding to each character is combined with the common background model corresponding to the preset corresponding character to obtain the Verifying the feature vectors corresponding to each character in the voice information includes: 将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,得到所述验证语音信息中各个字符对应的特征向量。Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character, and obtain the The feature vector corresponding to each character in the voice information is verified. 如权利要求5所述的声纹识别方法,其特征在于,所述将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,从而得到所述验证语音信息中各个字符对应的特征向量包括:The voiceprint recognition method according to claim 5, wherein the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as training sample data, and the maximum posterior probability algorithm is used Adjusting the mean supervector of the common background model corresponding to the preset corresponding character, so as to obtain the feature vector corresponding to each character in the verification voice information, including: 将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合预设的超向量子空间矩阵,得到所述验证语音信息中各个字符对应的特征向量。The voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the maximum a posteriori probability algorithm is used to adjust the mean supervector of the common background model corresponding to the preset corresponding character. And combining the preset super vector subspace matrix to obtain a feature vector corresponding to each character in the verification voice information. 如权利要求6所述的声纹识别方法,其特征在于,所述将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述训练样本数据,采用所述最大后验概率算法对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,并结合所述预设的超向量子空间矩阵,得到所述验证语音信息中各个字符对应的特征向量包括:The voiceprint recognition method according to claim 6, wherein the voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the maximum posterior probability algorithm is used. Adjusting the mean supervector of the common background model corresponding to the preset corresponding character, and combining the preset supervector subspace matrix, obtaining the feature vector corresponding to each character in the verification voice information includes: 将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述训练样本数据,采用下式对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted The general background model corresponding to the corresponding character has the largest posterior probability: M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量, m代表调整前的相应字符的通用背景模型的均值超向量,T为所述预设的超向量子空间矩阵,ω为所述验证语音信息中相应字符对应的特征向量。M=m+Tω, where M represents the mean supervector of the general background model of the adjusted character. m represents the mean supervector of the universal background model of the corresponding character before the adjustment, T is the preset supervector subspace matrix, and ω is the feature vector corresponding to the corresponding character in the verification speech information. 如权利要求6所述的声纹识别方法,其特征在于,所述预设的超向量子空间矩阵为根据所述通用背景模型中各个高斯模块的权重之间的相关性确定得到的。The voiceprint recognition method according to claim 6, wherein the preset super-vector sub-space matrix is determined according to a correlation between weights of respective Gaussian modules in the universal background model. 如权利要求5所述的声纹识别方法,其特征在于,所述通用背景模型的数学表达式为:The voiceprint recognition method according to claim 5, wherein the mathematical expression of the universal background model is: P(x)=∑t=1...CaiN(x|μi,∑i)P(x)=∑ t=1...C a i N(x|μ i ,∑ i ) 其中,P(x)代表UBM的概率分布,C代表UBM中共有C个高斯模块进行加和,ai代表第i个高斯模块的权重,μi代表第i个高斯模块的均值,∑i代表第i个高斯模块的方差,N(x)代表高斯分布,x代表输入的样本的声纹特征。Where P(x) represents the probability distribution of UBM, C represents the total of C Gaussian modules in UBM, a i represents the weight of the i-th Gaussian module, μ i represents the mean of the i-th Gaussian module, ∑ i represents The variance of the i-th Gaussian module, N(x) represents a Gaussian distribution, and x represents the voiceprint characteristics of the input sample. 如权利要求3所述的声纹识别方法,其特征在于,所述计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数包括:The voiceprint recognition method according to claim 3, wherein the calculating a similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information includes: : 计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值确定为所述相似度分数。Calculating a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information, and determining the cosine distance value as the similarity score. 如权利要求10所述的声纹识别方法,其特征在于,通过下式计算验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值:The voiceprint recognition method according to claim 10, wherein the cosine distance value between the feature vector corresponding to each character in the voice information and the feature vector corresponding to the corresponding character in the preset registered voice information is calculated by the following formula :
Figure PCTCN2017087911-appb-100001
Figure PCTCN2017087911-appb-100001
其中,下标i表示第i个验证语音信息和注册语音信息中共有的字符,ωi(tar)表示该字符在验证语音信息中对应的特征向量,ωi(test)表示该字符在注册语音信息中对应的特征向量。 Wherein, the subscript i indicates a character common to the i-th verification voice information and the registration voice information, ω i (tar) indicates a corresponding feature vector of the character in the verification voice information, and ω i (test) indicates that the character is in the registered voice. The corresponding feature vector in the message.
如权利要求2所述的声纹识别方法,其特征在于,所述对所述验证语音信息进行语音识别得到所述验证语音信息中包含的分别与所述第一字符串中的多个字符对应的语音片段包括:The voiceprint recognition method according to claim 2, wherein the voice recognition of the verification voice information is performed, and the verification voice information is respectively included corresponding to a plurality of characters in the first character string. The voice clips include: 识别所述验证语音信息中的有效语音片段和无效语音片段;Identifying valid voice segments and invalid voice segments in the verification voice information; 对所述有效语音片段进行语音识别得到分别与所述第一字符串中的多个字符对应的语音片段。Performing speech recognition on the valid speech segment to obtain a speech segment respectively corresponding to a plurality of characters in the first character string. 如权利要求1所述的声纹识别方法,其特征在于,所述将所述验证用户确定为所述注册语音信息对应的所述注册用户之前,所述方法还包括:The voiceprint recognition method according to claim 1, wherein the method further comprises: before the authenticating user is determined as the registered user corresponding to the registered voice information, the method further comprises: 确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序是否一致;以及Determining whether the ordering of the voice segments of the plurality of characters in the verification voice information is consistent with the ordering of the corresponding characters in the first character string; 在所述相似度达到预设的门限值,并且所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序一致的情况下,将所述验证用户确定为所述注册语音信息对应的所述注册用户。And if the similarity reaches a preset threshold value, and the order of the voice segments of the plurality of characters in the verification voice information is consistent with the ordering of the corresponding characters in the first character string, The verification user determines the registered user corresponding to the registration voice information. 如权利要求1-11中任一项所述的声纹识别方法,其特征在于,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述方法还包括:The voiceprint recognition method according to any one of claims 1 to 11, wherein the method further comprises: before acquiring the verification voice information generated by the verification user reading the first character string include: 随机生成所述第一字符串,并显示所述第一字符串。The first character string is randomly generated, and the first character string is displayed. 一种声纹识别装置,其特征在于,所述装置包括:A voiceprint recognition device, characterized in that the device comprises: 语音获取模块,用于获取验证用户朗读第一字符串所产生的验证语音信息;a voice acquiring module, configured to obtain verification voice information generated by the user to read the first character string; 语音片段识别模块,用于对所述验证语音信息进行语音识别,得到所述验证语音信息中包含的分别与所述第一字符串中的多个字符对应的语音片段;a voice segment identification module, configured to perform voice recognition on the verification voice information, and obtain a voice segment respectively included in the verification voice information corresponding to multiple characters in the first character string; 声纹特征提取模块,用于提取验证语音信息中各个字符对应的语音片段的声纹特征;a voiceprint feature extraction module, configured to extract a voiceprint feature of a voice segment corresponding to each character in the verification voice information; 特征模型训练模块,用于根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个 字符对应的特征向量;a feature model training module, configured to perform training according to a voice pattern feature of the voice segment corresponding to each character, and a common background model corresponding to the corresponding corresponding character, to obtain each of the verification voice information a feature vector corresponding to the character; 相似度判断模块,用于计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数,作为验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度;The similarity judging module is configured to calculate a similarity score of the feature vector corresponding to each character in the verification voice information and the corresponding character in the preset registration voice information, as a voice segment corresponding to each character in the verification voice information The similarity of the voiceprint features of the voice segment corresponding to the corresponding character in the preset registration voice information; 用户识别模块,用于在所述相似度达到预设的门限值时,将验证用户验证为所述注册语音信息对应的注册用户。The user identification module is configured to verify the verification user as the registered user corresponding to the registered voice information when the similarity reaches a preset threshold. 如权利要求15所述的声纹识别装置,其特征在于,A voiceprint recognition device according to claim 15, wherein: 所述语音获取模块,还用于获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符;The voice acquiring module is further configured to obtain registration voice information generated by the registered user reading the second character string, where the second character string and the first character string have at least one of the same characters; 所述特征模型训练模块,还用于根据所述注册语音信息中各个字符对应的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述注册语音信息中各个字符对应的特征向量。The feature model training module is further configured to: according to the voiceprint feature corresponding to each character in the registered voice information, and the common background model corresponding to the corresponding corresponding character, to obtain the feature corresponding to each character in the registered voice information. vector. 如权利要求15所述的声纹识别装置,其特征在于,所述特征模型训练模块用于:The voiceprint recognition device according to claim 15, wherein said feature model training module is configured to: 将所述验证语音信息中各个字符对应的语音片段的声纹特征作为训练样本数据,采用下式对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:The voiceprint feature of the voice segment corresponding to each character in the verification voice information is used as the training sample data, and the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted corresponding The general background model corresponding to the character has the largest posterior probability: M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为所述预设的超向量子空间矩阵,ω为所述验证语音信息中相应字符对应的特征向量,所述预设的超向量子空间矩阵为根据所述高斯混合模型的均值超向量中各个维度向量之间的相关性确定得到的;M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector a spatial matrix, ω is a feature vector corresponding to a corresponding character in the verification voice information, and the preset super-vector sub-space matrix is determined according to a correlation between each dimension vector in the mean super-vector of the Gaussian mixture model. of; 所述相似度判断模块用于:The similarity judgment module is used to: 计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值确定为所述相似度分数。 Calculating a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information, and determining the cosine distance value as the similarity score . 如权利要求15所述的声纹识别装置,其特征在于,所述语音片段识别模块包括:The voiceprint recognition device according to claim 15, wherein the voice segment recognition module comprises: 有效片段识别单元,用于识别所述验证语音信息中的有效语音片段和无效语音片段;a valid segment identification unit, configured to identify a valid speech segment and an invalid speech segment in the verification speech information; 语音识别单元,用于对所述有效语音片段进行语音识别得到分别与所述第一字符串中的多个字符对应的语音片段。The voice recognition unit is configured to perform voice recognition on the valid voice segment to obtain a voice segment corresponding to multiple characters in the first character string. 如权利要求15所述的声纹识别装置,其特征在于,还包括:The voiceprint recognition device according to claim 15, further comprising: 字符排序确定模块,用于确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序是否一致;a character order determining module, configured to determine whether a sorting of the voice segments of the plurality of characters in the verification voice information is consistent with a ranking of the corresponding characters in the first character string; 所述用户识别模块,还用于在所述相似度达到预设的门限值,并且所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序一致的情况下,将所述验证用户确定为所述注册语音信息对应的所述注册用户。The user identification module is further configured to: when the similarity reaches a preset threshold, and verify the order of the voice segments of the plurality of characters in the voice information and the corresponding characters in the first string In the case that the sorting is consistent, the verification user is determined as the registered user corresponding to the registered voice information. 如权利要求15-19中任一项所述的声纹识别装置,其特征在于,所述声纹识别装置还包括:The voiceprint recognition device according to any one of claims 15 to 19, wherein the voiceprint recognition device further comprises: 字符串显示模块,用于随机生成所述第一字符串,并显示所述第一字符串。a string display module, configured to randomly generate the first string, and display the first string. 一种声纹识别装置包括:A voiceprint recognition device includes: 用户接口,用于获取语音信息;a user interface for obtaining voice information; 存储器,存储计算机可执行程序代码;以及a memory storing computer executable program code; 处理器,用于调用所述计算机可执行程序代码以执行以下操作:a processor for invoking the computer executable program code to perform the following operations: 通过所述用户接口获取验证用户朗读第一字符串所产生的验证语音信息;Acquiring, by the user interface, verification voice information generated by the user to read the first character string; 对所述验证语音信息进行语音识别,得到所述验证语音信息中包含的分别与所述第一字符串中的各个字符对应的语音片段;Performing voice recognition on the verification voice information, and obtaining a voice segment respectively corresponding to each character in the first character string included in the verification voice information; 提取各个字符对应的语音片段的声纹特征;Extracting a voiceprint feature of the voice segment corresponding to each character; 根据所述各个字符对应的语音片段的声纹特征,结合预设的相应字符对应 的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量;Corresponding to the preset corresponding characters according to the voiceprint features of the voice segments corresponding to the respective characters The universal background model is trained to obtain a feature vector corresponding to each character in the verification voice information; 计算所述验证语音信息中各个字符对应的特征向量与所述预设的注册语音信息中相应字符对应的特征向量的相似度分数,作为验证语音信息中各个字符对应的语音片段与预设的注册语音信息中相应字符对应的语音片段的声纹特征的相似度;Calculating a similarity score of the feature vector corresponding to each character in the verification voice information and the feature vector corresponding to the corresponding character in the preset registration voice information, as a voice segment corresponding to each character in the verification voice information and a preset registration The similarity of the voiceprint features of the voice segments corresponding to the corresponding characters in the voice information; 当所述相似度达到预设的门限值时,将验证用户验证为所述注册语音信息对应的注册用户。When the similarity reaches a preset threshold, the user verification is verified as the registered user corresponding to the registered voice information. 如权利要求21所述的声纹识别装置,其特征在于,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,之前,所述处理器还调用所述计算机可执行程序代码执行以下操作:The voiceprint recognition apparatus according to claim 21, wherein said processor further calls said processor before said verifying said verification voice information generated by said user reading said first character string The computer executable code performs the following actions: 获取注册用户朗读第二字符串所产生的注册语音信息,所述第二字符串与所述第一字符串拥有至少一个相同的字符;Obtaining registration voice information generated by the registered user reading the second character string, the second character string having at least one of the same characters as the first character string; 根据所述注册语音信息中各个字符对应的声纹特征,结合预设的相应字符对应的通用背景模型训练,得到所述注册语音信息中各个字符对应的特征向量。The feature vector corresponding to each character in the registered voice information is obtained according to the voiceprint feature corresponding to each character in the registered voice information, combined with the common background model corresponding to the preset corresponding character. 如权利要求21所述的声纹识别装置,其特征在于,所述处理器调用所述计算机可执行程序代码执行以下操作以根据所述各个字符对应的语音片段的声纹特征,结合所述预设的相应字符对应的通用背景模型训练,得到所述验证语音信息中各个字符对应的特征向量:A voiceprint recognition apparatus according to claim 21, wherein said processor calls said computer executable program code to perform an operation of combining said sound according to a voiceprint feature of said voice segment corresponding to said respective character The general background model corresponding to the corresponding character is trained to obtain the feature vector corresponding to each character in the verification voice information: 将所述验证语音信息中各个字符对应的语音片段的声纹特征作为所述训练样本数据,采用下式对所述预设的相应字符对应的通用背景模型的均值超向量进行调整,使得调整后的相应字符对应的通用背景模型的后验概率最大:Using the voiceprint feature of the voice segment corresponding to each character in the verification voice information as the training sample data, the mean supervector of the common background model corresponding to the preset corresponding character is adjusted by using the following formula, so that the adjusted The general background model corresponding to the corresponding character has the largest posterior probability: M=m+Tω,其中M代表调整后的某个字符的通用背景模型的均值超向量,m代表调整前的相应字符的通用背景模型的均值超向量,T为所述预设的超向量子空间矩阵,ω为所述验证语音信息中相应字符对应的特征向量,所述预设的超向量子空间矩阵为根据所述通用背景模型中各个高斯模块的权重之间的相关性确定得到的;M=m+Tω, where M represents the mean supervector of the universal background model of the adjusted character, m represents the mean supervector of the universal background model of the corresponding character before adjustment, and T is the preset supervector a spatial matrix, ω is a feature vector corresponding to a corresponding character in the verification voice information, and the preset super-vector sub-space matrix is determined according to a correlation between weights of respective Gaussian modules in the universal background model; 所述处理器调用所述计算机可执行程序代码执行以下操作以计算所述验 证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量的相似度分数:The processor invoking the computer executable program code to perform the following operations to calculate the test The similarity score of the feature vector corresponding to each character in the voice information and the corresponding character in the preset registered voice information: 计算所述验证语音信息中各个字符对应的特征向量与预设的注册语音信息中相应字符对应的特征向量之间的余弦距离值,并将所述余弦距离值确定为所述相似度分数。Calculating a cosine distance value between a feature vector corresponding to each character in the verification voice information and a feature vector corresponding to a corresponding character in the preset registration voice information, and determining the cosine distance value as the similarity score. 如权利要求21所述的声纹识别装置,其特征在于,所述处理器调用所述计算机可执行程序代码执行以下操作以对所述验证语音信息进行语音识别得到所述验证语音信息中包含的分别与所述第一字符串中的多个字符对应的语音片段:A voiceprint recognition apparatus according to claim 21, wherein said processor calls said computer executable program code to perform the following operations to perform voice recognition on said verification voice information to obtain said voice information included in said verification voice information a speech segment corresponding to a plurality of characters in the first character string: 识别所述验证语音信息中的有效语音片段和无效语音片段;Identifying valid voice segments and invalid voice segments in the verification voice information; 对所述有效语音片段进行语音识别得到分别与所述第一字符串中的多个字符对应的语音片段。Performing speech recognition on the valid speech segment to obtain a speech segment respectively corresponding to a plurality of characters in the first character string. 如权利要求21所述的声纹识别装置,其特征在于,所述将所述验证用户确定为所述注册语音信息对应的所述注册用户之前,所述处理器还调用所述计算机可执行程序代码执行以下操作:The voiceprint identifying apparatus according to claim 21, wherein said processor further calls said computer executable program before said verifying user determines said registered user corresponding to said registered voice information The code does the following: 确定所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序是否一致;以及Determining whether the ordering of the voice segments of the plurality of characters in the verification voice information is consistent with the ordering of the corresponding characters in the first character string; 在所述相似度达到预设的门限值,并且所述验证语音信息中的多个字符的语音片段的排序与所述第一字符串中的相应字符的排序一致的情况下,将所述验证用户确定为所述注册语音信息对应的所述注册用户。And if the similarity reaches a preset threshold value, and the order of the voice segments of the plurality of characters in the verification voice information is consistent with the ordering of the corresponding characters in the first character string, The verification user determines the registered user corresponding to the registration voice information. 如权利要求21-25中任一项所述的声纹识别装置,其特征在于,所述获取所述验证用户朗读所述第一字符串所产生的所述验证语音信息之前,所述处理器还调用所述计算机可执行程序代码执行以下操作:A voiceprint recognition apparatus according to any one of claims 21 to 25, wherein said processor is obtained before said verification voice information generated by said verification user reading said first character string The computer executable program code is also invoked to perform the following operations: 随机生成所述第一字符串,并通过所述用户接口显示所述第一字符串。The first character string is randomly generated, and the first character string is displayed through the user interface. 一种存储介质,其特征在于,所述存储介质中存储有计算机程序,所 述计算机程序用以执行如权利要求1-14中任一项所述的声纹识别方法。 A storage medium, characterized in that a computer program is stored in the storage medium The computer program is for performing the voiceprint recognition method according to any one of claims 1-14.
PCT/CN2017/087911 2016-06-12 2017-06-12 Voiceprint recognition method and device Ceased WO2017215558A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201610416650.3 2016-06-12
CN201610416650.3A CN106098068B (en) 2016-06-12 2016-06-12 A kind of method for recognizing sound-groove and device

Publications (1)

Publication Number Publication Date
WO2017215558A1 true WO2017215558A1 (en) 2017-12-21

Family

ID=57846666

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/087911 Ceased WO2017215558A1 (en) 2016-06-12 2017-06-12 Voiceprint recognition method and device

Country Status (2)

Country Link
CN (1) CN106098068B (en)
WO (1) WO2017215558A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109147767A (en) * 2018-08-16 2019-01-04 平安科技(深圳)有限公司 Digit recognition method, device, computer equipment and storage medium in voice
CN110738998A (en) * 2019-09-11 2020-01-31 深圳壹账通智能科技有限公司 Voice-based personal credit evaluation method, device, terminal and storage medium
CN111199729A (en) * 2018-11-19 2020-05-26 阿里巴巴集团控股有限公司 Voiceprint recognition method and device
CN112037815A (en) * 2020-08-28 2020-12-04 中移(杭州)信息技术有限公司 Audio fingerprint extraction method, server and storage medium
CN112435673A (en) * 2020-12-15 2021-03-02 北京声智科技有限公司 Model training method and electronic terminal
CN115550075A (en) * 2022-12-01 2022-12-30 中网道科技集团股份有限公司 Anti-counterfeiting processing method and device for public welfare activity data of community correction object
CN115602177A (en) * 2022-10-12 2023-01-13 中国电信股份有限公司(Cn) Voiceprint recognition method and device and computer-readable storage medium
CN115641852A (en) * 2022-10-18 2023-01-24 中国电信股份有限公司 Voiceprint recognition method and device, electronic equipment and computer readable storage medium
WO2024077588A1 (en) * 2022-10-14 2024-04-18 Qualcomm Incorporated Voice-based user authentication

Families Citing this family (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106098068B (en) * 2016-06-12 2019-07-16 腾讯科技(深圳)有限公司 A kind of method for recognizing sound-groove and device
CN110169014A (en) * 2017-01-03 2019-08-23 诺基亚技术有限公司 Device, method and computer program product for certification
CN108447471B (en) * 2017-02-15 2021-09-10 腾讯科技(深圳)有限公司 Speech recognition method and speech recognition device
CN107610708B (en) * 2017-06-09 2018-06-19 平安科技(深圳)有限公司 Identify the method and apparatus of vocal print
CN109102812B (en) * 2017-06-21 2021-08-31 北京搜狗科技发展有限公司 Voiceprint recognition method and system and electronic equipment
CN107492379B (en) 2017-06-30 2021-09-21 百度在线网络技术(北京)有限公司 Voiceprint creating and registering method and device
CN107248410A (en) * 2017-07-19 2017-10-13 浙江联运知慧科技有限公司 The method that Application on Voiceprint Recognition dustbin opens the door
CN109559759B (en) * 2017-09-27 2021-10-08 华硕电脑股份有限公司 Electronic device with incremental registration unit and method thereof
CN110310647B (en) * 2017-09-29 2022-02-25 腾讯科技(深圳)有限公司 A voice identity feature extractor, classifier training method and related equipment
CN107886943A (en) * 2017-11-21 2018-04-06 广州势必可赢网络科技有限公司 Voiceprint recognition method and device
CN108154588B (en) * 2017-12-29 2020-11-27 深圳市艾特智能科技有限公司 Unlocking method and system, readable storage medium and intelligent device
CN110047491A (en) * 2018-01-16 2019-07-23 中国科学院声学研究所 A kind of relevant method for distinguishing speek person of random digit password and device
CN108269590A (en) * 2018-01-17 2018-07-10 广州势必可赢网络科技有限公司 Vocal cord recovery scoring method and device
CN108447489B (en) * 2018-04-17 2020-05-22 清华大学 A continuous voiceprint authentication method and system with feedback
CN110875044B (en) * 2018-08-30 2022-05-03 中国科学院声学研究所 A speaker recognition method based on word correlation score calculation
CN109117622B (en) * 2018-09-19 2020-09-01 北京容联易通信息技术有限公司 Identity authentication method based on audio fingerprints
CN109257362A (en) * 2018-10-11 2019-01-22 平安科技(深圳)有限公司 Method, apparatus, computer equipment and the storage medium of voice print verification
CN109473107B (en) * 2018-12-03 2020-12-22 厦门快商通信息技术有限公司 Text semi-correlation voiceprint recognition method and system
CN111669350A (en) * 2019-03-05 2020-09-15 阿里巴巴集团控股有限公司 Identity verification method, verification information generation method, payment method and payment device
CN110600041B (en) * 2019-07-29 2022-04-29 华为技术有限公司 Method and device for voiceprint recognition
CN110517695A (en) * 2019-09-11 2019-11-29 国微集团(深圳)有限公司 Verification method and device based on vocal print
CN110971763B (en) * 2019-12-10 2021-01-26 Oppo广东移动通信有限公司 Arrival reminding method and device, storage medium and electronic equipment
CN110956732A (en) * 2019-12-19 2020-04-03 重庆特斯联智慧科技股份有限公司 Safety entrance guard based on thing networking
CN111081256A (en) * 2019-12-31 2020-04-28 苏州思必驰信息科技有限公司 Digital string voiceprint password verification method and system
CN111081260A (en) * 2019-12-31 2020-04-28 苏州思必驰信息科技有限公司 Method and system for identifying voiceprint of awakening word
CN111597531A (en) * 2020-04-07 2020-08-28 北京捷通华声科技股份有限公司 Identity authentication method and device, electronic equipment and readable storage medium
CN111613230A (en) * 2020-06-24 2020-09-01 泰康保险集团股份有限公司 Voiceprint verification method, voiceprint verification device, voiceprint verification equipment and storage medium
CN112487384B (en) * 2020-11-25 2024-12-03 华为技术有限公司 Identity verification method and system
CN112820299B (en) * 2020-12-29 2021-09-14 马上消费金融股份有限公司 Voiceprint recognition model training method and device and related equipment
CN113113022A (en) * 2021-04-15 2021-07-13 吉林大学 Method for automatically identifying identity based on voiceprint information of speaker
CN113570754B (en) * 2021-07-01 2022-04-29 汉王科技股份有限公司 Voiceprint lock control method and device and electronic equipment
CN114357417B (en) * 2021-12-31 2025-04-08 中国科学院声学研究所东海研究站 A self-learning dynamic voiceprint identity authentication method based on unknown corpus
CN114782141A (en) * 2022-05-07 2022-07-22 中国工商银行股份有限公司 Product interaction method and device based on 5G message, electronic equipment and medium
CN115019808B (en) * 2022-06-01 2025-07-11 科大讯飞股份有限公司 Voiceprint extraction method, device, equipment and readable storage medium
CN116530944B (en) * 2023-07-06 2023-10-20 荣耀终端有限公司 Sound processing method and electronic equipment
CN116978368B (en) * 2023-09-25 2023-12-15 腾讯科技(深圳)有限公司 Wake-up word detection method and related device
CN120279894B (en) * 2025-05-14 2025-10-21 广东公信智能会议股份有限公司 Human-computer interaction method and system for speech recognition

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050033573A1 (en) * 2001-08-09 2005-02-10 Sang-Jin Hong Voice registration method and system, and voice recognition method and system based on voice registration method and system
CN102238189A (en) * 2011-08-01 2011-11-09 安徽科大讯飞信息科技股份有限公司 Voiceprint password authentication method and system
CN105096121A (en) * 2015-06-25 2015-11-25 百度在线网络技术(北京)有限公司 Voiceprint authentication method and device
CN105656887A (en) * 2015-12-30 2016-06-08 百度在线网络技术(北京)有限公司 Artificial intelligence-based voiceprint authentication method and device
CN106098068A (en) * 2016-06-12 2016-11-09 腾讯科技(深圳)有限公司 A kind of method for recognizing sound-groove and device

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102254559A (en) * 2010-05-20 2011-11-23 盛乐信息技术(上海)有限公司 Identity authentication system and method based on vocal print
CN102314877A (en) * 2010-07-08 2012-01-11 盛乐信息技术(上海)有限公司 Voiceprint identification method for character content prompt
CN101997689B (en) * 2010-11-19 2012-08-08 吉林大学 USB (universal serial bus) identity authentication method based on voiceprint recognition and system thereof
CN102163427B (en) * 2010-12-20 2012-09-12 北京邮电大学 Method for detecting audio exceptional event based on environmental model
CN102737634A (en) * 2012-05-29 2012-10-17 百度在线网络技术(北京)有限公司 Authentication method and device based on voice
CN103679452A (en) * 2013-06-20 2014-03-26 腾讯科技(深圳)有限公司 Payment authentication method, device thereof and system thereof
CN104282303B (en) * 2013-07-09 2019-03-29 威盛电子股份有限公司 Method for voice recognition by voiceprint recognition and electronic device thereof
CN104064189A (en) * 2014-06-26 2014-09-24 厦门天聪智能软件有限公司 Vocal print dynamic password modeling and verification method
CN104575504A (en) * 2014-12-24 2015-04-29 上海师范大学 Method for personalized television voice wake-up by voiceprint and voice identification
CN104901808A (en) * 2015-04-14 2015-09-09 时代亿宝(北京)科技有限公司 Voiceprint authentication system and method based on time type dynamic password

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050033573A1 (en) * 2001-08-09 2005-02-10 Sang-Jin Hong Voice registration method and system, and voice recognition method and system based on voice registration method and system
CN102238189A (en) * 2011-08-01 2011-11-09 安徽科大讯飞信息科技股份有限公司 Voiceprint password authentication method and system
CN105096121A (en) * 2015-06-25 2015-11-25 百度在线网络技术(北京)有限公司 Voiceprint authentication method and device
CN105656887A (en) * 2015-12-30 2016-06-08 百度在线网络技术(北京)有限公司 Artificial intelligence-based voiceprint authentication method and device
CN106098068A (en) * 2016-06-12 2016-11-09 腾讯科技(深圳)有限公司 A kind of method for recognizing sound-groove and device

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109147767A (en) * 2018-08-16 2019-01-04 平安科技(深圳)有限公司 Digit recognition method, device, computer equipment and storage medium in voice
CN111199729A (en) * 2018-11-19 2020-05-26 阿里巴巴集团控股有限公司 Voiceprint recognition method and device
CN111199729B (en) * 2018-11-19 2023-09-26 阿里巴巴集团控股有限公司 Voiceprint recognition method and voiceprint recognition device
CN110738998A (en) * 2019-09-11 2020-01-31 深圳壹账通智能科技有限公司 Voice-based personal credit evaluation method, device, terminal and storage medium
CN112037815A (en) * 2020-08-28 2020-12-04 中移(杭州)信息技术有限公司 Audio fingerprint extraction method, server and storage medium
CN112435673A (en) * 2020-12-15 2021-03-02 北京声智科技有限公司 Model training method and electronic terminal
CN112435673B (en) * 2020-12-15 2024-05-14 北京声智科技有限公司 Model training method and electronic terminal
CN115602177A (en) * 2022-10-12 2023-01-13 中国电信股份有限公司(Cn) Voiceprint recognition method and device and computer-readable storage medium
WO2024077588A1 (en) * 2022-10-14 2024-04-18 Qualcomm Incorporated Voice-based user authentication
CN115641852A (en) * 2022-10-18 2023-01-24 中国电信股份有限公司 Voiceprint recognition method and device, electronic equipment and computer readable storage medium
CN115550075A (en) * 2022-12-01 2022-12-30 中网道科技集团股份有限公司 Anti-counterfeiting processing method and device for public welfare activity data of community correction object
CN115550075B (en) * 2022-12-01 2023-05-09 中网道科技集团股份有限公司 Anti-counterfeiting processing method and equipment for community correction object public welfare activity data

Also Published As

Publication number Publication date
CN106098068B (en) 2019-07-16
CN106098068A (en) 2016-11-09

Similar Documents

Publication Publication Date Title
WO2017215558A1 (en) Voiceprint recognition method and device
US11727942B2 (en) Age compensation in biometric systems using time-interval, gender and age
KR102239129B1 (en) End-to-end speaker recognition using deep neural network
RU2738325C2 (en) Method and device for authenticating an individual
Dey et al. Speech biometric based attendance system
CN103475490B (en) A kind of auth method and device
Das et al. Development of multi-level speech based person authentication system
WO2017113658A1 (en) Artificial intelligence-based method and device for voiceprint authentication
US11348590B2 (en) Methods and devices for registering voiceprint and for authenticating voiceprint
WO2017114307A1 (en) Voiceprint authentication method capable of preventing recording attack, server, terminal, and system
WO2019019256A1 (en) Electronic apparatus, identity verification method and system, and computer-readable storage medium
Mansour et al. Voice recognition using dynamic time warping and mel-frequency cepstral coefficients algorithms
WO2017162053A1 (en) Identity authentication method and device
Korshunov et al. Impact of score fusion on voice biometrics and presentation attack detection in cross-database evaluations
US10909991B2 (en) System for text-dependent speaker recognition and method thereof
EP2879130A1 (en) Methods and systems for splitting a digital signal
Saquib et al. A survey on automatic speaker recognition systems
CN110111798B (en) Method, terminal and computer readable storage medium for identifying speaker
CN107346568A (en) The authentication method and device of a kind of gate control system
JP2007133414A (en) Speech discrimination capability estimation method and apparatus, and speaker authentication registration and evaluation method and apparatus
CN110379433A (en) Method, apparatus, computer equipment and the storage medium of authentication
Shirvanian et al. Quantifying the breakability of voice assistants
JP7259981B2 (en) Speaker authentication system, method and program
JP2016166927A (en) Parameter learning device, speaker recognition device, parameter learning method, speaker recognition method, and program
Georgescu et al. GMM-UBM modeling for speaker recognition on a Romanian large speech corpora

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17812669

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17812669

Country of ref document: EP

Kind code of ref document: A1