Detailed Description
In order that those skilled in the art will better understand the present application, a technical solution in the embodiments of the present application will be clearly and completely described below with reference to the accompanying drawings in which it is apparent that the described embodiments are only some embodiments of the present application, not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
The terms first, second and the like in the description and in the claims and in the above-described figures are used for distinguishing between different objects and not necessarily for describing a sequential or chronological order. Furthermore, the terms "comprise" and "have," as well as any variations thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those listed steps or elements but may include other steps or elements not listed or inherent to such process, method, article, or apparatus.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the application. The appearances of such phrases in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those of skill in the art will explicitly and implicitly appreciate that the embodiments described herein may be combined with other embodiments.
In the embodiment of the application, the appointed equipment can comprise a smart Phone (such as an Android Mobile Phone, an iOS Mobile Phone, a Windows Phone Mobile Phone and the like), a palm computer, a tablet personal computer, a Bluetooth sound box, a smart television, a smart refrigerator, a smart robot, a driving recorder, a notebook computer, a Mobile internet device (Mobile INTERNET DEVICES, MID) or a wearable device (such as smart AI glasses, smart bracelets, smart watches and Bluetooth headphones) and the like, which are merely examples, but are not limited to the appointed equipment, and the appointed equipment can also comprise a server, such as a cloud server.
Embodiments of the present application are described in detail below.
Referring to fig. 1, fig. 1 is a flow chart of an expression processing method provided by an embodiment of the present application, which is applied to intelligent AI glasses, wherein the intelligent AI glasses include a sensor and a communication module, and as shown in the figure, the expression processing method includes:
101. and capturing facial expression images of a target object through the sensor, wherein the target object is a user wearing the intelligent AI glasses or is a shooting object shot by the intelligent AI glasses.
The target object may be a user wearing the intelligent AI glasses, or the target object may be a shooting object shot by the intelligent AI glasses.
As shown in fig. 2, the smart AI glasses include a sensor for capturing a facial expression to obtain a facial expression image, and a communication module for implementing a communication function, where the communication module may include at least one of a mobile communication module (2G, 3G, 4G, 5G, etc.), a wireless fidelity (WIRELESS FIDELITY, wi-Fi), a bluetooth communication module, an infrared communication module, a millimeter wave communication module, a radar communication module, etc., which are not limited herein.
In the embodiment of the application, the sensor can comprise one or more sensors, and the sensor can comprise at least one of a camera, a temperature sensor, a humidity sensor, a substance detection sensor, a myoelectric sensor and the like, and is not limited herein.
Wherein the substance detection sensor may be used to detect facial skin parameters of the target object and the myoelectric sensor may be used to detect facial muscle movement parameters of the target object.
In a specific implementation, when the target object is a user wearing the intelligent AI glasses, the sensor may capture a facial expression image of the target object, for example, the facial expression image may be obtained by photographing the face of the target object, or when the target object is a photographing object photographed by the intelligent AI glasses, the sensor may capture a facial expression image of the target object, for example, the facial expression image may be obtained by photographing the face of the photographing object.
102. And extracting the facial expression image to obtain a target feature set.
In a specific implementation, the target feature set may include at least one feature, which may include at least one feature selected from the group consisting of feature points, feature textures, feature regions, feature values, feature vectors, and the like, without limitation.
Specifically, the facial expression image can be extracted to obtain a target feature set, and the target feature set can be used for realizing expression recognition.
103. And generating a target expression image in a preset format according to the target feature set.
The preset format may be preset or default, for example, the preset format may be any expression packet format, and the expression packet format may be a static expression packet format or a dynamic expression packet format, for example, the expression packet format may include an emoji expression packet format.
In a specific implementation, an emotion symbol generator may be utilized to generate a target emotion image in a preset format according to a target feature set, for example, the target emotion image may be sent to a friend by capturing a facial expression of a user and converting the facial expression into a corresponding emoji. The emoji generator may be preset or default to the system, e.g., the emoji generator may include machine learning models, neural network models, large models, etc., without limitation.
104. And sending the target expression image to a designated device through the communication module.
The designated device may be preset or default, and the designated device may be a part of the smart AI glasses or may also be a device that communicates with other than the smart AI glasses.
Wherein the designated device may comprise one or more devices.
In the specific implementation, the target expression image can be sent to the appointed equipment through the communication module, so that the expression image corresponding to the user can be generated through the intelligent AI glasses, the target expression image is obtained and then sent to the appointed equipment, and the intelligence and the interestingness of the intelligent AI glasses can be improved.
Optionally, the step 102 of extracting an image of the facial expression image to obtain a target feature set may include the following steps:
Extracting features of the facial expression image to obtain a first feature set;
Determining a target expression type corresponding to the first feature set;
And determining the characteristics corresponding to the target expression type according to the first characteristic set to obtain the target characteristic set.
In a specific implementation, the facial expression image may be subjected to feature extraction to obtain a first feature set, and the first feature set may be input into a classification network to obtain a corresponding expression type, where the expression type may include at least one of smiling, laughing, smiling, bending, embarrassing, gazing, and the like, and is not limited herein. The classification network may also be preset or default to the system, e.g., the classification network may include at least one of convolutional neural networks, large models, etc., without limitation.
In a specific implementation, the target expression type corresponding to the first feature set and different expression types can be determined based on the classification network, so that the required features are different when the target expression image is generated, and further, the features corresponding to the target expression type can be determined according to the first feature set to obtain the target feature set, and thus, the intelligent and interesting of the intelligent AI glasses can be improved based on the corresponding features of the preset format and the target expression type.
Optionally, the step of determining the feature corresponding to the target expression type according to the first feature set to obtain the target feature set may include the following steps:
Acquiring a feature selection rule set corresponding to the preset format to obtain a plurality of feature selection rule sets, wherein each feature selection rule set corresponds to one expression type;
determining a target feature selection rule set corresponding to the target expression type from the feature selection rule sets;
And screening the first feature set according to the target feature selection rule set to obtain the target feature set.
In a specific implementation, a feature selection rule set corresponding to a preset format may be obtained to obtain a plurality of feature selection rule sets, where each feature selection rule set corresponds to one expression type. The feature selection rule set may include at least one feature selection rule, and the feature selection rule may include at least one of which type of feature (e.g., feature point, feature value, feature vector, feature texture, etc.), which location or region of the feature is selected, which degree of feature is selected, and so on, without limitation.
Then, a target feature selection rule set corresponding to the target expression type can be determined from the feature selection rule sets, and then the first feature set is screened according to the target feature selection rule set to obtain a target feature set.
Optionally, the step 103 of generating the target expression image in the preset format according to the target feature set may include the following steps:
acquiring a reference expression template set corresponding to the preset format, wherein the reference expression template set comprises a plurality of reference expression templates, and each reference expression template corresponds to one expression type;
determining a target reference expression template corresponding to the target expression type from the reference expression template set;
and determining the target expression image according to the target feature set and the target reference expression template.
In a specific implementation, a reference expression template set corresponding to a preset format can be obtained, the reference expression template set can comprise a plurality of reference expression templates, each reference expression template corresponds to one expression type, then a target reference expression template corresponding to a target expression type can be determined from the reference expression template set, and then a target expression image is determined according to the target feature set and the target reference expression template, so that a target reference expression template corresponding to the target expression type can be selected, and then a corresponding expression image is synthesized based on the target reference expression template and the corresponding features of the preset format and the target expression type, thereby being beneficial to improving the intelligence and the interestingness of the intelligent AI glasses.
Optionally, the step of determining the target expression image according to the target feature set and the target reference expression template may include the steps of:
acquiring a feature position area corresponding to each feature in the target feature set to obtain at least one feature position area;
Marking corresponding areas in the target reference expression template according to the at least one characteristic position area to obtain at least one area;
And filling the target feature set into a corresponding region in the at least one region in the target reference expression template to obtain the target expression image.
In specific implementation, a feature position area corresponding to each feature in the target feature set can be obtained to obtain at least one feature position area, the corresponding area is marked according to the at least one feature position area in the target reference expression template to obtain at least one area, finally, the target feature set is filled into the corresponding area in the at least one area in the target reference expression template to obtain a target expression image, namely the target feature set can be fused to the corresponding position in the target reference expression template, so that the target expression image is fused to some features of the target object, the target expression image is related to the facial feature depth of the target object, and the intelligent AI glasses are improved in intelligence and interestingness.
Optionally, in the step 102, feature extraction is performed on the facial expression image to obtain a first feature set, which may include the following steps:
Acquiring target shooting parameters of the target expression image;
acquiring target skin color parameters of the target expression image;
determining a target feature extraction algorithm corresponding to the target skin color parameters;
determining target algorithm control parameters of the target feature extraction algorithm corresponding to the target shooting parameters;
and carrying out feature extraction on the facial expression image according to the target feature extraction algorithm and the target algorithm control parameter to obtain the first feature set.
The target shooting parameters may include at least one of shooting angle, shooting distance, sensitivity, exposure time, etc., which are not limited herein.
The target skin tone parameters may include at least one of skin tone type, skin tone degree, skin tone age, etc., without limitation herein.
In specific implementation, the target shooting parameters of the target expression image can be obtained, the target skin color parameters of the target expression image can be obtained, the mapping relation between the preset skin color parameters and the feature extraction algorithm can be stored in advance, the target feature extraction algorithm corresponding to the target skin color parameters can be determined based on the mapping relation, the mapping relation between the preset shooting parameters and the algorithm control parameters of the target feature extraction algorithm can be stored in advance, and the target algorithm control parameters of the target feature extraction algorithm corresponding to the target shooting parameters can be determined based on the mapping relation.
The target algorithm control parameter may be used to control a feature extraction effect of the target feature extraction algorithm, where the feature extraction effect may include at least one of a feature type, a feature extraction degree, a feature extraction speed, a feature extraction area, and the like, and is not limited herein.
The target feature extraction algorithm may include one or more feature extraction algorithms, each of which may be used to extract one type of feature.
Then, feature extraction can be performed on the facial expression image according to a target feature extraction algorithm and a target algorithm control parameter to obtain a first feature set, so that on one hand, the corresponding target feature extraction algorithm can be determined based on skin color parameters of a target object, and further, the feature extraction integrity and the accuracy are preliminarily ensured, and on the other hand, the algorithm control parameter of the target feature extraction algorithm can be optimized based on shooting parameter depth, so that the final algorithm control parameter depth accords with the characteristics of the image, and thus, the feature extraction integrity and the accuracy can be ensured in depth, thereby helping to ensure that a target expression image matched with the depth of the target object is obtained, and further, the intelligent and interesting of the intelligent AI glasses are promoted.
Optionally, the smart AI glasses further communicate with the wearable device of the target object, and the step of determining the target algorithm control parameter of the target feature extraction algorithm corresponding to the target shooting parameter may include the following steps:
acquiring the capturing moment of the facial expression image;
Acquiring, by the communication module, a target physiological state parameter of the target object acquired by the wearable device;
Determining a first algorithm control parameter of the target feature extraction algorithm corresponding to the target shooting parameter;
determining a first optimization parameter corresponding to the target physiological state parameter;
And determining the target algorithm control parameter according to the first optimization parameter and the first algorithm control parameter.
Wherein the target physiological state parameter may include at least one of blood pressure, muscle movement parameter, blood sugar, blood fat, brain wave parameter, vein parameter, electrocardiogram, etc., without limitation. The target physiological state parameter reflects to some extent the emotional condition of the target subject.
In a specific implementation, the capturing moment of the facial expression image can be obtained, the target physiological state parameter of the target object acquired by the wearable device can be obtained through the communication module, the mapping relation between the preset physiological state parameter and the optimization parameter can be stored in advance, and further, the first optimization parameter corresponding to the target physiological state parameter can be determined based on the mapping relation.
The mapping relation between the preset shooting parameters and the algorithm control parameters of the target feature extraction algorithm can be stored in advance, further, the first algorithm control parameters of the target feature extraction algorithm corresponding to the target shooting parameters can be determined based on the mapping relation, and the target algorithm control parameters can be determined according to the first optimization parameters and the first algorithm control parameters, namely, the target algorithm control parameters= (1+the first optimization parameters) & gt the first algorithm control parameters, so that the algorithm control parameters can be optimized based on the physiological feature depth of the user, the final algorithm control parameters further deeply conform to the emotion of the target object, the feature extraction integrity and the accuracy can be guaranteed in depth, and accordingly, the target expression image which is matched with the depth of the target object can be guaranteed, and the intelligent and the interestingness of the intelligent AI glasses can be improved.
It can be seen that the expression processing method described in the embodiment of the application is applied to intelligent AI glasses, the intelligent AI glasses comprise a sensor and a communication module, facial expression images of target objects are captured through the sensor, the target objects are users wearing the intelligent AI glasses, or the target objects are shooting objects shot by the intelligent AI glasses, image extraction is performed on the facial expression images to obtain a target feature set, a target expression image with a preset format is generated according to the target feature set, and the target expression image is sent to a designated device through the communication module, so that facial expression images of users can be collected, corresponding target expression images with the preset format are generated, and are shared with other users (friends), and therefore intelligence and interestingness of the intelligent AI glasses can be improved, and user experience is improved.
In accordance with the above embodiment, please refer to fig. 3, fig. 3 is a schematic structural diagram of another smart AI glasses provided in an embodiment of the present application, as shown in the drawing, the smart AI glasses include a processor, a memory, a communication interface, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the processor, and in the embodiment of the present application, the smart AI glasses include a sensor and a communication module, and the programs include instructions for executing the following steps:
capturing a facial expression image of a target object through the sensor, wherein the target object is a user wearing the intelligent AI glasses or is a shooting object shot by the intelligent AI glasses;
extracting the facial expression image to obtain a target feature set;
generating a target expression image in a preset format according to the target feature set;
and sending the target expression image to a designated device through the communication module.
Optionally, in the aspect of extracting the image of the facial expression image to obtain a target feature set, the program includes instructions for performing the following steps:
Extracting features of the facial expression image to obtain a first feature set;
Determining a target expression type corresponding to the first feature set;
And determining the characteristics corresponding to the target expression type according to the first characteristic set to obtain the target characteristic set.
Optionally, in the aspect of determining the feature corresponding to the target expression type according to the first feature set to obtain the target feature set, the program includes instructions for executing the following steps:
Acquiring a feature selection rule set corresponding to the preset format to obtain a plurality of feature selection rule sets, wherein each feature selection rule set corresponds to one expression type;
determining a target feature selection rule set corresponding to the target expression type from the feature selection rule sets;
And screening the first feature set according to the target feature selection rule set to obtain the target feature set.
Optionally, in the aspect of generating the target expression image in a preset format according to the target feature set, the program includes instructions for executing the following steps:
acquiring a reference expression template set corresponding to the preset format, wherein the reference expression template set comprises a plurality of reference expression templates, and each reference expression template corresponds to one expression type;
determining a target reference expression template corresponding to the target expression type from the reference expression template set;
and determining the target expression image according to the target feature set and the target reference expression template.
Optionally, in said determining said target expression image from said target feature set and said target reference expression template, the above procedure comprises instructions for performing the steps of:
acquiring a feature position area corresponding to each feature in the target feature set to obtain at least one feature position area;
Marking corresponding areas in the target reference expression template according to the at least one characteristic position area to obtain at least one area;
And filling the target feature set into a corresponding region in the at least one region in the target reference expression template to obtain the target expression image.
Optionally, in the aspect of extracting features of the facial expression image to obtain a first feature set, the program includes instructions for:
Acquiring target shooting parameters of the target expression image;
acquiring target skin color parameters of the target expression image;
determining a target feature extraction algorithm corresponding to the target skin color parameters;
determining target algorithm control parameters of the target feature extraction algorithm corresponding to the target shooting parameters;
and carrying out feature extraction on the facial expression image according to the target feature extraction algorithm and the target algorithm control parameter to obtain the first feature set.
Optionally, the smart AI glasses further communicate with the wearable device of the target object, and the step of determining the target algorithm control parameter of the target feature extraction algorithm corresponding to the target shooting parameter may include the following steps:
acquiring the capturing moment of the facial expression image;
Acquiring, by the communication module, a target physiological state parameter of the target object acquired by the wearable device;
Determining a first algorithm control parameter of the target feature extraction algorithm corresponding to the target shooting parameter;
determining a first optimization parameter corresponding to the target physiological state parameter;
And determining the target algorithm control parameter according to the first optimization parameter and the first algorithm control parameter.
It can be seen that, the intelligent AI glasses described in the embodiments of the present application include a sensor and a communication module, the sensor captures a facial expression image of a target object, the target object is a user wearing the intelligent AI glasses, or the target object is a shooting object shot by the intelligent AI glasses, image extraction is performed on the facial expression image to obtain a target feature set, a target expression image in a preset format is generated according to the target feature set, and the communication module sends the target expression image to a designated device, so that the facial expression image of the user can be acquired, and a target expression image in a corresponding preset format is generated and shared with other users (friends), thereby improving the intelligence and the interestingness of the intelligent AI glasses and improving the user experience.
Fig. 4 is a block diagram of functional units of a smart AI glasses 400 according to an embodiment of the application. The smart AI glasses 400 include an acquisition unit 401, an extraction unit 402, a generation unit 403, and an interaction unit 404, wherein,
The acquiring unit 401 is configured to capture, by using the sensor, a facial expression image of a target object, where the target object is a user wearing the smart AI glasses, or the target object is a shooting object shot by the smart AI glasses;
the extracting unit 402 is configured to perform image extraction on the facial expression image to obtain a target feature set;
the generating unit 403 is configured to generate a target expression image in a preset format according to the target feature set;
The interaction unit 404 is configured to send, through the communication module, the target expression image to a specified device.
Optionally, in the aspect of performing image extraction on the facial expression image to obtain a target feature set, the extracting unit 402 is specifically configured to:
Extracting features of the facial expression image to obtain a first feature set;
Determining a target expression type corresponding to the first feature set;
And determining the characteristics corresponding to the target expression type according to the first characteristic set to obtain the target characteristic set.
Optionally, in the aspect that the features corresponding to the target expression type are determined according to the first feature set, so as to obtain the target feature set, the extracting unit 402 is specifically configured to:
Acquiring a feature selection rule set corresponding to the preset format to obtain a plurality of feature selection rule sets, wherein each feature selection rule set corresponds to one expression type;
determining a target feature selection rule set corresponding to the target expression type from the feature selection rule sets;
And screening the first feature set according to the target feature selection rule set to obtain the target feature set.
Optionally, in the aspect of generating the target expression image in the preset format according to the target feature set, the generating unit 403 is specifically configured to:
acquiring a reference expression template set corresponding to the preset format, wherein the reference expression template set comprises a plurality of reference expression templates, and each reference expression template corresponds to one expression type;
determining a target reference expression template corresponding to the target expression type from the reference expression template set;
and determining the target expression image according to the target feature set and the target reference expression template.
Optionally, in the aspect of determining the target expression image according to the target feature set and the target reference expression template, the generating unit 403 is specifically configured to:
acquiring a feature position area corresponding to each feature in the target feature set to obtain at least one feature position area;
Marking corresponding areas in the target reference expression template according to the at least one characteristic position area to obtain at least one area;
And filling the target feature set into a corresponding region in the at least one region in the target reference expression template to obtain the target expression image.
Optionally, in the aspect of performing feature extraction on the facial expression image to obtain a first feature set, the extracting unit 402 is specifically configured to:
Acquiring target shooting parameters of the target expression image;
acquiring target skin color parameters of the target expression image;
determining a target feature extraction algorithm corresponding to the target skin color parameters;
determining target algorithm control parameters of the target feature extraction algorithm corresponding to the target shooting parameters;
and carrying out feature extraction on the facial expression image according to the target feature extraction algorithm and the target algorithm control parameter to obtain the first feature set.
Optionally, the smart AI glasses further communicate with a wearable device of the target object, and in the aspect of determining a target algorithm control parameter of the target feature extraction algorithm corresponding to the target shooting parameter, the extraction unit 402 is specifically configured to:
acquiring the capturing moment of the facial expression image;
Acquiring, by the communication module, a target physiological state parameter of the target object acquired by the wearable device;
Determining a first algorithm control parameter of the target feature extraction algorithm corresponding to the target shooting parameter;
determining a first optimization parameter corresponding to the target physiological state parameter;
And determining the target algorithm control parameter according to the first optimization parameter and the first algorithm control parameter.
It can be seen that, the intelligent AI glasses described in the embodiments of the present application include a sensor and a communication module, the sensor captures a facial expression image of a target object, the target object is a user wearing the intelligent AI glasses, or the target object is a shooting object shot by the intelligent AI glasses, image extraction is performed on the facial expression image to obtain a target feature set, a target expression image in a preset format is generated according to the target feature set, and the communication module sends the target expression image to a designated device, so that the facial expression image of the user can be acquired, and a target expression image in a corresponding preset format is generated and shared with other users (friends), thereby improving the intelligence and the interestingness of the intelligent AI glasses and improving the user experience.
It may be understood that the functions of each program module of the smart AI glasses of the present embodiment may be specifically implemented according to the method in the foregoing method embodiment, and the specific implementation process may refer to the relevant description of the foregoing method embodiment, which is not repeated herein.
The embodiment of the present application also provides a computer storage medium storing a computer program for electronic data exchange, where the computer program causes a computer to execute some or all of the steps of any one of the methods described in the above method embodiments.
Embodiments of the present application also provide a computer program product comprising a non-transitory computer-readable storage medium storing a computer program operable to cause a computer to perform part or all of the steps of any one of the methods described in the method embodiments above. The computer program product may be a software installation package.
It should be noted that, for simplicity of description, the foregoing method embodiments are all described as a series of acts, but it should be understood by those skilled in the art that the present application is not limited by the order of acts described, as some steps may be performed in other orders or concurrently in accordance with the present application. Further, those skilled in the art will also appreciate that the embodiments described in the specification are all preferred embodiments, and that the acts and modules referred to are not necessarily required for the present application.
In the foregoing embodiments, the descriptions of the embodiments are emphasized, and for parts of one embodiment that are not described in detail, reference may be made to related descriptions of other embodiments.
In the several embodiments provided by the present application, it should be understood that the disclosed apparatus may be implemented in other manners. For example, the above-described apparatus embodiments are merely illustrative, such as the above-described division of units, merely a division of logic functions, and there may be additional manners of dividing in actual implementation, such as multiple units or components may be combined or integrated into another system, or some features may be omitted, or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, or may be in electrical or other forms.
The units described above as separate components may or may not be physically separate, and components shown as units may or may not be physical units, may be located in one place, or may be distributed over a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in the embodiments of the present application may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units may be implemented in hardware or in software functional units.
The integrated units described above, if implemented in the form of software functional units and sold or used as stand-alone products, may be stored in a computer readable memory. Based on such understanding, the technical solution of the present application may be embodied in essence or a part contributing to the prior art or all or part of the technical solution in the form of a software product stored in a memory, comprising several instructions for causing a computer device (which may be a personal computer, a server or a network device, etc.) to perform all or part of the steps of the above-mentioned method of the various embodiments of the present application. The Memory includes a U disk, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), a removable hard disk, a magnetic disk, or an optical disk, etc. which can store the program codes.
Those of ordinary skill in the art will appreciate that all or part of the steps in the various methods of the above embodiments may be implemented by a program to instruct related hardware, where the program may be stored in a computer readable Memory, and the Memory may include a flash disk, a Read-Only Memory (ROM), a random access Memory (Random Access Memory, RAM), a magnetic disk or an optical disk, etc.
The foregoing has outlined rather broadly the more detailed description of embodiments of the application, wherein the principles and embodiments of the application are explained in detail using specific examples, the above examples being provided solely to facilitate the understanding of the method and core concepts of the application; meanwhile, as those skilled in the art will have variations in the specific embodiments and application scope in accordance with the ideas of the present application, the present description should not be construed as limiting the present application in view of the above.