US20250200855A1 - Method for real-time generation of empathy expression of virtual human based on multimodal emotion recognition and artificial intelligence system using the method - Google Patents
Method for real-time generation of empathy expression of virtual human based on multimodal emotion recognition and artificial intelligence system using the method Download PDFInfo
- Publication number
- US20250200855A1 US20250200855A1 US18/657,909 US202418657909A US2025200855A1 US 20250200855 A1 US20250200855 A1 US 20250200855A1 US 202418657909 A US202418657909 A US 202418657909A US 2025200855 A1 US2025200855 A1 US 2025200855A1
- Authority
- US
- United States
- Prior art keywords
- emotion
- user
- conversation
- terminal
- multimodal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/40—3D [Three Dimensional] animation of characters, e.g. humans, animals or virtual beings
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/002—Specific input/output arrangements not covered by G06F3/01 - G06F3/16
- G06F3/005—Input arrangements through a video camera
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
- G06F40/35—Discourse or dialogue representation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/004—Artificial life, i.e. computing arrangements simulating life
- G06N3/006—Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/091—Active learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/092—Reinforcement learning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T13/00—Animation
- G06T13/20—3D [Three Dimensional] animation
- G06T13/205—3D [Three Dimensional] animation driven by audio data
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/80—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level
- G06V10/809—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of classification results, e.g. where the classifiers operate on the same input data
- G06V10/811—Fusion, i.e. combining data from various sources at the sensor level, preprocessing level, feature extraction level or classification level of classification results, e.g. where the classifiers operate on the same input data the classifiers operating on different input data, e.g. multi-modal recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04R—LOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
- H04R1/00—Details of transducers, loudspeakers or microphones
- H04R1/08—Mouthpieces; Microphones; Attachments therefor
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/0475—Generative networks
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/098—Distributed learning, e.g. federated learning
Definitions
- the disclosure relates to a method for a virtual human to empathize with a user and express the degree of empathy in an environment in which the virtual human and the user converse with each other, and a conversational artificial intelligence (AI) system using the method.
- AI conversational artificial intelligence
- conversations with artificial intelligence are largely performed through a text-based conversation system or a voice-based conversation system.
- the text-based conversation system which has mainly the form of a chatbot or a virtual assistant, is used to generate a reply or perform an operation in response to a question input by a user.
- a language model such as GPT-3, which has excellent performance in understanding and generating of a natural language, has been used in the text-based conversation system.
- the voice-based conversation system understands a voice command and generates an appropriate response by using both of a voice recognition technique and a natural language processing technique.
- These conversation systems are basically based on pre-trained models and dynamically generate responses according to an input of a user.
- Rashkin, et al. have proposed an AI model by forming 25K conversation sets for enabling an empathetic conversation with a user, and Kumano, et al. have introduced a model for predicting a facial expression and a gaze of the eyes through a Bayesian method.
- a method for real-time generation of an empathy expression of a virtual human based on multimodal emotion recognition and a conversational artificial intelligence (AI) system using the method.
- a method for real-time generation of an empathy expression of a virtual human, based on multimodal emotion recognition whereby a voice-based conversation and user information of a facial expression based on an image are recognized based on a multimodal method (or approach), and a conversational AI system using the method.
- a method for real-time generation of an empathy expression of a virtual human based on multimodal emotion recognition, whereby a text and/or voice-based conversation and an expression of a user during a conversation are detected and emotions of an expression and a conversation of a virtual human are expressed to be synchronized with an emotion of the user, and a conversational AI system using the method.
- a conversational artificial intelligence (AI) system based on real-time multimodal emotion recognition includes a model server configured to provide a machine learning-based conversational model, a terminal configured to perform a conversation with the machine learning-based conversational model through the model server, display a virtual human responding to a user during a conversation with the user, and capture a facial image of the user during the conversation, and a multimodal empathetic conversation-generation system configured to access the model server and receive a response to a question of the user from the terminal, and assess an emotion of the user from the facial image of the user and control, based on the assessed emotion, an expression of the virtual human displayed on the terminal.
- AI artificial intelligence
- the model server may include reinforcement learning from human feedback (RLHF)-based large language models (LLMs).
- RLHF human feedback
- LLMs large language models
- the terminal may include a capturing unit including a camera photographing a face of the user and a recording unit including a microphone generating an electrical voice signal of the user.
- the terminal may further include an input window configured to display text returned from the multimodal empathetic conversation-generation system to be correctable.
- the multimodal empathetic conversation-generation system or the terminal further may include a voice-text conversion unit and an emotion recognition unit configured to recognize an emotion inherent in a voice of the user by analyzing text from the voice-text conversion unit.
- the multimodal empathetic conversation-generation system may further include an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image and a composite emotion recognition unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image.
- the multimodal empathetic conversation-generation system may further include an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image and an empathetic expression generation unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image and control the expression of the virtual human displayed on the terminal.
- the multimodal empathetic conversation-generation system may further include an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image, a composite emotion recognition unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image, and an empathetic expression generation unit configured to control, based on the composite emotion, the expression of the virtual human displayed on the terminal.
- an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image
- a composite emotion recognition unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image
- an empathetic expression generation unit configured to control, based on the composite emotion, the expression of the virtual human displayed on the terminal.
- the empathetic expression generation unit may be further configured to control or manipulate the expression of the virtual human by applying the composite emotion to a rule for converting an emotion recognition result value into a variable for manipulating the expression of the virtual human.
- the model server, the terminal, and the multimodal empathetic conversation-generation system may be connected to one another through a communication network, and the communication network may be accessed by a database storing information related to a conversation between the user and the virtual human.
- a conversation generation method based on real-time multimodal emotion recognition includes providing, via a model server, a conversational model, displaying a virtual human through a display and obtaining a facial image of a user and recording a voice of the user, via a terminal used by the user to perform a voice conversation with the conversational model, via a multimodal empathetic conversation-generation system, accessing the model server and receiving a response to a question of the user from the terminal, assessing a composite emotion of the user based on an emotion inherent in the question of the user and the facial image, and based on the assessed composite emotion, controlling an expression of the virtual human displayed on the terminal.
- the terminal may be configured to transmit, to the multimodal empathetic conversation-generation system, image data obtained by photographing a face of the user and voice data obtained by recording a voice of the user, and the multimodal empathetic conversation-generation system may be configured to comprehensively recognize an emotion of the user by analyzing each of the image data and the voice data and manipulate the expression of the virtual human displayed on the terminal.
- the terminal may further be configured to provide an input window configured to display text returned from the multimodal empathetic conversation-generation system to be correctable.
- any one of the multimodal empathetic conversation-generation system or the terminal may be configured to convert the voice of the user into text, and the multimodal empathetic conversation-generation system may further be configured to recognize an emotion inherent in the voice of the user by analyzing the text.
- FIG. 2 is a block diagram of schematic structures of a multimodal empathetic conversation-generation system according to the disclosure
- the multimodal empathetic conversation-generation system 200 may provide a response to a question, etc. of the user or text or sound corresponding to a detected emotion of the user, at an appropriate point in time, and may control an expression, etc. of a virtual human displayed on the terminal 400 to correspond to or comply with an emotion of the user.
- a generative LLM for example, GPT
- GPT generative LLM
- a response to the question may be received as a sentence (“Disappointment is implied in the word of the user.”) or may be qualitatively received (“the level of arousal of the word of the user is 0.7 (between 0 and 1) and the valence of an emotion is a positive force of 0.7 (between 0 and 1)”).
- EPD End point detection
- Operation S 305 Multimodal Emotion Recognition and Empathy Expression.
- an emotion recognition result (value) analyzed based on text data and an emotion recognition result analyzed based on a facial image collected during the utterance may be combined to analyze a final emotion result of the user.
- an expression of the virtual human may be manipulated and the sentiment of the word delivered by the virtual human to the user may be manipulated to express an emotion similar to an emotional state of the user.
- the manipulation of the expression of the virtual human may be gradually performed, and a response of the virtual human may be delivered at a certain point in time at which the expression manipulation is performed.
- the two emotion recognition result values when it is assumed that the two emotion recognition result values infer the same emotional state (for example, arousal or relaxation or a positive or negative emotion) and are represented as continual values (for example, represented as values between ⁇ 1 to 1, wherein the emotion is negative as the value is closer to ⁇ 1 and positive as the value is closer to 1), the two emotion recognition result values may be averaged, or a different weight may be given to each of the two emotion recognition result values and then the two emotion recognition result values given with the different weights may be summed.
- the emotion value of the image may be multiplied by the weight of 0.7 and the emotion value of the text may be multiplied by the weight of 0.3, and then, the two values may be summed.
- An empathetic expression may be generated based on the analyzed final emotion-recognition result value.
- a rule to manipulate the emotion-recognition result value into a variable for manipulating the expression of the virtual human for example, BlendShape
- the rule may be directly set, or a model trained based on an ML based-modeling method, such as a deep learning algorithm, may be used.
- appearance features may be detected in the facial area of the user.
- the appearance features of the facial area of the user may include eyebrows, the eyes, the nose, the mouth, and the chin, and the positions of the appearance features may be detected by using 68 landmarks.
- the landmarks may be defined, for example, based on the facial action coding system (FACS) of Ekman, and the landmarks may use a good-feature-to-track (GFTT) algorithm and a facial landmark detection (FLD) algorithm.
- FACS facial action coding system
- GFTT good-feature-to-track
- FLD facial landmark detection
- FIG. 7 illustrates a process in which an expression of a virtual human is changed by mapping an image emotion detected from a facial image of a user with the virtual human.
- a target expression that has to be generated through the process of the generation of an empathetic expression is represented as a set of variables of expression manipulation of a virtual human
- an increase or a decrease of the variables of the expression manipulation of the virtual human may be gradually represented for a change from a current expression to the target expression.
- a linear function for increasing the same value in frame units may be used or a non-linear function, such as a log function or an exponential function, may be used.
- a time point at which the voice is transmitted may be arbitrarily set by the user.
- the response may be delivered after the expression manipulation process described in B. of operation S 305 is ended, or may be delivered simultaneously with the expression manipulation process.
- the expression may return to operation S 301 above.
- a method of adjusting a retaining time and a retaining intensity of the expression may be used.
- the process in which the retained expression is returned to the state of operation S 301 and is changed to a new expression may be gradually performed based on a method described in D. of operation S 305 .
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Acoustics & Sound (AREA)
- Medical Informatics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Signal Processing (AREA)
- Databases & Information Systems (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Provided are a conversational artificial intelligence (AI) system and method based on real-time multimodal emotion recognition. The system includes a model server configured to provide a machine learning-based conversational model, a terminal configured to perform a conversation with the machine learning-based conversational model through the model server, display a virtual human responding to a user during a conversation with the user, and capture a facial image of the user during the conversation, and a multimodal empathetic conversation-generation system configured to access the model server and receive a response to a question of the user from the terminal, and assess an emotion of the user from the facial image of the user and control, based on the assessed emotion, an expression of the virtual human displayed on the terminal.
Description
- This application is based on and claims priority under 35 U.S.C. § 119 to Korean Patent Application No. 10-2023-0181152, filed on Dec. 13, 2023, in the Korean Intellectual Property Office, the disclosure of which is incorporated by reference herein in its entirety.
- The disclosure relates to a method for a virtual human to empathize with a user and express the degree of empathy in an environment in which the virtual human and the user converse with each other, and a conversational artificial intelligence (AI) system using the method.
- At the present technical level, conversations with artificial intelligence (AI) are largely performed through a text-based conversation system or a voice-based conversation system.
- The text-based conversation system, which has mainly the form of a chatbot or a virtual assistant, is used to generate a reply or perform an operation in response to a question input by a user. Recently, a language model such as GPT-3, which has excellent performance in understanding and generating of a natural language, has been used in the text-based conversation system.
- The voice-based conversation system understands a voice command and generates an appropriate response by using both of a voice recognition technique and a natural language processing technique. These conversation systems are basically based on pre-trained models and dynamically generate responses according to an input of a user.
- Rashkin, et al. have proposed an AI model by forming 25K conversation sets for enabling an empathetic conversation with a user, and Kumano, et al. have introduced a model for predicting a facial expression and a gaze of the eyes through a Bayesian method.
- However, these models are based on unimodal methods and thus are not capable of delivering more complex emotions and are limited in expressing emotions.
- Provided are a method for real-time generation of an empathy expression of a virtual human based on multimodal emotion recognition, and a conversational artificial intelligence (AI) system using the method.
- Provided are a method for real-time generation of an empathy expression of a virtual human, based on multimodal emotion recognition, whereby a voice-based conversation and user information of a facial expression based on an image are recognized based on a multimodal method (or approach), and a conversational AI system using the method.
- Provided are a method for real-time generation of an empathy expression of a virtual human, based on multimodal emotion recognition, whereby a text and/or voice-based conversation and an expression of a user during a conversation are detected and emotions of an expression and a conversation of a virtual human are expressed to be synchronized with an emotion of the user, and a conversational AI system using the method.
- Additional aspects will be set forth in part in the description which follows and, in part, will be apparent from the description, or may be learned by practice of the presented embodiments of the disclosure.
- According to an aspect of the disclosure, a conversational artificial intelligence (AI) system based on real-time multimodal emotion recognition includes a model server configured to provide a machine learning-based conversational model, a terminal configured to perform a conversation with the machine learning-based conversational model through the model server, display a virtual human responding to a user during a conversation with the user, and capture a facial image of the user during the conversation, and a multimodal empathetic conversation-generation system configured to access the model server and receive a response to a question of the user from the terminal, and assess an emotion of the user from the facial image of the user and control, based on the assessed emotion, an expression of the virtual human displayed on the terminal.
- According to one or more embodiments, the model server may include reinforcement learning from human feedback (RLHF)-based large language models (LLMs).
- According to one or more embodiments, the terminal may include a capturing unit including a camera photographing a face of the user and a recording unit including a microphone generating an electrical voice signal of the user.
- According to one or more embodiments, the terminal may further include an input window configured to display text returned from the multimodal empathetic conversation-generation system to be correctable.
- According to one or more embodiments, the multimodal empathetic conversation-generation system or the terminal further may include a voice-text conversion unit and an emotion recognition unit configured to recognize an emotion inherent in a voice of the user by analyzing text from the voice-text conversion unit.
- According to one or more embodiments, the multimodal empathetic conversation-generation system may further include an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image and a composite emotion recognition unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image.
- According to one or more embodiments, the multimodal empathetic conversation-generation system may further include an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image and an empathetic expression generation unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image and control the expression of the virtual human displayed on the terminal.
- According to one or more embodiments, the multimodal empathetic conversation-generation system may further include an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image, a composite emotion recognition unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image, and an empathetic expression generation unit configured to control, based on the composite emotion, the expression of the virtual human displayed on the terminal.
- According to one or more embodiments, the empathetic expression generation unit may be further configured to control or manipulate the expression of the virtual human by applying the composite emotion to a rule for converting an emotion recognition result value into a variable for manipulating the expression of the virtual human.
- According to one or more embodiments, the model server, the terminal, and the multimodal empathetic conversation-generation system may be connected to one another through a communication network, and the communication network may be accessed by a database storing information related to a conversation between the user and the virtual human.
- According to another aspect of the disclosure, a conversation generation method based on real-time multimodal emotion recognition includes providing, via a model server, a conversational model, displaying a virtual human through a display and obtaining a facial image of a user and recording a voice of the user, via a terminal used by the user to perform a voice conversation with the conversational model, via a multimodal empathetic conversation-generation system, accessing the model server and receiving a response to a question of the user from the terminal, assessing a composite emotion of the user based on an emotion inherent in the question of the user and the facial image, and based on the assessed composite emotion, controlling an expression of the virtual human displayed on the terminal.
- According to one or more embodiments, the terminal may be configured to transmit, to the multimodal empathetic conversation-generation system, image data obtained by photographing a face of the user and voice data obtained by recording a voice of the user, and the multimodal empathetic conversation-generation system may be configured to comprehensively recognize an emotion of the user by analyzing each of the image data and the voice data and manipulate the expression of the virtual human displayed on the terminal.
- According to one or more embodiments, the terminal may further be configured to provide an input window configured to display text returned from the multimodal empathetic conversation-generation system to be correctable.
- According to one or more embodiments, any one of the multimodal empathetic conversation-generation system or the terminal may be configured to convert the voice of the user into text, and the multimodal empathetic conversation-generation system may further be configured to recognize an emotion inherent in the voice of the user by analyzing the text.
- According to one or more embodiments, the multimodal empathetic conversation-generation system may further be configured to recognize an emotion of the user in the image by analyzing the facial image from the terminal, recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the facial image, and control the expression of the virtual human displayed on the terminal.
- According to one or more embodiments, the multimodal empathetic conversation-generation unit may further be configured to control or manipulate the expression of the virtual human by applying the composite emotion to a rule for converting an emotion recognition result value into a variable for manipulating the expression of the virtual human.
- The above and other aspects, features, and advantages of certain embodiments of the disclosure will be more apparent from the following description taken in conjunction with the accompanying drawings, in which:
-
FIG. 1 schematically illustrates a conversational artificial intelligence (AI) system using a method for real-time generation of an empathy expression of a virtual human based on multimodal emotion recognition, according to the disclosure; -
FIG. 2 is a block diagram of schematic structures of a multimodal empathetic conversation-generation system according to the disclosure; -
FIG. 3 is a schematic flowchart of an image-based empathy expression method according to the disclosure; -
FIG. 4 illustrates an arrangement of landmarks of a face, defined by a facial action coding system (FACS); -
FIG. 5 shows action units (AUs) defining motions of facial muscles for determining a change in a facial expression; -
FIG. 6 illustrates an expressionless state of a virtual human in (A) and a controlled state of the virtual human in a slightly smiling state with both corners of the mouth turned up in (B); and -
FIG. 7 illustrates manipulation of an expression of another virtual human empathizing with an emotion of a user, according to the disclosure. - Reference will now be made in detail to embodiments, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to like elements throughout. In this regard, the present embodiments may have different forms and should not be construed as being limited to the descriptions set forth herein. Accordingly, the embodiments are merely described below, by referring to the figures, to explain aspects of the present description. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. Expressions such as “at least one of,” when preceding a list of elements, modify the entire list of elements and do not modify the individual elements of the list.
- Hereinafter, embodiments will be described in detail with reference to the accompanying drawings. However, embodiments may be modified in various different forms and the scope of the disclosure shall not be interpreted to be limited to the embodiments described hereinafter. It is desirable to understand that the embodiments are described to relatively more fully explain the disclosure to one of ordinary skill in the art. The same reference numerals denote the same elements throughout. Furthermore, various elements and areas in the drawings are schematically illustrated. Thus, the disclosure is not limited by relative sizes or distances shown in the accompanying drawings.
- The terms, such as “first,” “second,” etc., may be used to describe various components, but the components shall not be limited by the terms. These terms are used only for distinguishing one element from another element. For example, without deviating from the scope of the claims of the disclosure, a first element may be referred to as a second element, and in contrast, the second element may be referred to as the first element.
- Terms used in the present application are used only for describing particular embodiments of the disclosure and are not intended to limit the disclosure. A singular expression may include a plural expression, unless an apparently different meaning is indicated in the context. With respect to the present application, it will be further understood that the expressions “comprises” and “comprising” used herein specify the presence of stated features, integers, steps, operations, members, components, and/or groups thereof, but do not preclude the presence or addition of one or more other features, integers, operations, members, components, and/or groups thereof.
- Unless defined otherwise, all of the terms used herein, including technical terms and scientific terms, have the same meaning as the meaning commonly understood by one of ordinary skill in the art. Also, the terms commonly used and having the meanings as defined in dictionaries shall be understood to have consistent meanings with the corresponding terms in the context of relevant arts, and unless explicitly defined so herein, the meanings of the terms shall not be understood to be excessively formal.
- When a certain embodiment may be implemented differently, a specific process order may be performed differently from the described order. For example, two consecutively described processes may be performed substantially at the same time or performed in an order opposite to the described order.
- According to an embodiment, a virtual human displayed on a terminal, for example, a personal computer (PC) or a portable terminal, performs, based on a large language model (LLM), a smooth and context-based empathetic conversation with a user, recognizes an emotion of the user though comprehensive analysis (multimodal emotion recognition) of a facial expression (an image) of the user and conversation content (text) spoken by the user during the conversation, and based on a recognition result value, shows an expression of an emotional state similar to the emotion of the user and performs an emotional conversation in compliance with the emotional state of the user, thereby empathizing with the user. Also, according to a method according to the disclosure, even in a situation where a user is not speaking, empathy may be expressed by performing emotion recognition based on a facial expression of the user.
- Hereinafter, according to one or more embodiments, a method and system for real-time generation of an empathy expression of a virtual human, based on multimodal emotion recognition, will be described in detail.
-
FIG. 1 illustrates a schematic structure of a conversational artificial intelligence (AI)system 1, that is, an AI-based conversational system using a method for real-time generation of an empathy expression of a virtual human based on multimodal emotion recognition, according to the disclosure. - Referring to
FIG. 1 , theconversational AI system 1 according to the disclosure includes aconversational model 100, a multimodal empathetic conversation-generation system 200, a terminal 400, and adatabase 500 configured to communicate with one another through acommunication network 300, such as the Internet, etc. - The
communication network 300 may include most of existing wired and wireless communication systems, such as the Internet, a wide area network (WAN), a wired local area network (LAN), a wireless LAN, a 3rd generation mobile network (3G), a 4th generation mobile network (4G), a 5th generation mobile network (5G), long term evolution (LTE), etc., and may also use various communication protocols. The technical scope of thecommunication network 300 according to the disclosure is not limited by predetermined systems and predetermined protocols. - The
database 500 may be accessed by the multimodal empathetic conversation-generation system 200 and may be accessed by a local area network (LAN) in a local directory or may be arranged in the multimodal empathetic conversation-generation system 200. Thedatabase 500 may store personal information, access information, content of a conversation of a user, etc., so that based on the stored content of the conversation, the user may continue the subsequent conversation with reference to the content of the conversation. For example, if the user said in yesterday's conversation that he/she was sick, the content may be stored in thedatabase 500, and in the next day's conversation, he/she could be asked, for example, how he/she feels. - The
conversational model 100 at the highest layer of theconversational AI system 1 according to the disclosure may be configured to understand a natural language sentence and extract the meaning thereof, based on a natural language processing technique which is input by the user through the terminal 400, and may be configured to generate a response based on knowledge or data that is previously learnt. Theconversational model 100 may use, for example, a generative pre-trained transformer (GPT) model configured to perform language understanding and generation by being previously trained with a large volume of text data, or a model in a form of a new outcome of transfer learning, in particular, a new model which is capable of effectively processing the Korean language or other languages. - In detail, the
conversational model 100 of a model server may include a model, for example, ChatGPT, which is obtained by applying the reinforcement learning from human feedback (RLHF) technique to LLMs trained by using a large volume of data. The LLM may be a type of natural language processing model trained based on a large volume of text data and may be designed to generate relatively more natural and consistent text by receiving original text or sentences as an input by using a different natural language generation model in the model. The RLHF technique, one of methods of training a machine learning model, is used for improving a model by using human feedback and is related to reinforcement learning. - By using the terminal 400, the user may subsequently access the multimodal empathetic conversation-
generation system 200 through the Web. Through the terminal 400, a question of the user may be input and a response thereto may be generated. Also, the terminal 400 may display a virtual human performing a conversation with the user. The virtual human may show various emotional expressions during the conversation with the user. - The question of the user may basically include a text input, and according to another embodiment, a voice signal may be received and may be changed into text. The changing of the voice signal into the text may be performed directly by the terminal 400, and according to another embodiment, may be performed by a general algorithm in the multimodal empathetic conversation-
generation system 200 accessed by theterminal 400. The terminal 400 may extract a facial area of the user from a facial image obtained through a photographing unit having a video camera, for example, a webcam, which is a video-obtaining device, and may analyze an emotion of the user indicated in the extracted facial area. Based on an emotion-analysis result value, the terminal 400 may manipulate an expression of the virtual human to represent an emotional state similar to an emotional state of the user. - The terminal 400 may indicate all types of electronic devices including a desktop PC, a portable PC, a notebook computer, an Android terminal, an iOS terminal, etc., which may access the multimodal empathetic conversation-
generation system 200 and perform a conversation. - While a conversation with the user is being maintained through the terminal 400, the multimodal empathetic conversation-
generation system 200 may assess an emotion of the user by analyzing a voice and an image transmitted from the terminal 400 and continue the conversation. Simultaneously, the multimodal empathetic conversation-generation system 200 may change the expression of the virtual human according to a result of the assessment of the emotion through the analysis of the image and the voice of the user. According to the present embodiment, it is described that the analysis of the image is generated in the multimodal empathetic conversation-generation system 200. However, according to another embodiment, the analysis of the image may be generated in theterminal 400. This aspect may denote that analysis of an emotional state through an image may be executed in any component of theAI system 1. -
FIG. 2 illustrates a process in which image data and voice data are processed by the multimodal empathetic conversation-generation system 200 and the terminal 400 and shows a relationship between the multimodal empathetic conversation-generation system 200 and the terminal 400, which is realized in the multimodal empathetic conversation-generation system 200. - The terminal 400 may have a general structure and may include a
display 401, atext input unit 402, avoice recorder 403, avoice reproducer 404, and a photographingunit 405 as hardware and application-based internal devices. - The
display 401 may display a dynamic virtual human and may display content of an exchanged conversation. According to an embodiment, thedisplay 401 may be in the form of a graphics user interface (GUI) and may provide thetext input unit 402 and one or more instruction buttons for start/end operations, etc. - The
text input unit 402 may receive user input text. However, thetext input unit 402 may also show voice data converted into text to the user so that the user may correct the converted text, before thetext input unit 402 transmits the converted text to an LLM. - The
voice recorder 403 may record a voice of the user and the recorded voice data may be digitalized and transmitted to the multimodal empathetic conversation-generation system 200. Thevoice reproducer 404 may provide a voice signal received form the multimodal empathetic conversation-generation system 200 to the user. The photographingunit 405 having a camera and an image recorder may photograph a face of the user and the captured image data may be transmitted to the multimodal empathetic conversation-generation system 200. - The terminal 400 may indicate all types of electronic devices including a desktop PC, a portable PC, a notebook computer, an Android terminal, an iOS terminal, etc., which may access the multimodal empathetic conversation-
generation system 200 and perform a conversation. - While a conversation with the user is being performed by the terminal 400, the multimodal empathetic conversation-
generation system 200 may provide a response to a question, etc. of the user or text or sound corresponding to a detected emotion of the user, at an appropriate point in time, and may control an expression, etc. of a virtual human displayed on the terminal 400 to correspond to or comply with an emotion of the user. - The multimodal empathetic conversation-
generation system 200 may transmit the sound of a conversational question or response generated by the text to speech (TTS) technique, wherein the sound may be embodied in various speaking styles and voices. Also, an emotion-expression generation unit 206 may transmit virtual human control data for controlling an expression of the virtual human displayed on the terminal 400, based on a question and an expression of the user. - A voice trigger recognition unit may determine whether or not a voice input from the
voice recorder 403 of theuser terminal 400 or a text input from theinput unit 402 is started, that is, whether or not a voice trigger is started, and when the voice trigger is recognized, may receive a voice signal and a text signal and transmit resultant text to an emotion recognition unit. A voice-text conversion unit 201 of the voice trigger recognition unit may convert the voice signal from the terminal 400 into text. - The emotion recognition unit may include an image-
emotion recognition unit 202, a text-emotion recognition unit 203, and a composite-emotion recognition unit 204. The image-emotion recognition unit 202 may recognize an emotion of a user from an image signal from the terminal 400, and the text-emotion recognition unit 203 may recognize an emotion inherent in text. - Results from the image-
emotion recognition unit 202 and the text-emotion recognition unit 203 may be transmitted to the composite-emotion recognition unit 204. The composite-emotion recognition unit 204 may recognize an emotion of the user by combining the emotion obtained from the image and the emotion obtained from the text. - The emotion-
expression generation unit 206 described above may receive result values from the image-emotion recognition unit 202 and the composite-emotion recognition unit 204 and may generate a signal for controlling an expression of a virtual human displayed on the terminal 4000 and transmit the signal to the terminal 400. - A response generation unit may perform a conversation between the user and the
conversational model 100. That is, the text that is converted from the voice data from the user or the text corrected by the user after being converted may be transmitted to theconversational model 100 and the text retrieved therefrom may be transmitted to theterminal 400 of the user. - The voice data from the terminal 400 may be input to the voice-
text conversion unit 201 of the multimodal empathetic conversation-generation system 200 and may be converted into text, the converted text may be transmitted to the text-emotion recognition unit 203, and the text-emotion recognition unit 203 may assess an emotion of the user in the text. Also, the image data from the terminal 400 may be transmitted to the image-emotion recognition unit 202 of the emotion recognition unit, and an emotion inherent in the image may be assessed. - The image emotion information and the text emotion information obtained from the image-
emotion recognition unit 202 and the text-emotion recognition unit 203, respectively, may be transmitted to the composite-emotion recognition unit 204 configured to determine a composite emotion of the user. A finally assessed emotion obtained from the composite-emotion recognition unit 204 may be transmitted to the emotion-expression generation unit 206, so that a virtual human having an expression corresponding to the finally assessed emotion may be generated or virtual human expression control data therefor may be formed. - The text-
voice conversion unit 205 may transmit, to the terminal 400, a response of an LLM, etc. in response to the user voice input from the voice-text conversion unit 201. - Hereinafter, by referring to
FIG. 2 above andFIG. 3 that is a flowchart of an overall conversation process, a specific method, performed by the system, of generating an empathetic conversation, will be described in detail. - Described below are real-time orders of the technique according to the disclosure in a conversational environment with a user, the real-time orders being illustrated in
FIG. 3 . In the description of the orders, specific techniques used in each order are included. - Operation S301 is performed in a state in which a conversation has not yet been performed. In operation S301, a facial area of a user may be extracted through a video photographing device, such as a webcam, etc., and an emotion of the user may be analyzed based on the extracted facial area. An emotional state similar as an emotional state of the user may be represented by manipulating, based on a result value of the analyzed emotion, an expression of a virtual human displayed on a terminal.
- To detect the facial area from a facial image, a face detection technique is used. As the face detection technique, a model trained based on a machine-learning (ML)-based modeling method, such as a computer vision algorithm such as Viola Jones or a deep learning algorithm, may be used.
- For the inference of the emotion from the extracted facial area, a facial expression recognition technique may be used. For example, a rule-based model may be used by extracting features using a computer-vision algorithm, such as a histogram of gradient (HOG) or a local binary pattern (LBP), or a model trained based on an ML-based modeling method, such as a deep learning algorithm, may be used.
- Data for controlling the expression of the virtual human may be generated based on the emotion recognition result, and the expression of the virtual human displayed on the terminal may be controlled or manipulated, which may be understood in detail with reference to the description of operation S305.
- Operation S301 above may be repetitively performed, unless there is utterance from the user and unless there is forcible control by the user.
- When a voice trigger of the user, that is, utterance, is detected (S302), a sound input device (a microphone) may be activated to receive a voice signal of the user, and the voice signal of the user may be recorded through the
voice recorder 403. The voice of the user may be converted into text by using an automatic speech recognition (ASR) technique. Before the voice of the user converted into the text is transmitted to an LLM, the converted text may be shown to the user so that the user may correct the text by using an input unit of the terminal to increase the degree of completion of a sentence. This correction process may be optional and may thus be omitted. The text of the completed sentence may be used to recognize an emotional state of the user by using a sentimental analysis method. - An automatic voice trigger detection technique that is triggered when a predetermined keyword is detected, etc. may be used. Representative voice trigger words may include “Hi Google,” “Hello Bixby,” and “Aria (a Nugu speaker of the SK Telecom).” When a voice trigger is triggered, a microphone of the terminal may be open, and all signals of subsequent utterance of the user may be input and processed. Alternatively, the voice trigger may not be used, and the microphone may be configured to be open when a user presses on a predetermined button. A code for processing the voice trigger may not be stored in a cloud. Rather, the code may be stored in the system in which the virtual human is displayed, for example, the terminal of the user.
- An exclusive ASR technique may be used, or a commercialized service, such as Naver CLOVA, etc. may be used. Whether the ASR technique is used or the Naver CLOVA is used, communication (input and output operations) with an ASR AI decoder may be necessary. Input information may be a voice trigger signal (analog) of the user, and output information may be a text string obtained by digitalizing the analog voice signal. The ASR model may be stored in a local directory, a cloud or a conversation-generation system. In the case of the latter, a communication module may have to be provided in the local terminal.
- A word in which a sentiment of a sentence is embedded may be detected from input text, and the word in which the sentiment is embedded may be used to estimate in which emotional state the corresponding sentence is uttered. (For example, “happy”: a happy sentiment, “joyful”: an excited sentiment, “irritated”: a vexed sentiment, “sad”: a mournful sentiment, etc.) As another method, an AI model may be trained by labelling a sentiment for each sentence, and the trained model may be used. (For example, labelling “I have received a lot of birthday presents!” with “happy” or continual values). A method of estimating which text data corresponds to which emotion by using an AI model trained by gathering the data above may also be used. As another method, a generative LLM (for example, GPT) may be asked which emotion is implied in conversation content uttered by the user. A response to the question may be received as a sentence (“Disappointment is implied in the word of the user.”) or may be qualitatively received (“the level of arousal of the word of the user is 0.7 (between 0 and 1) and the valence of an emotion is a positive force of 0.7 (between 0 and 1)”). These methods may be used to estimate which emotional state is indicated by the content of the word of the user.
- After this process is ended, whether or not utterance is ended may be determined in operation S304, wherein end is Yes, and the process may subsequently proceed to operation S305, and when the utterance is not ended, the process may return to operation S303 and operation S303 may be performed. End point detection (EPD) of the utterance of the user may be additionally received by the user, for example, through a button input, or an automatic utterance end detection technique, etc. may be used.
- When the user utterance is ended in operation S304, an emotion recognition result (value) analyzed based on text data and an emotion recognition result analyzed based on a facial image collected during the utterance may be combined to analyze a final emotion result of the user. Based on the analyzed final emotion result value, an expression of the virtual human may be manipulated and the sentiment of the word delivered by the virtual human to the user may be manipulated to express an emotion similar to an emotional state of the user. Here, the manipulation of the expression of the virtual human may be gradually performed, and a response of the virtual human may be delivered at a certain point in time at which the expression manipulation is performed.
- Combination of a text-based emotion recognition result and an image-based emotion recognition result: when it is assumed that the two emotion recognition result values infer the same emotional state (for example, arousal or relaxation or a positive or negative emotion) and are represented as continual values (for example, represented as values between −1 to 1, wherein the emotion is negative as the value is closer to −1 and positive as the value is closer to 1), the two emotion recognition result values may be averaged, or a different weight may be given to each of the two emotion recognition result values and then the two emotion recognition result values given with the different weights may be summed. (For example, the emotion value of the image may be multiplied by the weight of 0.7 and the emotion value of the text may be multiplied by the weight of 0.3, and then, the two values may be summed.)
- An empathetic expression may be generated based on the analyzed final emotion-recognition result value. To this end, a rule to manipulate the emotion-recognition result value into a variable for manipulating the expression of the virtual human, for example, BlendShape, may be defined. The rule may be directly set, or a model trained based on an ML based-modeling method, such as a deep learning algorithm, may be used.
- An example of the rule may include a method, by which the variable for manipulating the expression of the virtual human is mapped with an action unit (AU) of Ekman and the analyzed emotion result value is linked to an adjacent basic emotion expression (Ekman), which will be described in detail with reference to (1) to (4) below. Here, the BlendShape described hereinafter may be mainly used in three-dimensional (3D) character animation and modeling, and in particular, may be an example variable used to generate various facial expressions.
- (1) Mapping of the AU with a BlendShape variable (For example, AU6 of “both corners of the mouth turned up” may be mapped with, for example, a variable of “Mouth_Smile” from among the BlendShape (AU) variables.
FIG. 6 illustrates an expressionless state of a virtual human in (A) and a controlled state of the virtual human in a slightly smiling Mouth_Smile state with both corners of the mouth turned up in (B). - To this end, appearance features may be detected in the facial area of the user. In an AU detection process, the appearance features of the facial area of the user may include eyebrows, the eyes, the nose, the mouth, and the chin, and the positions of the appearance features may be detected by using 68 landmarks. The landmarks may be defined, for example, based on the facial action coding system (FACS) of Ekman, and the landmarks may use a good-feature-to-track (GFTT) algorithm and a facial landmark detection (FLD) algorithm.
-
FIG. 4 illustrates an arrangement of landmarks defined in the FACS,FIG. 5 shows a list of the AUs (BlendShapes) defining the motion of facial muscles for determining a change in a facial expression, and Table 1 below describes the AUs with respect toFIG. 4 and a landmark included in each AU. -
TABLE 1 AU Description Facial Landmarks AU 1 Inner Brow Raise 18, 19, 20, 21 (Left) | 22, 23, 24, 25(Right) AU 2 Outer Brow Raise 17, 18, 19 (Left) | 24, 25, 26 (Right) AU 4 Brow Lowerer 17, 18, 25, 26 AU 5 Upper Lid Raise 36, 37, 38, 39 (Left) | 42, 43, 44, 45 (Right) AU 6 Check Raise 41, 2, 31 (Left) | 46, 14, 35(Right) AU 7 Lids Tight 36, 41, 39, 38 (Left) | 42, 46, 45, 43 (Right) AU 9 Nose Wrinkle 30, 31, 33 (Left) | 30, 33, 35 (Right) AU 12 Lip Corner Puller 3, 48, 31 (Left) | 13, 54, 35 (Right) AU 15 Lip Corner Depressor 4, 6, 48 (Left) | 12, 54, 10 (Right) AU 16 Lower Lip Depress 59, 6, 7, 8, 9, 10, 55, 56, 57, 58 AU 20 Lip Stretch 3, 4, 5, 48 (Left) | 11, 12, 13, 54 (Right) AU 23 Lip Funneler 50, 58, 56, 52 AU 26 Jaw Drop 6, 8, 10, 57 - (2) In order to control or change the expression of the virtual human, the AUs and the basic emotions may be mapped with each other, for example, based on the AUs of basic emotional expressions (happiness, sadness, anger, fear, distaste, surprise) of Ekman. For example, according to the definition of Ekman, happiness=AU6+AU12.
- The virtual human may be generated by using, for example, REALLUSION Character Creator3.
- (3) After an extracted emotion recognition result value is mapped with an adjacent basic emotion, a weight is assigned thereto. For example, when a final emotion recognition result value corresponds to the degree of arousal of 0.3 and the positive valence of 0.7, the degree of arousal and the positive valence are classified into predetermined ranges (for example, with the degree of arousal >0 and the positive valence >0, the emotion is classified as “a positive and aroused emotion”), and the corresponding range and a basic emotion are mapped with each other (“positive and aroused”=“happiness”+“surprised,” and “negative and aroused”=“fear”+“distaste”+“anger”). Then, a weight is assigned to each basic emotion (because the positive valence is higher than the degree of arousal, a weight of 0.7 is assigned to the AU of happiness, and a weight of 0.3 is assigned to the AU of surprise). Alternatively, the degree of arousal and the positive valence are represented as points on a two-dimensional plane and are converted into a polar coordinate system, and the weight is determined through a relative rate as to which of 0 degrees and 90 degrees a theta value is close to. Alternatively, an additional image module for classifying the basic emotion ranges may be used, and while the user is speaking, a relative ratio of a recognition frequency value or a probability value of “happiness” and “surprise” may be used.
- (4) Setting an Expressional Intensity of an Expression For example, in the process (iii.) above, an R value calculated when the degree of arousal and the positive valence are converted into the polar coordinate system may be used to set the expressional intensity.
FIG. 7 illustrates a process in which an expression of a virtual human is changed by mapping an image emotion detected from a facial image of a user with the virtual human. - When a target expression that has to be generated through the process of the generation of an empathetic expression, is represented as a set of variables of expression manipulation of a virtual human, an increase or a decrease of the variables of the expression manipulation of the virtual human may be gradually represented for a change from a current expression to the target expression. A linear function for increasing the same value in frame units may be used or a non-linear function, such as a log function or an exponential function, may be used.
- Response text of the virtual human generated based on the LLM may be converted to a voice of a human being by using the TTS technique, and the converted voice data may be delivered to the user through a sound output device (for example, a speaker). When the response is generated through the LLM, the LLM may be instructed to generate the response having the content in compliance with an emotion of the user by reflecting the multimodal recognition result with respect to the emotion of the user. For example, a response with respect to the word of the user “hello” as a prompt sentence of the LLM may be generated, wherein an emotion of the response should be the arousal of A and the valence of B (A and B are the arousal and valence states of the user, estimated in the multimodal emotion recognition in B. of operation S305 above). The content of the word generated by the LLM may be converted into the voice through the TTS technique. Here, the tone and the manner of the voice of the virtual human may have to have the same arousal and valence as the content of the word generated by the LLM. When a gap occurs between the content and the arousal and the valence, a feeling of distance may occur between the content of the word and the voice expression, and thus, a hearer may feel it unnatural. For example, when the virtual human says “hello!” the tone and the manner must not be too dark or slow.
- A time point at which the voice is transmitted may be arbitrarily set by the user. For example, the response may be delivered after the expression manipulation process described in B. of operation S305 is ended, or may be delivered simultaneously with the expression manipulation process.
- After the expression generated in operation S305 above is retained, the expression may return to operation S301 above. A method of adjusting a retaining time and a retaining intensity of the expression may be used. The process in which the retained expression is returned to the state of operation S301 and is changed to a new expression may be gradually performed based on a method described in D. of operation S305.
- To summarize the embodiment described in detail above, according to the system, the virtual human empathizes with the user and expresses a degree of empathy in the environment in which the user converses with the virtual human through the terminal. During the conversation, the virtual human may perform a natural and context-based empathetic conversation with the user. During the conversation, the virtual human may comprehensively analyze the facial expression of the user and the utterance of the user based on the multimodal emotion recognition and may express an emotional state similar to an emotional state of the user and perform the conversation in compliance with the emotional state of the user.
- It should be understood that embodiments described herein should be considered in a descriptive sense only and not for purposes of limitation. Descriptions of features or aspects within each embodiment should typically be considered as available for other similar features or aspects in other embodiments. While one or more embodiments have been described with reference to the figures, it will be understood by one of ordinary skill in the art that various changes in form and details may be made therein without departing from the spirit and scope as defined by the following claims.
Claims (19)
1. A conversational artificial intelligence (AI) system based on real-time multimodal emotion recognition, the conversational artificial intelligence (AI) system comprising:
a model server configured to provide a machine learning-based conversational model;
a terminal configured to perform a conversation with the machine learning-based conversational model through the model server, display a virtual human responding to a user during a conversation with the user, and capture a facial image of the user during the conversation; and
a multimodal empathetic conversation-generation system configured to access the model server and receive a response to a question of the user from the terminal, and assess an emotion of the user from the facial image of the user and control, based on the assessed emotion, an expression of the virtual human displayed on the terminal.
2. The conversational AI system of claim 1 , wherein the model server comprises reinforcement learning from human feedback (RLHF)-based large language models (LLMs).
3. The conversational AI system of claim 1 , wherein the terminal comprises:
a capturing unit including a camera photographing a face of the user; and
a recording unit including a microphone generating an electrical voice signal of the user.
4. The conversational AI system of claim 3 , wherein the terminal further comprises an input window configured to display text returned from the multimodal empathetic conversation-generation system to be correctable.
5. The conversational AI system of claim 4 , wherein the multimodal empathetic conversation-generation system or the terminal further comprises a voice-text conversion unit and an emotion recognition unit configured to recognize an emotion inherent in a voice of the user by analyzing text from the voice-text conversion unit.
6. The conversational AI system of claim 5 , wherein the multimodal empathetic conversation-generation system further comprises:
an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image; and
a composite emotion recognition unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image.
7. The conversational AI system of claim 5 , wherein the multimodal empathetic conversation-generation system further comprises:
an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image; and
an empathetic expression generation unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image and control the expression of the virtual human displayed on the terminal.
8. The conversational AI system of claim 5 , wherein the multimodal empathetic conversation-generation system further comprises:
an image emotion recognition unit configured to analyze the facial image from the terminal to recognize an emotion of the user in the image;
a composite emotion recognition unit configured to recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the image; and
an empathetic expression generation unit configured to control, based on the composite emotion, the expression of the virtual human displayed on the terminal.
9. The conversational AI system of claim 7 , wherein the empathetic expression generation unit is further configured to control or manipulate the expression of the virtual human by applying the composite emotion to a rule for converting an emotion recognition result value into a variable for manipulating the expression of the virtual human.
10. The conversational AI system of claim 1 , wherein the model server, the terminal, and the multimodal empathetic conversation-generation system are connected to one another through a communication network, and
the communication network is accessed by a database storing information related to a conversation between the user and the virtual human.
11. A conversation generation method based on real-time multimodal emotion recognition, the conversation generation method comprising:
providing, via a model server, a conversational model;
displaying a virtual human through a display and obtaining a facial image of a user and recording a voice of the user via a terminal used by the user to perform a voice conversation with the conversational model;
via a multimodal empathetic conversation-generation system, accessing the model server and receiving a response to a question of the user from the terminal, assessing a composite emotion of the user based on an emotion inherent in the question of the user and the facial image, and based on the assessed composite emotion, controlling an expression of the virtual human displayed on the terminal.
12. The conversation generation method of claim 11 , wherein the model server comprises reinforcement learning from human feedback (RLHF)-based large language models (LLMs).
13. The conversation generation method of claim 12 , wherein the terminal is configured to transmit, to the multimodal empathetic conversation-generation system, image data obtained by photographing a face of the user and voice data obtained by recording a voice of the user, and
the multimodal empathetic conversation-generation system is configured to comprehensively recognize an emotion of the user by analyzing each of the image data and the voice data and manipulate the expression of the virtual human displayed on the terminal.
14. The conversation generation method of claim 13 , wherein the terminal is further configured to provide an input window configured to display text returned from the multimodal empathetic conversation-generation system to be correctable.
15. The conversation generation method of claim 11 , wherein the terminal is configured to transmit, to the multimodal empathetic conversation-generation system, image data obtained by photographing a face of the user and voice data obtained by recording a voice of the user, and
the multimodal empathetic conversation-generation system is configured to comprehensively recognize an emotion of the user by analyzing each of the image data and the voice data and manipulate the expression of the virtual human displayed on the terminal.
16. The conversation generation method of 15 , wherein the terminal is further configured to provide an input window configured to display text returned from the multimodal empathetic conversation-generation system to be correctable.
17. The conversation generation method of claim 11 , wherein any one of the multimodal empathetic conversation-generation system or the terminal is configured to convert the voice of the user into text, and the multimodal empathetic conversation-generation system is further configured to recognize an emotion inherent in the voice of the user by analyzing the text.
18. The conversation generation method of claim 17 , wherein the multimodal empathetic conversation-generation system is further configured to recognize an emotion of the user in the image by analyzing the facial image from the terminal, recognize a composite emotion by combining the emotion obtained from the text with the emotion obtained from the facial image, and control the expression of the virtual human displayed on the terminal.
19. The conversation generation method of claim 18 , wherein the multimodal empathetic conversation-generation unit is further configured to control or manipulate the expression of the virtual human by applying the composite emotion to a rule for converting an emotion recognition result value into a variable for manipulating the expression of the virtual human.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR10-2023-0181152 | 2023-12-13 | ||
| KR1020230181152A KR20250090933A (en) | 2023-12-13 | 2023-12-13 | Method for real-time empathy expression of virtual human based on multimodal emotion recognition and Artificial Intelligence using the Method |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20250200855A1 true US20250200855A1 (en) | 2025-06-19 |
Family
ID=96022265
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US18/657,909 Pending US20250200855A1 (en) | 2023-12-13 | 2024-05-08 | Method for real-time generation of empathy expression of virtual human based on multimodal emotion recognition and artificial intelligence system using the method |
Country Status (2)
| Country | Link |
|---|---|
| US (1) | US20250200855A1 (en) |
| KR (1) | KR20250090933A (en) |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20250238991A1 (en) * | 2024-01-18 | 2025-07-24 | Purdue Research Foundation | System and method for authoring context-aware augmented reality instruction through generative artificial intelligence |
| CN120585332A (en) * | 2025-08-06 | 2025-09-05 | 湖南高阳通联信息技术有限公司 | An AI psychological counseling system with emotional intelligence and privacy computing dual engines |
| CN120951293A (en) * | 2025-10-15 | 2025-11-14 | 联通在线信息科技有限公司 | User emotion recognition method and device, electronic equipment and storage medium |
Family Cites Families (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| KR102507811B1 (en) | 2022-10-28 | 2023-03-10 | 주식회사 닥터송 | User psychology management system through chatbot service based on empathic psychology |
| KR102507809B1 (en) | 2022-10-28 | 2023-03-10 | 주식회사 닥터송 | Artificial intelligence dialogue system for psychotherapy through consensus formation |
-
2023
- 2023-12-13 KR KR1020230181152A patent/KR20250090933A/en active Pending
-
2024
- 2024-05-08 US US18/657,909 patent/US20250200855A1/en active Pending
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20250238991A1 (en) * | 2024-01-18 | 2025-07-24 | Purdue Research Foundation | System and method for authoring context-aware augmented reality instruction through generative artificial intelligence |
| CN120585332A (en) * | 2025-08-06 | 2025-09-05 | 湖南高阳通联信息技术有限公司 | An AI psychological counseling system with emotional intelligence and privacy computing dual engines |
| CN120951293A (en) * | 2025-10-15 | 2025-11-14 | 联通在线信息科技有限公司 | User emotion recognition method and device, electronic equipment and storage medium |
Also Published As
| Publication number | Publication date |
|---|---|
| KR20250090933A (en) | 2025-06-20 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12367640B2 (en) | Virtual role-based multimodal interaction method, apparatus and system, storage medium, and terminal | |
| US11281945B1 (en) | Multimodal dimensional emotion recognition method | |
| CN111368609B (en) | Voice interaction method, intelligent terminal and storage medium based on emotion engine technology | |
| CN110688911B (en) | Video processing method, device, system, terminal equipment and storage medium | |
| US20250200855A1 (en) | Method for real-time generation of empathy expression of virtual human based on multimodal emotion recognition and artificial intelligence system using the method | |
| US20250078569A1 (en) | Object emotion analysis method and apparatus and electronic device | |
| CN114495927A (en) | Multimodal interactive virtual digital human generation method and device, storage medium and terminal | |
| CN114401438A (en) | Video generation method and device for virtual digital person, storage medium and terminal | |
| CN110688008A (en) | Virtual image interaction method and device | |
| CN113689879B (en) | Method, device, electronic equipment and medium for driving virtual person in real time | |
| US11810233B2 (en) | End-to-end virtual object animation generation method and apparatus, storage medium, and terminal | |
| CN114429767B (en) | Video generation method, device, electronic device, and storage medium | |
| US20200342896A1 (en) | Conference support device, conference support system, and conference support program | |
| CN119904901B (en) | Emotion recognition methods and related devices based on large models | |
| CN118760359B (en) | Multi-mode multi-task intelligent interaction method and device, electronic equipment and storage medium | |
| CN115409923A (en) | Method, device and system for generating three-dimensional virtual image facial animation | |
| CN119441403A (en) | A digital human control method, device and electronic device based on multi-modality | |
| CN118897887A (en) | An efficient digital human interaction system integrating multimodal information | |
| WO2024088321A1 (en) | Virtual image face driving method and apparatus, electronic device and medium | |
| JP6754154B1 (en) | Translation programs, translation equipment, translation methods, and wearable devices | |
| CN120563687A (en) | AI digital human construction method and device based on cross-modal joint representation and time series analysis | |
| US20250265756A1 (en) | Method for generating audio-based animation with controllable emotion values and electronic device for performing the same. | |
| US20240320519A1 (en) | Systems and methods for providing a digital human in a virtual environment | |
| JP2017182261A (en) | Information processing apparatus, information processing method, and program | |
| CN114694687A (en) | Emotion recognition method and device, storage medium and terminal equipment |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: SANGMYUNG UNIVERSITY INDUSTRY - ACADEMY COOPERATION FOUNDATION, KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PARK, SUNG JUN;YOON, DAEHO;LEE, JUNGMIN;SIGNING DATES FROM 20240416 TO 20240417;REEL/FRAME:067344/0010 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION COUNTED, NOT YET MAILED |